WO1998040876A1 - Speech recognition system employing discriminatively trained models - Google Patents

Speech recognition system employing discriminatively trained models Download PDF

Info

Publication number
WO1998040876A1
WO1998040876A1 PCT/US1998/004992 US9804992W WO9840876A1 WO 1998040876 A1 WO1998040876 A1 WO 1998040876A1 US 9804992 W US9804992 W US 9804992W WO 9840876 A1 WO9840876 A1 WO 9840876A1
Authority
WO
WIPO (PCT)
Prior art keywords
models
word
model
pdfs
parameters
Prior art date
Application number
PCT/US1998/004992
Other languages
French (fr)
Other versions
WO1998040876A9 (en
Inventor
Vladimir Sejnoha
Original Assignee
Lernout & Hauspie Speech Products N.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lernout & Hauspie Speech Products N.V. filed Critical Lernout & Hauspie Speech Products N.V.
Priority to AU64636/98A priority Critical patent/AU751310B2/en
Priority to DE69818231T priority patent/DE69818231T2/en
Priority to JP53986398A priority patent/JP2002500779A/en
Priority to AT98910384T priority patent/ATE250269T1/en
Priority to EP98910384A priority patent/EP0966736B1/en
Priority to CA002275712A priority patent/CA2275712A1/en
Publication of WO1998040876A1 publication Critical patent/WO1998040876A1/en
Publication of WO1998040876A9 publication Critical patent/WO1998040876A9/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/06Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
    • G10L15/063Training
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/14Speech classification or search using statistical models, e.g. Hidden Markov Models [HMMs]
    • G10L15/142Hidden Markov Models [HMMs]
    • G10L15/144Training of HMMs

Definitions

  • the function of automatic speech recognition (ASR) systems is to determine the lexical identity of spoken utterances.
  • the recognition process also referred to as classification, typically begins with the conversion of an analog acoustical signal into a stream of digitally represented spectral vectors or frames which describe important characteristics of the signal at successive time intervals.
  • the classification or recognition process is based upon the availability of reference models which describe aspects of the behavior of spectral frames corresponding to different words. A wide variety of models have been developed but they all share the property that they describe the temporal characteristics of spectra typical to particular words or subword segments.
  • the sequence of spectral vectors arising from an input utterance is compared with the models and the success with which models of different words predict the behavior of the input frames, determines the putative identity of the utterance.
  • HMM Hidden Markov Model
  • probabilities may be provided which determine transitions between states.
  • the process of computing the probability that an unknown input utterance corresponds to a given model is usually done in one of two standard ways.
  • the first approach isJ nown as Forward-Backward algorithm, and uses an efficient recursion to compute the match probability as the sum of the probabilities of all possible alignments of the input sequence and the model states permitted by the model topology.
  • An alternative, called the Viterbi algorithm approximates t summed match probability by finding the single sequence of m states with the maximum probability.
  • the Viterbi algorithm be viewed as simultaneously performing an alignment of the of the input utterance and the model and computing the probability of that alignment.
  • HMMs can be created to model entire words, or alternatively, a variety of subword linguistic units, such as phonemes or syllables.
  • Phone-level HMMs have the advantage that a relatively compact set of models can be used to build arbitrary new words given that their phonetic transcription is known. More sophisticated versions reflect the fact that contextual effects can cause large variations in the way different phones are realized. Such models are known as allophonic or context-dependent. A common approach is to initiate the search with relatively inexpensive context- independent models and re-evaluate a small number of promising candidates with context-dependent phonetic models.
  • the pdfs take on the form of vectors of probabilities, where each component represents the probability-of observing a particular prototype vector given a particular HMM state.
  • One of the advantages of this approach is that it makes no assumptions about the nature of such pdfs, but this is offset by the information loss incurred in the quantization stage.
  • continuous pdfs eliminates the quantization step, and the probability vectors are replaced by parametric functions which specify the probability of any arbitrary input spectral vector given a state.
  • the most common class of functions used for this purpose is the mixture of Gaussians, where arbitrary pdfs are modeled by a weighted sum of Normal distributions.
  • One drawback of using continuous pdfs is that, unlike in the case of the discrete pdf, the designer must make explicit assumptions about the nature of the pdf being modeled — something which can be quite difficult since the true distribution form for the speech signal is not known.
  • continuous pdf models are computationally far more expensive than discrete pdf models, since following vector quantization the computation of a discrete probability involves no more than a single table lookup.
  • the probability values in the discrete pdf case and the parameter values of the continuous pdf are most commonly trained using the Maximum Likelihood method. In this manner, the model parameters are adjusted so that the likelihood of observing the training data given the model is maximized.
  • this approach does not necessarily lead to the best recognition performance and this realization has led to the development of new training criteria, known as discriminative, the objective of which is to adjust model parameters so as to minimize the number of recognition errors rather than fit the distributions to the data.
  • discriminative training has been applied most successfully to small-vocabulary tasks.
  • it presents a number of new problems, such as how to appropriately smooth the discriminatively-trained pdfs and how to adapt these systems to a new user with a relatively small amount of training data.
  • a recognition system should use high- resolution models which are computationally expensive (e.g., context-dependent, discriminatively-trained continuous density models).
  • high- resolution models which are computationally expensive (e.g., context-dependent, discriminatively-trained continuous density models).
  • speedup techniques are usually used.
  • the vocabulary search is performed in multiple stages or passes, where each successive pass makes use of increasingly detailed and expensive models, applied to increasingly small lists of candidate models.
  • context independent, discrete models can be used first, followed by context-dependent continuous density models.
  • a separate simultaneous alignment and pdf evaluation is essentially carried out for each set.
  • the present invention represents a novel approach to the efficient use of high- resolution models in large vocabulary recognition.
  • the proposed method benefits from the use of a continuous density model and a discriminative training criterion which leads to a high recognition performance on a large vocabulary task at the cost of only a marginal increase of computation over a simple discrete pdf system.
  • Another novel feature of the new approach is its ability to make use of limited quantities of new data for rapid adaptation to a particular speaker.
  • the probability that an input utterance corresponds to a given HMM can be computed by the Viterbi algorithm, which finds the sequence of model states which maximizes this probability. This optimization can be viewed as a
  • the alignment paths obtained with relatively computationally inexpensive discrete pdf models can be of comparable quality to those obtained with computationally costly continuous density pdf models, even though the match probabilities or metrics generated by the discrete pdf alignment do not lead to sufficiently high accuracy for large vocabulary recognition.
  • a decoupling of the alignment and final probability computation tasks A discrete-pdf system is used to establish alignment paths of an input utterance and a reference model, while the final probability metric is obtained by post-processing frame-state pairs with more powerful, discriminatively trained continuous-density pdfs, but using the same alignment path.
  • the state models in the present system are thus associated with both a discrete (low-resolution) pdf and a discriminatively trained, continuous-density (high-resolution) pdf.
  • the high-resolution pdfs are trained using alignments of models and speech data obtained using the low-resolution pdfs, and thus the discriminative training incorporates knowledge of the characteristics of the discrete pdf system.
  • each input utterance is converted to a sequence of raw or unquantized vectors. For each raw
  • Each word model is represented by a sequence of states, the states being selected from a preselected group of states. However, for each word model state, there is provided both a discrete probability distribution function (pdf) and a continuous pdf characterized by preselected adjustable parameters.
  • PDF discrete probability distribution function
  • a stored table is provided which contains distance metric values for each combination of a quantized input vector with model state as characterized by the discrete pdfs.
  • Word models are aligned with an input utterance using the respective discrete PDFs and initial match scores are generated using the stored table. From well matching word models identified from the initial match scores, a ranked scoring of those models is generated using the respective continuous pdfs and the raw vector information. After each utterance, the preselected parameters are adjusted to increase, by a small proportion, the difference in scoring between the top and next ranking models.
  • a re-adjustment of the continuous pdf parameters is accomplished by performing, on the current state of the parameters, an adjustment opposite to that performed with the original recognition event and performing on the then current state of the parameters an adjustment equal to that which would have been performed if the newly identified different word model had been the best scoring.
  • Fig. 1 is a block diagram of a speech recognition system in accordance with the present invention
  • Fig. 2 illustrates vocabulary word models used in the speech recognition system of the present invention
  • Fig. 3 illustrates a recursion procedure used in the speech recognition system of the present invention
  • Fig. 4 illustrates a training data structure set used in training word models
  • Fig. 5 is a flow chart illustrating initial, batch training of word models
  • Fig. 6 is a flow chart illustrating on-line adaptive training of word models.
  • the present invention is particularly concerned with the provision of discriminatively trained multi-resolution vocabulary models which increase accuracy and reduce computational load in an automatic speech recognition (ASR) system.
  • ASR automatic speech recognition
  • the computer system illustrated there is of the type generally referred to as a personal computer.
  • the computer runs under the MS DOS or WINDOWS® operating system and is organized around a system bus, designated generally by reference character 11.
  • the system bus may be of the so called EISA type (Extended Industry Standards Association).
  • the computer system utilizes a microprocessor, designated by reference character 13, which may, for
  • the system is also provided with an appropriate amount of local or random access memory, e.g., 32 megabytes, designated by reference character 15. Additional storage capacity is provided by a hard disk 17 and floppy diskette drive 19 which operate in conjunction with a controller 23 which couples them to the system bus.
  • a digital signal processor is provided as indicated by reference character 16, typically this processor being configured as an add-in circuit card coupled to the system bus 11.
  • the digital signal processor takes in analog signals from a microphone, designated by reference character 18, converts those signals to digital form and processes them e.g., by performing a Fast Fourier Transform (FFT), to obtain a series of spectral frames or vectors which digitally characterize the speech input at successive points in time.
  • FFT Fast Fourier Transform
  • these input vectors are referred to as the raw input vectors.
  • acoustic vectors (X u ) are generated at a rate of one every 10 ms, and have
  • the raw vectors are subjected to a gender-normalizing linear
  • the transformed acoustic frames are vector quantized with a codebook of
  • VQ vector quantizer
  • Reference vocabulary models are composed of sequences of states
  • Yj (y l ,—,y l m ,--,y lMi ), where . is the length of a model and i is the model index.
  • Each model state y 1>m is a pointer into a common set of R DTMR states
  • VQLP is essentially a table of precomputed log-probabilities and thus the evaluation of the discrete-pdf models consists of a very fast table lookup.
  • the second pool of pdfs is made up of continuous distributions which give the probability of observing a specific spectrum Xt given a particular state s ⁇ referenced by y ija , i.e., Pr(X t ly i n )
  • CDLP(X t ly im ) -log(Pr(X t ly i4n )).
  • the continuous pdfs are parametric models and thus the probabilities cannot be precomputed. Rather than storing pre-computed probabilities as is the case for the discrete pdfs, we store the pdf parameters themselves and use them to compute the log-probabilities for specific input frames. Note that individual palls in each set may be shared by acoustically similar states in different models.
  • Fig. 2 illustrates the relationship between model states and the two sets of log-pdfs. The vector-quantized input utterances V are matched against reference
  • the algorithm evaluates -log(Pr(VIY j )), the negative logarithm of the probability of observing the input sequence given the model by finding the best alignment path between the input utterance and the model
  • the original acoustic vector at a particular path point p can thus be identified as
  • Gaussian Mixtures Experimental evidence revealed that with the use of discriminative training there was no advantage to using the full mixture models over the simplified version. In addition, reducing the number of free parameters in the model significantly improves their trainability with limited quantities of data.
  • the standard Gaussian Mixture log-probability density function GMLP is defined as follows:
  • GMLP (x(t),s r ) -log( ⁇ a(s r ,k)N(x(ty, ⁇ (s r ,ky, ⁇ (s r ,k))) (3) k
  • a(s r ,k) is the weight of mixture component k in state s r and N(x; ⁇ ; ⁇ ) denotes the probability of observing x(t) given a multivariate Gaussian with mean ⁇
  • N(s r ) is the number of mixture components.
  • CDLP is not a true log-probability, and thus is not interchangeable with the discrete log-probabilities VQLP. This incompatibility is not an issue, however,
  • the rescored models are then re-sorted according to their new scores.
  • the first step in the training of the continuous density pdfs is the
  • the total number of mean vectors can be set to reflect the variance of
  • the next step consists of the discriminative training of the mean vectors.
  • modified parameters is computed and the parameters are adjusted further.
  • a training database is preprocessed by obtaining for each training
  • Each candidate list contains
  • Each list is sorted by the score D ; , and an augmented alignment path
  • b is used to store the index of the best mean vector at a particular path point.
  • FIG. 4 illustrates the training structure set for an input utterance.
  • the sizes of the sets C and I can be controlled to determine how many correct
  • Values greater than 0.5 represent recognition errors while values less than 0.5
  • the parameter ⁇ controls the amount of influence
  • the score D, between the training utterance and the target model / is
  • ⁇ (s,k,l) ⁇ (s,k,l) + ⁇ (s,k,l) (15) where w is a weight which determines the magnitude of the change to the
  • candidate models are selected using the discrete density pdfs as
  • step 101 Again using the discrete pdfs, the input utterances aligned
  • the scores are sorted as indicated at block 105
  • the models are then re-sorted based on the scores obtained with the continuous density pdfs.
  • the gradient is not accumulated but is applied
  • ⁇ (s,k,l) ⁇ (s,k,l) u _ l + ⁇ ⁇ M ⁇ (_s,A:,/) u _ 1 (16)
  • the notation ⁇ u means that the utterance u is used to compute the gradient
  • the weighting used in the on-line adaptation ⁇ ' is set much smaller than
  • the candidate alignment paths are regenerated and the utterance gradient
  • the delayed correction algorithm is as follows:
  • the candidate list and alignment paths are stored as indicated at block 159. If the user does not make a correction, the
  • utterance path is incremented, as indicated at block 163, and, if there are no
  • the procedure returns to the initial point to await a new utterance.
  • This information can either be

Abstract

In the speech recognition system disclosed herein, each input utterance is converted to a sequence of raw vectors. For each raw vector, the system identifies that one of a preselected plurality of quantized vectors which best matches the raw vector. The raw vector information is, however, retained for subsequent utilization. Each model of a vocabulary word to be recognized is in turn represented by sequence of states, the states being selected from a preselected group of states. However, for each word module state, there is provided both a discrete probability distribution function (pdf) and a continuous pdf characterized by preselected adjustable parameters. A stored table is provided which contains distance metric values for each combination of a quantized input vector with model state as characterized by the discrete pdfs. Word models are aligned with an input utterance using the respective discrete pdfs and initial match scores are generated using the stored table. From well matching word models identified from the initial match scores, a ranked scoring of those models is generated using the respective continuous pdfs and the raw vector information. After each utterance, the preselected continuous pdfs parameters are adjusted to increase, by a small proportion, the difference in scoring between the top and next ranking models. Preferably, if a user corrects a prior recognition event by selecting a different word model from the respective selected group, a re-adjustment of the continuous pdfs parameters is accomplished by performing, on the current state of the parameters, an adjustment opposite to that performed with the original recognition event and performing on the then current state of the parameters an adjustment equal to that which would have been performed if the newly identified different word model had been the best scoring.

Description

SPEECH RECOGNITION SYSTEM EMPLOYING DISCRIMINATIVELY
TRAINED MODELS
Background of the Invention
The function of automatic speech recognition (ASR) systems is to determine the lexical identity of spoken utterances. The recognition process, also referred to as classification, typically begins with the conversion of an analog acoustical signal into a stream of digitally represented spectral vectors or frames which describe important characteristics of the signal at successive time intervals. The classification or recognition process is based upon the availability of reference models which describe aspects of the behavior of spectral frames corresponding to different words. A wide variety of models have been developed but they all share the property that they describe the temporal characteristics of spectra typical to particular words or subword segments. The sequence of spectral vectors arising from an input utterance is compared with the models and the success with which models of different words predict the behavior of the input frames, determines the putative identity of the utterance.
Currently most systems utilize some variant of a statistical model called the Hidden Markov Model (HMM). Such models consist of sequences of states connected by arcs, and a probability density function (pdf) associated with each state describes the likelihood of observing any given spectral vector at that state. A separate set of
probabilities may be provided which determine transitions between states.
The process of computing the probability that an unknown input utterance corresponds to a given model, also known as decoding, is usually done in one of two standard ways. The first approach isJ nown as Forward-Backward algorithm, and uses an efficient recursion to compute the match probability as the sum of the probabilities of all possible alignments of the input sequence and the model states permitted by the model topology. An alternative, called the Viterbi algorithm, approximates t summed match probability by finding the single sequence of m states with the maximum probability. The Viterbi algorithm be viewed as simultaneously performing an alignment of the of the input utterance and the model and computing the probability of that alignment.
HMMs can be created to model entire words, or alternatively, a variety of subword linguistic units, such as phonemes or syllables. Phone-level HMMs have the advantage that a relatively compact set of models can be used to build arbitrary new words given that their phonetic transcription is known. More sophisticated versions reflect the fact that contextual effects can cause large variations in the way different phones are realized. Such models are known as allophonic or context-dependent. A common approach is to initiate the search with relatively inexpensive context- independent models and re-evaluate a small number of promising candidates with context-dependent phonetic models.
As in the case of the phonetic models, various levels of modeling power are available in the case of the probability densities describing the observed spectra associated with the states of the HMM. There are two major approaches: the discrete pdf and the continuous pdf. In the former, the spectral vectors corresponding to the input
speech are first quantized with a vector quantizer which assigns each input frame an
index corresponding to the closest vector from a codebook of prototypes. Given this
encoding of the input, the pdfs take on the form of vectors of probabilities, where each component represents the probability-of observing a particular prototype vector given a particular HMM state. One of the advantages of this approach is that it makes no assumptions about the nature of such pdfs, but this is offset by the information loss incurred in the quantization stage.
The use of continuous pdfs eliminates the quantization step, and the probability vectors are replaced by parametric functions which specify the probability of any arbitrary input spectral vector given a state. The most common class of functions used for this purpose is the mixture of Gaussians, where arbitrary pdfs are modeled by a weighted sum of Normal distributions. One drawback of using continuous pdfs is that, unlike in the case of the discrete pdf, the designer must make explicit assumptions about the nature of the pdf being modeled — something which can be quite difficult since the true distribution form for the speech signal is not known. In addition, continuous pdf models are computationally far more expensive than discrete pdf models, since following vector quantization the computation of a discrete probability involves no more than a single table lookup.
The probability values in the discrete pdf case and the parameter values of the continuous pdf are most commonly trained using the Maximum Likelihood method. In this manner, the model parameters are adjusted so that the likelihood of observing the training data given the model is maximized. However, it is known that this approach does not necessarily lead to the best recognition performance and this realization has led to the development of new training criteria, known as discriminative, the objective of which is to adjust model parameters so as to minimize the number of recognition errors rather than fit the distributions to the data.
As used heretofore, discriminative training has been applied most successfully to small-vocabulary tasks. In addition, it presents a number of new problems, such as how to appropriately smooth the discriminatively-trained pdfs and how to adapt these systems to a new user with a relatively small amount of training data.
To achieve high recognition accuracies, a recognition system should use high- resolution models which are computationally expensive (e.g., context-dependent, discriminatively-trained continuous density models). In order to achieve real-time recognition, a variety of speedup techniques are usually used.
In one typical approach, the vocabulary search is performed in multiple stages or passes, where each successive pass makes use of increasingly detailed and expensive models, applied to increasingly small lists of candidate models. For example, context independent, discrete models can be used first, followed by context-dependent continuous density models. When multiple sets of models are used sequentially during the search, a separate simultaneous alignment and pdf evaluation is essentially carried out for each set.
In other prior art approaches, computational speedups are applied to the evaluation of the high-resolution pdfs. For example, Gaussian-mixture models are evaluated by a fast but approximate identification of those mixture components which are most likely to make a significant contribution to the probability and a subsequent evaluation of those components in full. Another approach speeds up the evaluation of Gaussian-mixture models by exploiting a geometric approximation of the computation. However, even with speedups the evaluation can be slow enough that only a small number can be carried out.
In another scheme, approximate models are first used to compute the state probabilities given the input speech. All state probabilities which exceed some threshold are then recomputed using the detailed model, the rest are retained as they are. Given the new, composite set of probabilities a new Viterbi search is performed to determine the optimal alignment and overall probability. In this method, the alignment has to be repeated, and in addition, the approximate and detailed probabilities must be similar, compatible quantities. If the detailed model generates probabilities which are significantly higher than those from the approximate models the combination of the two will most likely not lead to satisfactory performance. This requirement constrains this method to use approximate and detailed models which are fairly closely related and thus generate probabilities of comparable magnitude. It should also be noted that in this method there is no guarantee that all of the individual state probabilities that make up the final alignment probability come from detailed models. The present invention represents a novel approach to the efficient use of high- resolution models in large vocabulary recognition. The proposed method benefits from the use of a continuous density model and a discriminative training criterion which leads to a high recognition performance on a large vocabulary task at the cost of only a marginal increase of computation over a simple discrete pdf system. Another novel feature of the new approach is its ability to make use of limited quantities of new data for rapid adaptation to a particular speaker.
As was mentioned above, the probability that an input utterance corresponds to a given HMM can be computed by the Viterbi algorithm, which finds the sequence of model states which maximizes this probability. This optimization can be viewed as a
simultaneous probability computation and alignment of the input utterance and the model.
In accordance with one aspect of the present invention, it has been determined that the alignment paths obtained with relatively computationally inexpensive discrete pdf models can be of comparable quality to those obtained with computationally costly continuous density pdf models, even though the match probabilities or metrics generated by the discrete pdf alignment do not lead to sufficiently high accuracy for large vocabulary recognition. In accordance with another aspect of the invention, there is provided a decoupling of the alignment and final probability computation tasks. A discrete-pdf system is used to establish alignment paths of an input utterance and a reference model, while the final probability metric is obtained by post-processing frame-state pairs with more powerful, discriminatively trained continuous-density pdfs, but using the same alignment path.
Unlike conventional systems, where model states are characterized by one particular type of observed pdf, the state models in the present system are thus associated with both a discrete (low-resolution) pdf and a discriminatively trained, continuous-density (high-resolution) pdf. The high-resolution pdfs are trained using alignments of models and speech data obtained using the low-resolution pdfs, and thus the discriminative training incorporates knowledge of the characteristics of the discrete pdf system.
Brief Description of the Invention In the speech recognition system of the present invention, each input utterance is converted to a sequence of raw or unquantized vectors. For each raw
vector the system identifies that one of a preselected plurality of quantized vectors which best matches the raw vector. The raw vector information is, however, retained for subsequent utilization. Each word model is represented by a sequence of states, the states being selected from a preselected group of states. However, for each word model state, there is provided both a discrete probability distribution function (pdf) and a continuous pdf characterized by preselected adjustable parameters. A stored table is provided which contains distance metric values for each combination of a quantized input vector with model state as characterized by the discrete pdfs.
Word models are aligned with an input utterance using the respective discrete PDFs and initial match scores are generated using the stored table. From well matching word models identified from the initial match scores, a ranked scoring of those models is generated using the respective continuous pdfs and the raw vector information. After each utterance, the preselected parameters are adjusted to increase, by a small proportion, the difference in scoring between the top and next ranking models.
Preferably, if a user corrects a prior recognition event by selecting a different word model from the respective selected group, a re-adjustment of the continuous pdf parameters is accomplished by performing, on the current state of the parameters, an adjustment opposite to that performed with the original recognition event and performing on the then current state of the parameters an adjustment equal to that which would have been performed if the newly identified different word model had been the best scoring.
Brief Description of the Drawings
Fig. 1 is a block diagram of a speech recognition system in accordance with the present invention;
Fig. 2 illustrates vocabulary word models used in the speech recognition system of the present invention; Fig. 3 illustrates a recursion procedure used in the speech recognition system of the present invention;
Fig. 4 illustrates a training data structure set used in training word models; Fig. 5 is a flow chart illustrating initial, batch training of word models; and Fig. 6 is a flow chart illustrating on-line adaptive training of word models.
Corresponding reference characters indicate corresponding elements throughout the several views of the drawings.
Description of the Preferred Embodiment As indicated previously, the present invention is particularly concerned with the provision of discriminatively trained multi-resolution vocabulary models which increase accuracy and reduce computational load in an automatic speech recognition (ASR) system. At the outset, however, it is appropriate to describe in general terms the type of speech recognition system to which the present invention is applicable. Referring now to Figure 1, the computer system illustrated there is of the type generally referred to as a personal computer. The computer runs under the MS DOS or WINDOWS® operating system and is organized around a system bus, designated generally by reference character 11. The system bus may be of the so called EISA type (Extended Industry Standards Association). The computer system utilizes a microprocessor, designated by reference character 13, which may, for
example, be an Intel Pentium type processor. The system is also provided with an appropriate amount of local or random access memory, e.g., 32 megabytes, designated by reference character 15. Additional storage capacity is provided by a hard disk 17 and floppy diskette drive 19 which operate in conjunction with a controller 23 which couples them to the system bus.
User input to the computer system is conventionally provided by means of keyboard 25 and feedback to the user is provided by means of a CRT or other video display 27 operating from the bus through a video controller 29. External communications may be provided through an I O system designated by reference character 31 which supports a serial port 33 and a printer 35. Advantageously, a fax modem may be provided as indicated by reference character 37. This is particularly useful for forwarding structured medical reports as described in co-assigned U. S . Patent No. 5,168,548. To facilitate the use of the computer system for speech recognition, a digital signal processor is provided as indicated by reference character 16, typically this processor being configured as an add-in circuit card coupled to the system bus 11. As is understood by those skilled in the art, the digital signal processor takes in analog signals from a microphone, designated by reference character 18, converts those signals to digital form and processes them e.g., by performing a Fast Fourier Transform (FFT), to obtain a series of spectral frames or vectors which digitally characterize the speech input at successive points in time. As used herein, these input vectors are referred to as the raw input vectors. In the embodiment being described, acoustic vectors (Xu) are generated at a rate of one every 10 ms, and have
14 output dimensions.
Preferably, the raw vectors are subjected to a gender-normalizing linear
discriminant analysis, as described in my co-pending, coassigned application Serial No. 08/185,500, the disclosure of which is incorporated herein by reference. The purpose of the analysis is to transform the spectral frames so as to enhance the discriminability of different phonetic events. While the raw vectors are subsequently quantized for use in alignment and initial scoring, the data comprising the raw vectors is preserved for use in more precise final scoring using continuous pdfs as described hereinafter. Thus Xu = (xU)1,.„,xUιt,.„,x11>Tu), where Tu is the length and X„ t is the t* vector of size 14 in the u* input utterance.
The transformed acoustic frames are vector quantized with a codebook of
1024 standard vector prototypes and each original spectral frame xt (omitting tbe subscript u) is assigned a corresponding vector quantizer (VQ) label vt. Each sequence X thus gives rise to a VQ label sequence V = (v,,...,v„...,vτ).
Reference vocabulary models are composed of sequences of states
Yj = (yl ,—,yl m,--,ylMi), where . is the length of a model and i is the model index.
Each model state y1>m is a pointer into a common set of R DTMR states,
S =
Figure imgf000012_0001
of which is associated in turn with two distinct types of pdf selected from two common pdf pools.
The first type of pdf pool contains discrete distributions whicb express the probability of observing a quantized frame v„ given a state _sr referenced by y1>m which occurs at the m4 position in tbe 1th model, i.e., Pr(vtl y m) = Pr(vtlsr). The computation of the match probability is simplified if the pdfs are converted to negative logarithms and thus we define the quantity VQLP((vtl yιm)) = -log(Pr(vtl y1>m)). Note that VQLP is essentially a table of precomputed log-probabilities and thus the evaluation of the discrete-pdf models consists of a very fast table lookup.
The second pool of pdfs, on the other hand, is made up of continuous distributions which give the probability of observing a specific spectrum Xt given a particular state sτ referenced by yija, i.e., Pr(Xtlyi n)
Figure imgf000013_0001
As is the case for the discrete pdfs it is more convenient to use the continuous-density probabilities in the log-domain, and thus we define CDLP(Xtlyim) = -log(Pr(Xtlyi4n)).
The continuous pdfs are parametric models and thus the probabilities cannot be precomputed. Rather than storing pre-computed probabilities as is the case for the discrete pdfs, we store the pdf parameters themselves and use them to compute the log-probabilities for specific input frames. Note that individual palls in each set may be shared by acoustically similar states in different models. Fig. 2 illustrates the relationship between model states and the two sets of log-pdfs. The vector-quantized input utterances V are matched against reference
models Y{ by the Viterbi algorithm described in greater detail hereinafter using the
discrete-pdf section of the DTMR models. The algorithm evaluates -log(Pr(VIYj)), the negative logarithm of the probability of observing the input sequence given the model by finding the best alignment path between the input utterance and the model
states.
Define Sum(t,m) as the accumulated negative log-probabilities. The alignment algorithm used in this work can then be summarized as follows.
Initialization: Sum(0, 0) = 0
Evaluation: for t := l to Tdo for m := 1 to , do Sum(ϊ, m) = VQLP(v„ yι;π + ιrdn(Sum(f - 1, w.),(Sum(t - 1, m - 1), (Sum(t, - 1)) Pred,(t, m) - arg min,(Sum(t - 1, m),(Sum(t - 1, m - 1), (Sum(r, m - 1))
Predm(t, m) = erg min Sum^ - 1, /n),(Sum(t - 1, m - 1), (Sum(t, m - 1)) end; end; Termination: return Sum, = Sυm(T,M)/(T+M) where Predt(t, m) and Predm(t, m) are the indices of the best predecessor score at position t, m.
This basic recursion is also illustrated in Fig. 3 of the drawings. The following structures are set up to store the alignment path between an input utterance and a given reference model:
= ( ι,l'"- ι,p'—! ι,Pι) 0 )
> = føι.l>-...4,,p,...,2_,P_) (2) where /ιp is the input frame index and qιp is the state index at position p on the path for the ift refererce model, and P. is the path length. The best alignment path is recovered by using the predecessor arrays Predt(t, m) and Pred^t, m) in the following backtracking recursion:
Initialization: p = 0, t = T. m = M1
Backtracking: while t >_ 1 and m > 1 do
Figure imgf000014_0001
Λ m) mnev, = Pred t, m)
' 'new m = m„ew p++ end;
Termination: Pyp
The original acoustic vector at a particular path point p can thus be identified as
x(fιp) while the state index at path position p is directly given by qιp. The
normalized scores Sum, for all the reference models aligned with a particular input utterance can be sorted and only a small number of models with the lowest normalized score need be restored in the next recognition stage. In the rescoring stage the alignment paths for a small number of the models with the best discrete-pdf scores are traversed, and new scores are computed for the frame-state pairs defined by these paths using the set of discriminatively trained continuous density pdfs. The continuous density pdfs used in this work are a simplified form of
Gaussian Mixtures. Experimental evidence revealed that with the use of discriminative training there was no advantage to using the full mixture models over the simplified version. In addition, reducing the number of free parameters in the model significantly improves their trainability with limited quantities of data. The standard Gaussian Mixture log-probability density function GMLP is defined as follows:
N sr)
GMLP (x(t),sr) = -log( ∑ a(sr,k)N(x(ty,μ (sr,ky,Σ(sr,k))) (3) k where a(sr,k) is the weight of mixture component k in state sr and N(x; μ; Σ ) denotes the probability of observing x(t) given a multivariate Gaussian with mean μ
and covariance Σ. N(sr) is the number of mixture components.
The discriminatively trained continuous density log-pdf (CDLP)
used in tthis work is as follows
Figure imgf000015_0001
where
d(x(fιp),μ (qιpk)) = (x(fιp,l) - μ (qιp,k,l))2 (5)
Figure imgf000015_0002
The continuous pdf model for state qιp thus consists of N(qιp)
14-dimensional mean vectors μ. Due to the lack of normalizing terms in equation
(4), CDLP is not a true log-probability, and thus is not interchangeable with the discrete log-probabilities VQLP. This incompatibility is not an issue, however,
because once the alignment paths are established the discrete log-probabilities are
no longer used. The ability to utilize incompatible pdfs constitutes an advantage
over known schemes.
The new score for a path corresponding to an alignment of input utterance
with reference model .' is obtained as
Dy ^∑ CDLP(x(fip,q.p) (6) ri p=ι
The rescored models are then re-sorted according to their new scores.
The role of the discrete-density component of the DTMR models is two- fold: for each input utterance it is used to screen out the great majority of incorrect
models and produce a small set of likely candidates, and it is also used to obtain accurate alignment of the input utterance to reference models. It is, however, not
called upon to provide fine discrimination between highly confusable models.
Rather, that is the role of the continuous density rescoring pdfs.
For these reasons it is sufficient to rely on conventional Maximum-
Likelihood training for the discrete-density component, and apply a discriminative
criterion to the training of the continuous density component only. The continuous
pdf training however, users alignment paths established on the basis of the
discrete pdfs.
The first step in the training of the continuous density pdfs is the
initialization of the mean vectors μsr,k. This can be done by training a
conventional Maximum Likelihood Gaussian Mixture pdf for each model state
from the input utterance frames aligned with that state using the discrete-pdf component. The total number of mean vectors can be set to reflect the variance of
the data frames aligned with each state during the iterative training. Upon
convergence of the initial training, the mean vectors with significant probabilities
are retained, while all other parameters associated with a standard Gaussian
Mixture model are discarded.
The next step consists of the discriminative training of the mean vectors.
This is accomplished by defining an appropriate training objective function which
reflects recognition error-rate and optimizing the mean parameters so as to
minimize this function.
One common technique applicable to the minimization of the objective function is gradient descent optimization. In this approach, the objective function
is differentiated with respect to the model parameters, and the parameters are then
modified by the addition of the scaled gradient. A new gradient which reflects the
modified parameters is computed and the parameters are adjusted further. The
iteration is continued until convergence is attained, usually determined by monitoring the performance on evaluation data independent from the training
data.
A training database is preprocessed by obtaining for each training
utterance a short list of candidate recognition models. Each candidate list contains
some number of correct models (subset C) and a number of incorrect (subset I)
models. Each list is sorted by the score D;, and an augmented alignment path
structure is retained for each reference model in the list. The additional stored path
information is as follows:
Figure imgf000018_0001
b, is used to store the index of the best mean vector at a particular path point. For
example, if p connects the frame x(flιP) and state q, p,
b v = ar8 ι^^ [ (x(ftιp,μ (qtιp,k )] (8) Fig. 4 illustrates the training structure set for an input utterance. An error
function εu for a particular training utterance u is computed from the pairwise
error functions o,0:
Figure imgf000018_0002
where
o = ( l +e - --^rl (10) β is a sealer multiplier, D„ i e C is the alignment score of the input token and a
correct model i, and D,, j e I is the score of the token and an incorrect model j.
The sizes of the sets C and I can be controlled to determine how many correct
models and incorrect or potential intruder models are used in the training. o10 takes on values near 1 when the correct model score D, is much greater
(i.e., worse) than the intruder score DJ5 and near 0 when the converse is true.
Values greater than 0.5 represent recognition errors while values less than 0.5
represent correct recognitions. The parameter β controls the amount of influence
"near-errors" will have on the training.
The score D, between the training utterance and the target model / is
obtained by rescoring the alignment path as shown in equation (6). In practice the
normalization by path length P. can be ignored during training. Thus
Dy ∑ CDLP(x(fιp,qιp) (11) p=\ which can be rewritten as:
P, Dy ∑ d x{fιp),μ {qι p,bιp)) (12) p=l
A similar expression can be written for D
Differentiating the error function with respect to a particular component of
the mean vector μ (s,k,l) yields:
(k,b p) ) (13)
Figure imgf000019_0001
where δ (a,b) is the Kronecker delta and equals 1 if a = b, and 0 otherwise. The
gradient is averaged over all utterances and correct-incorrect pairs: Aμ (s,k,l) = — ∑ — ~-^— (14)
U u NCJu dμ (s,k,l) whereNc Iu is the number of correct-incorrect model pairs for utterance u. The mean components are modified by the addition of the scaled gradient:
ύ(s,k,l) = μ (s,k,l) + ωΔμ (s,k,l) (15) where w is a weight which determines the magnitude of the change to the
parameter set in one iteration. This procedure is illustrated in the flowchart of Fig.
5.
Initially, candidate models are selected using the discrete density pdfs as
indicated in step 101. Again using the discrete pdfs, the input utterances aligned
with the best models using the Viterbi algorithm and the traceback information is
stored as indicated at block 103. The scores are sorted as indicated at block 105
and then the top scoring models are re-scored as indicated at block 107 using the
continuous density pdfs, the rescoring being done along the alignment path
determined with the discrete pdfs. As indicated at block 109, the models are then re-sorted based on the scores obtained with the continuous density pdfs.
Correct and incorrect models are identified as indicated at block 111 and
for each pair of correct and incorrect models an error function is computed as
indicated at block 113. Since multiple models may be used for each vocabulary
word, the procedure provides for subsets rather than just individual correct and
incorrect examples.
As indicated at block 115, a gradient is accumulated for each pair trace
backed along the correct and incorrect paths. An accumulated gradient is applied
to the continuous density pdf parameters as indicated at block 119. A test for convergence is applied as indicated at block 121 and the procedure beginning at
block 117 is repeated until the models have converged.
In each iteration of the batch mode training of the DTMR models outlined
in the previous section, all training utterances are processed before the model
parameters are updated. It is however also possible to train the models with an on¬
line adaptive algorithm, where the models are updated after each training
utterance has been processed. The on-line training makes it possible to rapidly
adapt the DTMR models with limited amounts of speech from a new user of the
recognition system, and ideally this may be done in a fashion invisible to the user,
with speech produced in the course of doing useful work with the recognition
system.
Like batch training, on-line training requires the computation of the error
function gradient for all current model parameters specified by the correct and incorrect alignment paths in the candidate set for a particular input utterance u.
Unlike in the batch training case, the gradient is not accumulated but is applied
immediately to the model parameters:
ύ(s,k,l) = μ (s,k,l)u_l + ω ΔMμ (_s,A:,/)u_1 (16)
The notation Δu means that the utterance u is used to compute the gradient, and
the operation is performed on the current model μ(s,k,l)u.j (which was presumably
also adapted on the previous utterance u - 1).
The weighting used in the on-line adaptation ω' is set much smaller than
the weighting used in the batch-mode training since the reliability of the change estimated from a single utterance is considerably lower than the estimate from a
complete training set. In other words, ω'«ω.
A complicating factor in on-line adaptation is that the identity of the input
utterances is not known with certainty. Relying on the recognition system to
identify the input utterances will inevitably lead to errors and misadaptations of
the models. Delaying the adaptation to give the user a chance to make corrections is not desirable, because given the different correction strategies favored by
different users, it is difficult to predict how long the delay needs to be.
The solution to this problem provided by the present invention is to begin
by assuming that the top-choice recognition candidate is in fact the correct answer
and to update the models immediately. However, if the user makes a correction at
some subsequent time, the original misadaptation will be undone and a new
modification of the model parameters will be performed based on the corrected
information. In order to undo the incorrect adaptation at a later time, the original input
utterance corresponding to each candidate set must be retained, although the
candidate alignment paths need not be saved. When the user specifies the correct
answer, the candidate alignment paths are regenerated and the utterance gradient
term is recomputed. The weighted gradient is subtracted from the affected model
parameters. A new gradient term, reflecting the correct target model is calculated
and applied to the DTMR parameters. This sequence of operations does not
completely undo the original error because to so do would mean undoing all the
other updates that may have intervened between the original recognition and the correction action. In practice, however, the delayed correction has proved to be as
effective as supervised adaptation, i.e. where the correct answer is identified
before gradient computation.
The delayed correction algorithm is as follows:
for u := 1 to U do
Obtain candidate list for u using μ„.,
Identify subsets Ilop.chmce (assumed incorrect models) and Clop.cholce (assumed correct models). Compute Δuμ(s, k, /)u.ι for all s, k ,1 specified by alignment paths for all pairs in Ilop choice and Cl→mcc Update μ(s, k, /)„ = μ(s, k, l)aA +ω'Δvμ(s, k, l)aA. Save candidate list (including alignment paths) for ~ if user corrects result for utterance v, [1 <= v <= u] then
Retrieve utterance v
Retrieve candidate list and alignment paths for v Identify subsets Ilop _chmce and Clop.cholce Compute Δvμ(s, k, l)a = for all s, k, I specified by alignment paths for all pairs in Itop chmce and Clop. choice
Update μ(s, k, /)„ = μ(s, k, /)„ -ω' ≠s, l)a
Identify subsets Icorrecltd and Ccomcted Compute Δvμ(s, k, /)„ for all s, k, I specified by alignment paths for all pairs in Icorrecled and CcomcUd Update μ(s, k, T)n = μ(s, k, Z)_ - 'Δvμ(s, k, l)a end end
As indicated previously, adaptation is performed as recognition is
performed on each utterrance based on the assumption that the recognition is
correct and a re-adjustment is performed only when the user makes a correction,
even though that correction may be made after the user has issued several
intervening utterances. This procedure is illustrated in the flowchart of Fig. 6.
After a candidate list is obtained as indicated at block 151, correct (C) and
incorrect (I) subsets are identified as indicated at block 153. Corrections to model
parameters are computed for all pairs C and I as indicated at block 155 and the
corrections are added to the then current model parameters, as indicated at block
157, using a relatively low weight. The candidate list and alignment paths are stored as indicated at block 159. If the user does not make a correction, the
utterance path is incremented, as indicated at block 163, and, if there are no
pending utterances, as tested at block 165, the procedure returns to the initial point to await a new utterance.
If, at the block 161, the user corrects an earlier utterance, the stored data
corresponding to the item to be corrected is retrieved as indicated at block 171.
Likewise, the candidate set and alignment paths for the utterance to be corrected
are retrieved as indicated at block 173. The correct and incorrect subsets are
identified as indicated at block 175 and the correction term is computed for all
pairs in I and C as indicated at block 179. This information can either be
computed at the time of correction or stored from the initial recognition. The
corresponding correction factor is subtracted from the then extant model parameters without attempting to undue all intervening corrections which may have been applied. The subset for the C (after correction) and I (after correction)
are identified as indicated at block 183 and correction terms are computed for all
pairs in I and C as indicated at block 183. This correction factor is then added to
the model parameters as
indicated at block 191. As indicated, a relatively low weighting factor is used in
this adjustment since it is based on a single example rather than a batch of
examples as was the case of the adjustments made during the initial or batch training illustrated in Fig. 5.
In view of the foregoing it may be seen that several objects of the present
invention are achieved and other advantageous results have been attained.
As various changes could be made in the above constructions without departing from the scope of the invention, it should be understood that all matter
contained in the above description or shown in the accompanying drawings shall
be interpreted as illustrative and not in a limiting sense.

Claims

CLAIMSWhat is claimed is:
1. In a speech recognition system in which text is created from a sequence
of input utterances and in which each input utterance is converted to a sequence of
raw vectors, a method of identifying, from a vocabulary of word models, those
models which best correspond to said utterance and for improving the word
models based on corrections made by the user, said method comprising:
providing, for each word model state, a continuous probability distribution function (PDF) characterized by preselected parameters;
using the respective continuous PDFs, calculating a ranked scoring of a group of
word models; and
inserting the word corresponding to the best scoring model into the text being
created;
after each utterance, adjusting said preselected parameters to increase the
difference in scoring between the top and next ranking models;
storing said raw vectors for each of a predetermined plurality of successive utterances and
storing the identities of the word models in the respective selected group thereby creating a historical record of recognition events; if a user corrects a prior recognition event in said text by selecting a different word model
from the respective selected group
1) performing, on the current state of said parameters, an adjustment opposite to
that performed with the original recognition event
2) performing, on the then current state of said parameters, an adjustment
equivalent to that which would have been performed if said different word model
had been best scoring.
2. A method as set forth in claim 1 further comprising:
identifying, for each raw vector in an input utterance, that one of a preselected plurality of
quantized vectors which best matches the raw vector;
providing, for each model, a sequence of states representing the corresponding vocabulary word, said states being selected from a preselected group of states;
providing, for each word model state, both a discrete probability distribution function
(PDF) and said continuous PDF;
providing a stored table containing distance degree-of-match metric values for each
combination of quantized vector with model state as characterized by said discrete PDFs; using the respective discrete PDFs, performing an alignment of at least selected ones of
said word models with an input utterance and generating initial match scores using said
stored table and said alignments;
selecting a group of well matching word models based on said initial match scores for
further scoring based on said continuous PDFs.
3. A method as set forth in claim 2 wherein the scoring based on said continuous
PDFs follows the alignment derived from said discrete PDFs.
4. In a speech recognition system in which text is created from a sequence of
input utterances and in which each input utterance is converted to a sequence of raw vectors, a method of identifying, from a vocabulary of word models, those models which
best correspond (match) to said utterance, said method comprising:
identifying, for each raw vector in an input utterance, that one of a preselected plurality of
quantized vectors which best matches the raw vector;
providing, for each model, a sequence of states representing the corresponding
vocabulary word, said states being selected from a preselected group of states;
providing, for each word model state, both a discrete probability distribution function
(PDF) and a continuous PDF characterized by Preselected parameters; providing a stored table containing distance degree-of-match metric values for each
combination of quantized vector with model state as characterized by said discrete PDFs;
using the respective discrete PDFs, performing an alignment of at least selected
ones of said word models with an input utterance and generating initial match
scores using said stored table and said alignments;
selecting a group of well matching word models based on said initial match
scores;
using the respective continuous PDFs and following the alignments obtained
using said discrete PDFs, calculating a ranked scoring of said group of word
models; and
inserting the word corresponding to the best scoring model into the text beina
created;
after each utterance, adjusting said preselected parameters to increase (by a small
proportion) the difference in scoring between the top and next ranking models.
5. A method as set forth in claim 4 further comprising:
storing said raw vectors for each of a predetermined plurality of successive
utterances and storing the identities of the word models in the respective selected group thereby creating a historical record of recognition events;
if a user corrects a prior recognition event in said text by selecting a different
word model from the respective selected group
1) performing, on the current state of said parameters, an adjustment
opposite to that performed with the oriainal recoqnition event
2) performing, on the then current state of said parameters, an adjustment
equivalent to that which would have been performed if said different word
model had been best scoring.
59385
PCT/US1998/004992 1997-03-14 1998-03-13 Speech recognition system employing discriminatively trained models WO1998040876A1 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
AU64636/98A AU751310B2 (en) 1997-03-14 1998-03-13 Speech recognition system employing discriminatively trained models
DE69818231T DE69818231T2 (en) 1997-03-14 1998-03-13 METHOD FOR THE DISCRIMINATIVE TRAINING OF VOICE RECOGNITION MODELS
JP53986398A JP2002500779A (en) 1997-03-14 1998-03-13 Speech recognition system using discriminatively trained model
AT98910384T ATE250269T1 (en) 1997-03-14 1998-03-13 METHOD FOR DISCRIMINATIVE TRAINING OF SPEECH RECOGNITION MODELS
EP98910384A EP0966736B1 (en) 1997-03-14 1998-03-13 Method for discriminative training of speech recognition models
CA002275712A CA2275712A1 (en) 1997-03-14 1998-03-13 Speech recognition system employing discriminatively trained models

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US08/818,072 1997-03-14
US08/818,072 US6260013B1 (en) 1997-03-14 1997-03-14 Speech recognition system employing discriminatively trained models

Publications (2)

Publication Number Publication Date
WO1998040876A1 true WO1998040876A1 (en) 1998-09-17
WO1998040876A9 WO1998040876A9 (en) 1999-01-28

Family

ID=25224595

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1998/004992 WO1998040876A1 (en) 1997-03-14 1998-03-13 Speech recognition system employing discriminatively trained models

Country Status (8)

Country Link
US (1) US6260013B1 (en)
EP (1) EP0966736B1 (en)
JP (1) JP2002500779A (en)
AT (1) ATE250269T1 (en)
AU (1) AU751310B2 (en)
CA (1) CA2275712A1 (en)
DE (1) DE69818231T2 (en)
WO (1) WO1998040876A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1085498A2 (en) * 1999-09-13 2001-03-21 Mitsubishi Denki Kabushiki Kaisha Reference speech pattern generating method
WO2001033551A1 (en) * 1999-11-02 2001-05-10 Speechworks International, Inc. Method and apparatus for discriminative training of acoustic models of a speech recognition system
US7680659B2 (en) * 2005-06-01 2010-03-16 Microsoft Corporation Discriminative training for language modeling

Families Citing this family (164)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7043426B2 (en) 1998-04-01 2006-05-09 Cyberpulse, L.L.C. Structured speech recognition
EP1426923B1 (en) * 1998-12-17 2006-03-29 Sony Deutschland GmbH Semi-supervised speaker adaptation
US6678658B1 (en) * 1999-07-09 2004-01-13 The Regents Of The University Of California Speech processing using conditional observable maximum likelihood continuity mapping
US6789061B1 (en) * 1999-08-25 2004-09-07 International Business Machines Corporation Method and system for generating squeezed acoustic models for specialized speech recognizer
US7689416B1 (en) 1999-09-29 2010-03-30 Poirier Darrell A System for transferring personalize matter from one computer to another
US8645137B2 (en) 2000-03-16 2014-02-04 Apple Inc. Fast, language-independent method for user authentication by voice
JP3728177B2 (en) * 2000-05-24 2005-12-21 キヤノン株式会社 Audio processing system, apparatus, method, and storage medium
WO2002001549A1 (en) * 2000-06-15 2002-01-03 Intel Corporation Speaker adaptation using weighted feedback
DE10030105A1 (en) * 2000-06-19 2002-01-03 Bosch Gmbh Robert Speech recognition device
US7080011B2 (en) * 2000-08-04 2006-07-18 International Business Machines Corporation Speech label accelerators and techniques for using same
DE60002584D1 (en) * 2000-11-07 2003-06-12 Ericsson Telefon Ab L M Use of reference data for speech recognition
US7437289B2 (en) * 2001-08-16 2008-10-14 International Business Machines Corporation Methods and apparatus for the systematic adaptation of classification systems from sparse adaptation data
US8321427B2 (en) * 2002-10-31 2012-11-27 Promptu Systems Corporation Method and apparatus for generation and augmentation of search terms from external and internal sources
JP2004191705A (en) * 2002-12-12 2004-07-08 Renesas Technology Corp Speech recognition device
US7031915B2 (en) * 2003-01-23 2006-04-18 Aurilab Llc Assisted speech recognition by dual search acceleration technique
US20040148169A1 (en) * 2003-01-23 2004-07-29 Aurilab, Llc Speech recognition with shadow modeling
US6823493B2 (en) 2003-01-23 2004-11-23 Aurilab, Llc Word recognition consistency check and error correction system and method
US20040158468A1 (en) * 2003-02-12 2004-08-12 Aurilab, Llc Speech recognition with soft pruning
US20040193412A1 (en) * 2003-03-18 2004-09-30 Aurilab, Llc Non-linear score scrunching for more efficient comparison of hypotheses
US20040186819A1 (en) * 2003-03-18 2004-09-23 Aurilab, Llc Telephone directory information retrieval system and method
US20040186714A1 (en) * 2003-03-18 2004-09-23 Aurilab, Llc Speech recognition improvement through post-processsing
US7146319B2 (en) * 2003-03-31 2006-12-05 Novauris Technologies Ltd. Phonetically based speech recognition system and method
US20040210437A1 (en) * 2003-04-15 2004-10-21 Aurilab, Llc Semi-discrete utterance recognizer for carefully articulated speech
JP4349216B2 (en) * 2004-06-21 2009-10-21 富士ゼロックス株式会社 Distribution suitability test device, consumable supply timing judgment device, image forming device, distribution suitability test method and program
US7689615B2 (en) * 2005-02-25 2010-03-30 Microsoft Corporation Ranking results using multiple nested ranking
US7689520B2 (en) * 2005-02-25 2010-03-30 Microsoft Corporation Machine learning system and method for ranking sets of data using a pairing cost function
GB2428853A (en) * 2005-07-22 2007-02-07 Novauris Technologies Ltd Speech recognition application specific dictionary
US7472096B2 (en) * 2005-07-18 2008-12-30 Microsoft Corporation Training a learning system with arbitrary cost functions
US8677377B2 (en) 2005-09-08 2014-03-18 Apple Inc. Method and apparatus for building an intelligent automated assistant
US7707032B2 (en) * 2005-10-20 2010-04-27 National Cheng Kung University Method and system for matching speech data
US7617164B2 (en) * 2006-03-17 2009-11-10 Microsoft Corporation Efficiency of training for ranking systems based on pairwise training with aggregated gradients
US20080086311A1 (en) * 2006-04-11 2008-04-10 Conwell William Y Speech Recognition, and Related Systems
US20080005067A1 (en) * 2006-06-28 2008-01-03 Microsoft Corporation Context-based search, retrieval, and awareness
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US20080147579A1 (en) * 2006-12-14 2008-06-19 Microsoft Corporation Discriminative training using boosted lasso
US8977255B2 (en) 2007-04-03 2015-03-10 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US8856002B2 (en) * 2007-04-12 2014-10-07 International Business Machines Corporation Distance metrics for universal pattern processing tasks
GB0720473D0 (en) * 2007-10-19 2007-11-28 Univ Surrey Accoustic source separation
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US20090192782A1 (en) * 2008-01-28 2009-07-30 William Drewes Method for increasing the accuracy of statistical machine translation (SMT)
US8996376B2 (en) 2008-04-05 2015-03-31 Apple Inc. Intelligent text-to-speech conversion
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US20100030549A1 (en) 2008-07-31 2010-02-04 Lee Michael M Mobile device having human language translation capability with positional feedback
US8949124B1 (en) 2008-09-11 2015-02-03 Next It Corporation Automated learning for speech-based applications
WO2010067118A1 (en) 2008-12-11 2010-06-17 Novauris Technologies Limited Speech recognition involving a mobile device
US8661030B2 (en) 2009-04-09 2014-02-25 Microsoft Corporation Re-ranking top search results
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US20120311585A1 (en) 2011-06-03 2012-12-06 Apple Inc. Organizing task items that represent tasks to perform
US9431006B2 (en) 2009-07-02 2016-08-30 Apple Inc. Methods and apparatuses for automatic speech recognition
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
WO2011089450A2 (en) 2010-01-25 2011-07-28 Andrew Peter Nelson Jerram Apparatuses, methods and systems for a digital conversation management platform
US8682667B2 (en) 2010-02-25 2014-03-25 Apple Inc. User profiling for selecting user specific voice input processing information
KR20120046627A (en) * 2010-11-02 2012-05-10 삼성전자주식회사 Speaker adaptation method and apparatus
US10762293B2 (en) 2010-12-22 2020-09-01 Apple Inc. Using parts-of-speech tagging and named entity recognition for spelling correction
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US20120310642A1 (en) 2011-06-03 2012-12-06 Apple Inc. Automatically creating a mapping between text data and audio data
US8994660B2 (en) 2011-08-29 2015-03-31 Apple Inc. Text correction processing
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US9280610B2 (en) 2012-05-14 2016-03-08 Apple Inc. Crowd sourcing information to fulfill user requests
US8515750B1 (en) * 2012-06-05 2013-08-20 Google Inc. Realtime acoustic adaptation using stability measures
US9721563B2 (en) 2012-06-08 2017-08-01 Apple Inc. Name recognition system
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9547647B2 (en) 2012-09-19 2017-01-17 Apple Inc. Voice-based media searching
US9653070B2 (en) 2012-12-31 2017-05-16 Intel Corporation Flexible architecture for acoustic signal processing engine
KR102516577B1 (en) 2013-02-07 2023-04-03 애플 인크. Voice trigger for a digital assistant
US10652394B2 (en) 2013-03-14 2020-05-12 Apple Inc. System and method for processing voicemail
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
WO2014144579A1 (en) 2013-03-15 2014-09-18 Apple Inc. System and method for updating an adaptive speech recognition model
WO2014144949A2 (en) 2013-03-15 2014-09-18 Apple Inc. Training an at least partial voice command system
US9390708B1 (en) * 2013-05-28 2016-07-12 Amazon Technologies, Inc. Low latency and memory efficient keywork spotting
WO2014197334A2 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
WO2014197336A1 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
WO2014197335A1 (en) 2013-06-08 2014-12-11 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
EP3008641A1 (en) 2013-06-09 2016-04-20 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
CN105265005B (en) 2013-06-13 2019-09-17 苹果公司 System and method for the urgent call initiated by voice command
WO2015020942A1 (en) 2013-08-06 2015-02-12 Apple Inc. Auto-activating smart responses based on activities from remote devices
US9779724B2 (en) 2013-11-04 2017-10-03 Google Inc. Selecting alternates in speech recognition
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US10592095B2 (en) 2014-05-23 2020-03-17 Apple Inc. Instantaneous speaking of content on touch devices
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
EP3149728B1 (en) 2014-05-30 2019-01-16 Apple Inc. Multi-command single utterance input method
US10289433B2 (en) 2014-05-30 2019-05-14 Apple Inc. Domain specific language for encoding assistant dialog
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9858922B2 (en) 2014-06-23 2018-01-02 Google Inc. Caching speech recognition scores
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US10446141B2 (en) 2014-08-28 2019-10-15 Apple Inc. Automatic speech recognition based on user feedback
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US9299347B1 (en) 2014-10-22 2016-03-29 Google Inc. Speech recognition using associative mapping
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US9578173B2 (en) 2015-06-05 2017-02-21 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US9786270B2 (en) 2015-07-09 2017-10-10 Google Inc. Generating acoustic models
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10229672B1 (en) 2015-12-31 2019-03-12 Google Llc Training acoustic models using connectionist temporal classification
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
DK179309B1 (en) 2016-06-09 2018-04-23 Apple Inc Intelligent automated assistant in a home environment
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10586535B2 (en) 2016-06-10 2020-03-10 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
DK179415B1 (en) 2016-06-11 2018-06-14 Apple Inc Intelligent device arbitration and control
DK201670540A1 (en) 2016-06-11 2018-01-08 Apple Inc Application integration with a digital assistant
DK179049B1 (en) 2016-06-11 2017-09-18 Apple Inc Data driven natural language event detection and classification
DK179343B1 (en) 2016-06-11 2018-05-14 Apple Inc Intelligent task discovery
US20180018973A1 (en) 2016-07-15 2018-01-18 Google Inc. Speaker verification
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
DK201770439A1 (en) 2017-05-11 2018-12-13 Apple Inc. Offline personal assistant
DK179745B1 (en) 2017-05-12 2019-05-01 Apple Inc. SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT
DK179496B1 (en) 2017-05-12 2019-01-15 Apple Inc. USER-SPECIFIC Acoustic Models
DK201770431A1 (en) 2017-05-15 2018-12-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
DK201770432A1 (en) 2017-05-15 2018-12-21 Apple Inc. Hierarchical belief states for digital assistants
DK179560B1 (en) 2017-05-16 2019-02-18 Apple Inc. Far-field extension for digital assistant services
US10706840B2 (en) 2017-08-18 2020-07-07 Google Llc Encoder-decoder models for sequence to sequence mapping

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0376501A2 (en) * 1988-12-06 1990-07-04 Dragon Systems Inc. Speech recognition system
EP0559349A1 (en) * 1992-03-02 1993-09-08 AT&T Corp. Training method and apparatus for speech recognition
US5280563A (en) * 1991-12-20 1994-01-18 Kurzweil Applied Intelligence, Inc. Method of optimizing a composite speech recognition expert
WO1995009416A1 (en) * 1993-09-30 1995-04-06 Apple Computer, Inc. Continuous reference adaptation in a pattern recognition system

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5231670A (en) * 1987-06-01 1993-07-27 Kurzweil Applied Intelligence, Inc. Voice controlled system and method for generating text from a voice controlled input
JPH0833739B2 (en) * 1990-09-13 1996-03-29 三菱電機株式会社 Pattern expression model learning device
JPH05257492A (en) * 1992-03-13 1993-10-08 Toshiba Corp Voice recognizing system
US5233681A (en) * 1992-04-24 1993-08-03 International Business Machines Corporation Context-dependent speech recognizer using estimated next word context
US5546499A (en) * 1994-05-27 1996-08-13 Kurzweil Applied Intelligence, Inc. Speech recognition system utilizing pre-calculated similarity measurements
US5428707A (en) * 1992-11-13 1995-06-27 Dragon Systems, Inc. Apparatus and methods for training speech recognition systems and their users and otherwise improving speech recognition performance
US5420335A (en) 1993-09-30 1995-05-30 Birkhahn; Ronald H. Parenteral nutrients based on watersoluble glycerol bisacetoacetates
US5572624A (en) * 1994-01-24 1996-11-05 Kurzweil Applied Intelligence, Inc. Speech recognition system accommodating different sources
US5710864A (en) * 1994-12-29 1998-01-20 Lucent Technologies Inc. Systems, methods and articles of manufacture for improving recognition confidence in hypothesized keywords
US5710866A (en) * 1995-05-26 1998-01-20 Microsoft Corporation System and method for speech recognition using dynamically adjusted confidence measure
WO1997008686A2 (en) * 1995-08-28 1997-03-06 Philips Electronics N.V. Method and system for pattern recognition based on tree organised probability densities
US5852801A (en) * 1995-10-04 1998-12-22 Apple Computer, Inc. Method and apparatus for automatically invoking a new word module for unrecognized user input
US5822730A (en) * 1996-08-22 1998-10-13 Dragon Systems, Inc. Lexical tree pre-filtering in speech recognition
US5781887A (en) * 1996-10-09 1998-07-14 Lucent Technologies Inc. Speech recognition method with error reset commands
US5899976A (en) * 1996-10-31 1999-05-04 Microsoft Corporation Method and system for buffering recognized words during speech recognition
US5864805A (en) * 1996-12-20 1999-01-26 International Business Machines Corporation Method and apparatus for error correction in a continuous dictation system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0376501A2 (en) * 1988-12-06 1990-07-04 Dragon Systems Inc. Speech recognition system
US5280563A (en) * 1991-12-20 1994-01-18 Kurzweil Applied Intelligence, Inc. Method of optimizing a composite speech recognition expert
EP0559349A1 (en) * 1992-03-02 1993-09-08 AT&T Corp. Training method and apparatus for speech recognition
WO1995009416A1 (en) * 1993-09-30 1995-04-06 Apple Computer, Inc. Continuous reference adaptation in a pattern recognition system

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1085498A2 (en) * 1999-09-13 2001-03-21 Mitsubishi Denki Kabushiki Kaisha Reference speech pattern generating method
EP1085498A3 (en) * 1999-09-13 2005-02-09 Mitsubishi Denki Kabushiki Kaisha Reference speech pattern generating method
WO2001033551A1 (en) * 1999-11-02 2001-05-10 Speechworks International, Inc. Method and apparatus for discriminative training of acoustic models of a speech recognition system
US7680659B2 (en) * 2005-06-01 2010-03-16 Microsoft Corporation Discriminative training for language modeling

Also Published As

Publication number Publication date
AU751310B2 (en) 2002-08-15
ATE250269T1 (en) 2003-10-15
CA2275712A1 (en) 1998-09-17
JP2002500779A (en) 2002-01-08
EP0966736B1 (en) 2003-09-17
US6260013B1 (en) 2001-07-10
EP0966736A1 (en) 1999-12-29
DE69818231D1 (en) 2003-10-23
DE69818231T2 (en) 2004-06-24
AU6463698A (en) 1998-09-29

Similar Documents

Publication Publication Date Title
EP0966736B1 (en) Method for discriminative training of speech recognition models
WO1998040876A9 (en) Speech recognition system employing discriminatively trained models
US6490555B1 (en) Discriminatively trained mixture models in continuous speech recognition
US5857169A (en) Method and system for pattern recognition based on tree organized probability densities
US6493667B1 (en) Enhanced likelihood computation using regression in a speech recognition system
EP0635820B1 (en) Minimum error rate training of combined string models
EP0771461B1 (en) Method and apparatus for speech recognition using optimised partial probability mixture tying
US7672847B2 (en) Discriminative training of hidden Markov models for continuous speech recognition
US6076053A (en) Methods and apparatus for discriminative training and adaptation of pronunciation networks
GB2333877A (en) Speech recognition
JPH05257492A (en) Voice recognizing system
McDermott et al. Prototype-based minimum classification error/generalized probabilistic descent training for various speech units
Ney et al. The RWTH large vocabulary continuous speech recognition system
EP1465154A2 (en) Method of speech recognition using variational inference with switching state space models
US20030023438A1 (en) Method and system for the training of parameters of a pattern recognition system, each parameter being associated with exactly one realization variant of a pattern from an inventory
US6438519B1 (en) Apparatus and method for rejecting out-of-class inputs for pattern classification
US7454341B1 (en) Method, apparatus, and system for building a compact model for large vocabulary continuous speech recognition (LVCSR) system
McDermott et al. Prototype-based discriminative training for various speech units
JP2938866B1 (en) Statistical language model generation device and speech recognition device
Moreau et al. Confidence measure and incremental adaptation for the rejection of incorrect data
JP3216565B2 (en) Speaker model adaptation method for speech model, speech recognition method using the method, and recording medium recording the method
Modi et al. Discriminative utterance verification using multiple confidence measures.
EP1008983B1 (en) Maximum likelihood linear regression (MLLR) speaker adaptation using dynamic weighting
Choi et al. Speech recognition using an enhanced FVQ based on a codeword dependent distribution normalization and codeword weighting by fuzzy objective function
JP3406672B2 (en) Speaker adaptation device

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AU CA JP

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FI FR GB GR IE IT LU MC NL PT SE

CFP Corrected version of a pamphlet front page
CR1 Correction of entry in section i
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
COP Corrected version of pamphlet

Free format text: PAGES 1/5-5/5, DRAWINGS, REPLACED BY NEW PAGES 1/5-5/5; DUE TO LATE TRANSMITTAL BY THE RECEIVING OFFICE

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 64636/98

Country of ref document: AU

WWE Wipo information: entry into national phase

Ref document number: 1998910384

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 2275712

Country of ref document: CA

Ref document number: 2275712

Country of ref document: CA

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 1998 539863

Country of ref document: JP

Kind code of ref document: A

WWP Wipo information: published in national office

Ref document number: 1998910384

Country of ref document: EP

WWG Wipo information: grant in national office

Ref document number: 64636/98

Country of ref document: AU

WWG Wipo information: grant in national office

Ref document number: 1998910384

Country of ref document: EP

DPE2 Request for preliminary examination filed before expiration of 19th month from priority date (pct application filed from 20040101)