|Publication number||US5680509 A|
|Application number||US 08/312,584|
|Publication date||Oct 21, 1997|
|Filing date||Sep 27, 1994|
|Priority date||Sep 27, 1994|
|Publication number||08312584, 312584, US 5680509 A, US 5680509A, US-A-5680509, US5680509 A, US5680509A|
|Inventors||Ponani S. Gopalakrishnan, David Nahamoo, Mukund Padmanabhan, Michael Alan Picheny|
|Original Assignee||International Business Machines Corporation|
|Export Citation||BiBTeX, EndNote, RefMan|
|Patent Citations (5), Non-Patent Citations (4), Referenced by (27), Classifications (11), Legal Events (5)|
|External Links: USPTO, USPTO Assignment, Espacenet|
Havg k =Pr(SLopt k)H(p/SLopt k)+Pr(SLopt k)H(p/SL.sub.optk);
The invention relates to speech recognition systems, and in particular to the estimation of phone class probabilities in such systems.
The estimation of phone probabilities is an essential part of any speech recognition system. Typically, the recognition is carried out in a maximum likelihood framework, where the probability of a given acoustic feature vector, conditioned on the phone is computed (i.e., P(acoustic/phone)). Subsequently, for all words in the vocabulary, the recognizer hypothesizes that the word is the correct word and computes a probabilistic score for this word on the basis of these probabilities, and then chooses the word with the highest score to be the correct one. This probability is typically computed in most schemes using only a very limited acoustic context. This is done by making up a feature vector at every frame by splicing together the original acoustic vector at that frame and in the adjacent few frames. However, the extent of the acoustic context that can be considered in this scheme is limited, because as the dimension of the feature vector is proportional to the size of the context, using a very large context to define the feature vector makes it prohibitively large and computationally very expensive.
This invention proposes an alternative scheme to enable the use of adjoining acoustic context, at a very small cost, to predict a probability of the phone classes at any given time. The invention uses a nonlinear decision-tree-based approach to solve this problem. The quantized feature vectors at, and in the vicinity of, a current time are used to predict a probability distribution of the phone classes, with the mechanism of prediction being a decision tree. The decision tree is constructed from training data by designing binary questions about the predictors such that the uncertainty in the predicted class is minimized by asking the question. The technique described in A. Nadas and D. Nahamoo, "Method and Apparatus for Finding the Best Splits in a Decision Tree for a Language Model for a Speech Recognizer", U.S. Pat. No. 5,263,117, issued Nov. 16, 1993, ("Nadas") is used to design these questions for each predictor and is incorporated herein by reference.
This invention also describes a method whereby the probabilities so calculated are used in conjunction with the maximum-likelihood framework to cut down the search space of the speech recognizer.
It is an object of the invention to take the given vector-quantized feature vectors at the current time t, and the adjacent N time frames on either side, and devise a decision-tree that maps them to a distribution over the phone classes. Hence, the training data used for the construction of the decision tree consists of sets of records of 2N+1 predictors and the associated class (which is assumed to be known). The alphabet size of each predictor is in the hundreds, and the class alphabet size is also typically quite large (50 or so).
The invention uses a successive data partitioning and search strategy to determine the questions of the decision tree. Starting with all the training data at the root of the tree, the invention chooses one of the 2N+1 predictors and partitions the alphabet of the predictor into two non-overlapping sets. Subsequently, for all the training records at the current node, if the value of the chosen predictor lies in the first set, the record is assigned to the first set; otherwise it is assigned to the second set. Hence, the training data at the current node is distributed between two child nodes on the basis of the set membership of the selected predictor. The predictor and the partitioning of the alphabet are chosen in such a way that after the training data is partitioned as described above, the uncertainty in the predicted class is minimized. The procedure is repeated for each child of the current node, until the class uncertainty at a node (quantified by the entropy of the class distribution at the node) falls below a certain level, or until the amount of training data at a node falls below a certain level. After the tree is constructed, the phone class distribution at the terminal nodes of the tree is available, and is stored along with the questions of the tree.
For the case of a single predictor, Nadas describes a technique to find the best binary question that minimizes the uncertainty in the predicted class. At the current node, this technique is applied independently to each of the 2N+1 predictors, and the best question for each predictor is determined. Subsequently, the best one among the 2N+1 predictors is determined as the one that provides the maximum reduction in class uncertainty, and the question at the current node is formulated as the best question for this best predictor. Alternatively, the question at a node could also be made more complex, such that it depends on more than one predictor, or an inventory of fixed complex questions could be used, and the best question chosen as the one in this inventory that provides the maximum reduction in class uncertainty.
It is another object of the invention to describe means whereby the above described decision tree can be used in a speech recognizer. During recognition, the decision tree is traversed until it reaches one of the terminal nodes, and the phone class distribution at the terminal node of the decision tree is taken to be the phone class distribution at the current node. Subsequently, to determine whether a phone can occur in a specified time interval, the probability of the phone, as predicted by the decision tree over this interval, is compared to a threshold, and a short-list of allowed phones is constructed. If the predicted probability is larger than the threshold, it is concluded that the phone can occur at that time, and if not, then the phone cannot occur at that time. The subsequent search of the decoder is constrained to this short list, rather than the space of the entire phone alphabet.
This information is used in the maximum likelihood framework to determine whether to carry out a match for a given word. Before carrying out the match for a given phone in a word, the above defined measure is checked to see if the phone can possibly occur at the given time, and if the measure predicts that it cannot, then the match for the current word is discarded.
The method and apparatus according to the invention are advantageous because they provide a fast and accurate way of estimating a-posteriori phone probabilities, using a decision tree, that is capable of taking the surrounding acoustic context into account. The method is very fast, as the questions asked in the decision tree simply involve the set membership of the selected predictor.
The invention also describes a technique where the phone probabilities so calculated can be used to reduce the complexity of the speech recognizer by cutting down its search space.
FIG. 1 is a flow chart depicting the procedure for constructing a decision tree to predict the probability distribution of a phone class at a given time, in accordance with the invention.
FIG. 2 is a flow chart describing a method of obtaining probability thresholds subsequently used by a speech recognizer.
FIG. 3 is a schematic of a preferred apparatus for constructing a decision tree and obtaining probability thresholds in accordance with the invention.
FIG. 4 is a block diagram of an automatic speech recognition system using a decision tree according to the invention.
FIG. 1 is a flow chart depicting the procedure to construct a decision tree to predict a probability distribution of phone classes at time t, given the quantized feature vectors at times t-N, t-N+1, . . . , t, . . . t+N. For the purpose of explaining the working of the invention, the quantized feature vectors will henceforth be referred to as labels. The predictors used in the decision tree are the labels at times t-N, . . . , t, . . . , t+N, represented as l-N, . . . ,l0, . . . l+N, and the predicted quantity is a distribution of the phone classes at time t. The size of the phone alphabet is denoted as P, and the size of the label alphabet as L. Typically, P ranges from 50-100, and L is in the 100's; however, for the purpose of explaining the invention, we will assume that L=4, P=3, and N=1. We will represent these four predictor values as l1, l2, l3, and l4, and the three class values as p1, p2, and p3.
The training data consists of a number of transcribed sentences, with the acoustic corresponding to each sentence being quantized into a sequence of labels. Further, as the data is transcribed (FIG. 1, Block 100), it is also possible to assign a phone to every time frame.
If the event (li k,p) is defined as one where the value of the predictor lk is equal to li and the phone class value is equal to p, then a confusion matrix is created (FIG. 1, Block 102), which enumerates the counts of all possible events (li k,p). The matrix has L rows, and P columns, and the entry corresponding to the ith row and the jth column represents the number of times the value of the predictor lk equalled li, when the class value equalled pj, in the training data at the current node of the decision tree (at the root node, all the training data is used). These counts are then converted into joint probabilities by computing the sum of all entries in the matrix, and then dividing each entry of the matrix by this sum. As there are 2N+1 predictors, 2N+1 joint distribution matrices can be created, one for each predictor. An example of these joint distribution matrices is shown in Tables 1(a)-1(c), for the case of three predictors l-1, l0 and l+1.
TABLE 1(a)______________________________________l-1 p1 p2 p3______________________________________l1 0.1 0.067 0.033l2 0.067 0.167 0.033l3 0.133 0.033 0.1l4 0.033 0.067 0.167______________________________________
TABLE 1(b)______________________________________l0 p1 p2 p3______________________________________l1 0.133 0.05 0.033l2 0.067 0.2 0.034l3 0.1 0.034 0.067l4 0.033 0.05 0.2______________________________________
TABLE 1(c)______________________________________l+1 p1 p2 p3______________________________________l1 0.117 0.05 0.033l2 0.067 0.167 0.033l3 0.116 0.05 0.1l4 0.033 0.067 0.167______________________________________
The class distribution at the current node and its entropy is computed and stored at this point. The class distribution is obtained by summing up the rows of any one of the 2N+1 joint distribution matrices, i.e., ##EQU1## and the entropy of the class distribution is obtained as ##EQU2##
For the considered example, the class distribution and its entropy is given in Table 2.
TABLE 2______________________________________p1 p2 p3______________________________________Pr 0.333 0.334 0.333______________________________________ H(p) = 1.58
In Block 103, we start with the joint distribution of the tth predictor, lk, and the class p, and design a binary partitioning SLopt k,SLopt k, of the values of the predictor lk using the method of Nadas. In other words, for each predictor, the predictor alphabet l1, l2, l3, l4 ! is partitioned into two complementary sets, SLopt k and SLopt k, (for example, SLopt k = l1, l2 !, and SLopt k = l3, l4 !, with the criterion for the selection of the partition being the minimization of the class uncertainty. The entropy of the class distribution is used as a measure of the uncertainty. The details of this method are given in Nadas. This process is carried out for each predictor independently. For the considered example, one iteration of the procedure in Nadas, col 4, line 30 col 9, line 25!, leads to a nearly optimal partitioning of the different predictors as follows:
SLopt -1 = l1,l2 !,SLopt -1 = l3,l4 !,SLopt 0 = l1,l2,l3 !, SLopt 0 = l4 !,SLopt +1 = l1,l2 ! and SLopt +1 = l3,l4 !.
Now, for each one of the predictors lk, the training data at the current node may be split into two parts based on the partitioning SLopt k,SLopt k, and the probability of these two child nodes is given as ##EQU3##
Further, the class distribution conditioned on the partitioning at the two child nodes may be calculated as follows ##EQU4##
The entropy for each of these child nodes can be calculated just as for the parent node and the average entropy for the two child nodes computed as
Hk avg =Pr(SLk opt)H(p/SLk opt)+Pr(SLk opt)H(p/SL.sup.kopt).
For the considered example, these quantities are tabulated in Table 3.
______________________________________ p1 p2 p3______________________________________Pr(p/SL opt -1) 0.358 0.5 0.142Pr(p/SL opt -1) 0.312 0.188 0.5Pr(p/SL0 opt) 0.418 0.396 0.187Pr(p/SL0 opt) 0.117 0.177 0.707Pr(p/SL opt +1) 0.394 0.465 0.141Pr(p/SL opt +1) 0.28 0.22 0.5______________________________________Pr(SL opt -1) = 0.467 Pr(SL1 opt) = 0.533H(p/SL opt -1) = 1.43 H(p/SL1 opt) = 1.477H avg -1 = 1.455Pr(SL0 opt) = 0.717 Pr(SL0 opt) =0 0.283H(p/SL0 opt) = 1.508 H(p/SL0 opt) = 1.158 Hus,20 avg = 1.409Pr(SL opt +1) = 0.467 Pr(SL1 opt) = 0.533H(p/SL opt +1) = 1.442 H(p/SL opt +1) = 1.495 Havg +1 = 1.470______________________________________
In Block 104 of FIG. 1, the reduction in class uncertainty associated with the best question for each predictor is tabulated, and the predictor providing the largest reduction in uncertainty is selected. The reduction in uncertainty due to a partitioning based on SLopt k is computed as H(p)-Havg k. For the considered example, we have H(p)=1.58, Havg -1 =1.455, Havg 0 =1.409 and Havg +1 =1.470. Hence, the selected predictor is 1°, as this gives the maximum reduction in the uncertainty of the predicted class.
In Block 5, the training data at the current node is partitioned into two parts on the basis of the optimal partitioning of the selected predictor at the current node.
Subsequently, depending on the class uncertainty and the amount of training data at a child node, the process goes back to Block 2, and starts again by recomputing the joint distribution on the basis of only the training data at the child node. The processing at a child node terminates when the class uncertainty at the child node falls below a specified threshold, or if the amount of training data at a child node falls below a specified threshold.
FIG. 2 is a flow chart describing a method of obtaining the probability thresholds, which are subsequently used in the speech recognizer. Given a stream of labels to be decoded, the object of the invention is to use the decision tree to predict a short-list of phones at every frame. Once this short-list is created, the subsequent search in the decoder is constrained to this short-list of phones, rather than the entire phone alphabet. This greatly reduces the computational complexity of the acoustic search by reducing the search space.
The thresholds are obtained during the training process. The training data is gathered at block 200 of FIG. 2. Beginning at 202 of FIG. 2, for every time frame of every sentence in the training data, the decision tree is traversed, using the label at the current time and the labels at the adjacent times as the predictors, until the traversal terminates in a node. The probability distribution on the phone classes at that current time frame, Pr(pi,t), is then taken to be the stored distribution at the node of the tree.
Referring now to the flow chart of FIG. 2, the variable i is set initially to 1. (Block 204) The thresholds for the various phones are initially set at some high value, block 206, and the variable "instance" is set to 1, block 208.
Now, any phone in the alphabet occurs several times in the training data, and every one of these instances of a particular phone is typically aligned to several time frames in the training data. The probability of phone pi predicted by the decision tree over these time frames is searched for its maximum value, Block 210, and the minimum of these maximum values over all instances of the phone pi in the training data is used as the threshold for pi (Block 212). Note that rather than obtaining a threshold for a phone pi, an alternative embodiment of the invention allows the obtaining of thresholds for combinations of phones (diphones, triphones etc.), that enable the decision tree to predict whether a particular diphone or triphone, rather than a phone, can occur at a frame.
The process of blocks 210 and 212 repeats until all instances of a particular phone have been considered (Block 214). The process of Blocks 206-214 repeats until each possible phone has been analyzed (Block 216).
Once the thresholds have been determined, they are used during speech recognition by comparing the probability predicted by the decision tree to the threshold, and deciding on the basis of this, whether or not a phone is possible at a given frame. The decoder restricts its search to the space of possible phones only, rather than considering all phones in the alphabet.
FIG. 3 schematically shows a preferred apparatus for constructing the decision tree and obtaining the probability thresholds. The apparatus may comprise, for example, an appropriately programmed computer system. In this example, the apparatus comprises a general purpose digital processor 8 having a data entry keyboard 9, a display 10, a random access memory 11, and a storage device 12. From the training data, processor 8 computes the joint distribution of the predictor lk and the phone p, for all 2N+1 predictors, using all of the training data, and stores the estimated joint distribution, along with the class distribution, in storage device 12.
Next, processor 8 computes the best partitioning of each of the predictor values such that the maximum reduction in class uncertainty is obtained due to the partitioning, according to the algorithm of Nadas. Then processor 8 chooses the best predictor, l*, and partitions the training data into two child nodes based on the best partitioning for the predictor l*.
Still under the control of the program, the processor 10 repeats the above procedure for the data at each of the two child nodes, until the class entropy at the node falls below a specified threshold, or until the amount of training data at a node falls below a specified threshold.
After the decision tree is grown, still under control of the program, the processor computes a distribution on phone classes for every frame of the training data, based on the decision tree, and stores it in storage device 12. It also stores some initial specified values of the thresholds for the various phones in storage device 12.
Next, for every instance of a given phone, pi, in the phone alphabet, the processor searches for the maximum predicted probability of phone pi over the time frames that the current instance of pi has been aligned to, and replaces the current threshold of phone pi by this quantity, if it is less than the current threshold.
FIG. 4 is a block diagram of an automatic speech recognition system which utilizes the decision tree according to the present invention. A suitable system is also described, for example, in Nadas. The system in FIG. 4 includes a microphone 13 for converting an utterance into an electrical signal. The signal from the microphone is processed by an acoustic processor and label match 14 which finds the best-matched acoustic label prototype from the acoustic label prototype store 15. A probability distribution of phone classes 16 is then produced for every time frame using the decision tree described in the invention. These probabilities are used in conjunction with the probability thresholds 19 to select a subset of acoustic word models in store 19, and a fast acoustic word match processor 18 matches the label string from the acoustic processor 14 against this subset of abridged acoustic word models to produce an utterance signal.
The utterance signal output by the fast acoustic word match processor comprises at least one word. In general, however, the fast acoustic word match processor will output a number of candidate words.
Each word signal produced by the fast acoustic word match processor 18 is input into a word context match 20 which compares the word context to language models in store 21 and outputs at least one candidate word. From the recognition candidates produced by the fast acoustic match and the language model, the detailed acoustic match 22 matches the label string from the acoustic processor 14 against detailed acoustic word models in store 23 and outputs a word string corresponding to an utterance.
Given the acoustic label string from the acoustic processor 14, the context-dependent phone probability estimator 24 traverses the decision tree 25 for every time frame using the label at the current time and the labels in the adjacent times as the predictors, until it reaches a terminal node of the tree. Then the stored class distribution at that node is stored as the class distribution at the current time. These probabilities are used by the fast acoustic word match to decide whether a phone is possible at a given time.
The abridged acoustic word models in store 19a model a word as a sequence of phones and the fast acoustic word match processor 18 computes a score for each phone in the sequence for a word by computing that match between the labels produced by acoustic processor 14 and the phone. The number of such matches that have to be computed can be reduced by using the phone probabilities 16 to prune out a number of words. The probabilities predicted by the decision tree are searched for their maximum over the time frames corresponding to the hypothesized end time of the previous phone of the word (for which a fast match score was computed), and this maximum is compared to the threshold for the phone (or diphone, triphone etc.). If the maximum is less than the threshold, then it is hypothesized that the phone cannot occur in that interval, and the current word is discarded, i.e., the fast match computation is not done for the remainder of the word. Other measures rather than the maximum could also be used (such as the sum of the probabilities of the phone over the specified time interval, etc.), with the thresholds being adjusted accordingly.
|Cited Patent||Filing date||Publication date||Applicant||Title|
|US4759068 *||May 29, 1985||Jul 19, 1988||International Business Machines Corporation||Constructing Markov models of words from multiple utterances|
|US4852173 *||Oct 29, 1987||Jul 25, 1989||International Business Machines Corporation||Design and construction of a binary-tree system for language modelling|
|US5033087 *||Mar 14, 1989||Jul 16, 1991||International Business Machines Corp.||Method and apparatus for the automatic determination of phonological rules as for a continuous speech recognition system|
|US5263117 *||Oct 26, 1989||Nov 16, 1993||International Business Machines Corporation||Method and apparatus for finding the best splits in a decision tree for a language model for a speech recognizer|
|US5267345 *||Feb 10, 1992||Nov 30, 1993||International Business Machines Corporation||Speech recognition apparatus which predicts word classes from context and words from word classes|
|1||A. Nadas et al "An iterative flip-flop approximation of the . . . " Proc. of International Conf. on Acoustics, Speech etc. '91 pp. 565-568.|
|2||*||A. Nadas et al An iterative flip flop approximation of the . . . Proc. of International Conf. on Acoustics, Speech etc. 91 pp. 565 568.|
|3||L. R. Bahl et al, "A Fast Approximate Acoustic Match for Large . . . " IEEE Trans. on Speech & Audio Processing. V. 1 #1 Jan. 93 pp. 59-67.|
|4||*||L. R. Bahl et al, A Fast Approximate Acoustic Match for Large . . . IEEE Trans. on Speech & Audio Processing. V. 1 1 Jan. 93 pp. 59 67.|
|Citing Patent||Filing date||Publication date||Applicant||Title|
|US6058205 *||Jan 9, 1997||May 2, 2000||International Business Machines Corporation||System and method for partitioning the feature space of a classifier in a pattern classification system|
|US6061654 *||Dec 16, 1996||May 9, 2000||At&T Corp.||System and method of recognizing letters and numbers by either speech or touch tone recognition utilizing constrained confusion matrices|
|US6122612 *||Nov 20, 1997||Sep 19, 2000||At&T Corp||Check-sum based method and apparatus for performing speech recognition|
|US6137863 *||Dec 13, 1996||Oct 24, 2000||At&T Corp.||Statistical database correction of alphanumeric account numbers for speech recognition and touch-tone recognition|
|US6141661 *||Oct 17, 1997||Oct 31, 2000||At&T Corp||Method and apparatus for performing a grammar-pruning operation|
|US6154579 *||Aug 11, 1997||Nov 28, 2000||At&T Corp.||Confusion matrix based method and system for correcting misrecognized words appearing in documents generated by an optical character recognition technique|
|US6205261||Feb 5, 1998||Mar 20, 2001||At&T Corp.||Confusion set based method and system for correcting misrecognized words appearing in documents generated by an optical character recognition technique|
|US6205428||Nov 20, 1997||Mar 20, 2001||At&T Corp.||Confusion set-base method and apparatus for pruning a predetermined arrangement of indexed identifiers|
|US6208965 *||Nov 20, 1997||Mar 27, 2001||At&T Corp.||Method and apparatus for performing a name acquisition based on speech recognition|
|US6219453||Aug 11, 1997||Apr 17, 2001||At&T Corp.||Method and apparatus for performing an automatic correction of misrecognized words produced by an optical character recognition technique by using a Hidden Markov Model based algorithm|
|US6400805||Jun 15, 1998||Jun 4, 2002||At&T Corp.||Statistical database correction of alphanumeric identifiers for speech recognition and touch-tone recognition|
|US6542866 *||Sep 22, 1999||Apr 1, 2003||Microsoft Corporation||Speech recognition method and apparatus utilizing multiple feature streams|
|US6640207 *||Apr 27, 2001||Oct 28, 2003||Siemens Aktiengesellschaft||Method and configuration for forming classes for a language model based on linguistic classes|
|US7562249 *||Mar 17, 2006||Jul 14, 2009||Fujitsu Limited||RAID system, RAID controller and rebuilt/copy back processing method thereof|
|US7630899||Sep 7, 2006||Dec 8, 2009||At&T Intellectual Property Ii, L.P.||Concise dynamic grammars using N-best selection|
|US7937260||Jun 15, 1998||May 3, 2011||At&T Intellectual Property Ii, L.P.||Concise dynamic grammars using N-best selection|
|US8380506 *||Nov 29, 2007||Feb 19, 2013||Georgia Tech Research Corporation||Automatic pattern recognition using category dependent feature selection|
|US8515753 *||Mar 30, 2007||Aug 20, 2013||Gwangju Institute Of Science And Technology||Acoustic model adaptation methods based on pronunciation variability analysis for enhancing the recognition of voice of non-native speaker and apparatus thereof|
|US8682665||Apr 28, 2011||Mar 25, 2014||At&T Intellectual Property Ii, L.P.||Concise dynamic grammars using N-best selection|
|US8788498||Jun 15, 2009||Jul 22, 2014||Microsoft Corporation||Labeling data samples using objective questions|
|US9286887||Mar 24, 2014||Mar 15, 2016||At&T Intellectual Property Ii, L.P.||Concise dynamic grammars using N-best selection|
|US20070101187 *||Mar 17, 2006||May 3, 2007||Fujitsu Limited||RAID system, RAID controller and rebuilt/copy back processing method thereof|
|US20080077404 *||Sep 6, 2007||Mar 27, 2008||Kabushiki Kaisha Toshiba||Speech recognition device, speech recognition method, and computer program product|
|US20080147402 *||Nov 29, 2007||Jun 19, 2008||Woojay Jeon||Automatic pattern recognition using category dependent feature selection|
|US20090119105 *||Mar 30, 2007||May 7, 2009||Hong Kook Kim||Acoustic Model Adaptation Methods Based on Pronunciation Variability Analysis for Enhancing the Recognition of Voice of Non-Native Speaker and Apparatus Thereof|
|US20100318539 *||Jun 15, 2009||Dec 16, 2010||Microsoft Corporation||Labeling data samples using objective questions|
|US20110202343 *||Apr 28, 2011||Aug 18, 2011||At&T Intellectual Property I, L.P.||Concise dynamic grammars using n-best selection|
|U.S. Classification||704/240, 704/E15.008, 704/255, 704/E15.014, 704/231|
|International Classification||G10L15/08, G10L15/06|
|Cooperative Classification||G10L15/063, G10L15/08|
|European Classification||G10L15/063, G10L15/08|
|Sep 27, 1994||AS||Assignment|
Owner name: IBM CORPORATION, NEW YORK
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GOPALAKRISHNAN, PONANI S.;NAHAMOO, DAVID;PADMANABHAN, MUKUND;AND OTHERS;REEL/FRAME:007184/0645
Effective date: 19940927
|Jan 8, 2001||FPAY||Fee payment|
Year of fee payment: 4
|Jan 24, 2005||FPAY||Fee payment|
Year of fee payment: 8
|Mar 6, 2009||AS||Assignment|
Owner name: NUANCE COMMUNICATIONS, INC., MASSACHUSETTS
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTERNATIONAL BUSINESS MACHINES CORPORATION;REEL/FRAME:022354/0566
Effective date: 20081231
|Apr 21, 2009||FPAY||Fee payment|
Year of fee payment: 12