CA2088080C - Automatic speech recognizer - Google Patents
Automatic speech recognizerInfo
- Publication number
- CA2088080C CA2088080C CA002088080A CA2088080A CA2088080C CA 2088080 C CA2088080 C CA 2088080C CA 002088080 A CA002088080 A CA 002088080A CA 2088080 A CA2088080 A CA 2088080A CA 2088080 C CA2088080 C CA 2088080C
- Authority
- CA
- Canada
- Prior art keywords
- new
- word
- words
- database
- spoken
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
- G10L15/183—Speech classification or search using natural language modelling using context dependencies, e.g. language models
- G10L15/187—Phonemic context, e.g. pronunciation rules, phonotactical constraints or phoneme n-grams
Abstract
Apparatus and method for recording data in a speech recognition system and recognizing spoken data corresponding to the recorded data. The apparatus and method responds to entered data by generating a string of phonetic transcriptions from the entered data. The data and generated phonetic transcription string associated therewith is recorded in a vocabulary lexicon of the speech recognition system. The apparatus and method responds to receipt of spoken data by constructing a model of subwords characteristic of the spoken data and compares the constructed subword model with ones of the recorded lexicon vocabulary recorded phonetic transcription strings to recognize the spoken data as the data identified by and associated with a phonetic transcription string matching the constructed subword string.
Description
-1- 20~3~0 AUTOMATIC SPEECH RECOGNIZER
Field of the Invention The invention relates to a~p~alus and method for generating word models used in speech recognition systems and in particular to word models used 5 to recognize spoken words.
Back~round and Problem Speech recognition systems are increasingly used in applications to recognize and respond to spoken hlrc~lmation. In a typical application, a department store uses a speech recognition system in telephone sales operations to inform 10 calling CU~C~ of new merch~nrlise and to accept customer spoken orders.
Another application may be used by a stock brokerage firm to respond to a calling customer voice request for a stock quotation by verbally quoting current information relating to the calling customer's account.
Speech recognition systems usually have a cl~t~b~e lexicon storing 15 patterns, such as voice templ~tes or models, representing complete numbers, words and phrases used in various applications. In operation, received verbal hlrc,lmation is cclllpaled with templates to identify the received verbal information as specific numbers, words or phrases. The lexicon templates are created and llpd~t~d in a laboratory by technicians who collect data samples of speech from a large number20 of people. The speech data samples are verified manually and used to construct a template model of a number, word or phrase. The template model is subsequently added to the dflt~b~ce lexicon of the a~r~liate speech recognition system.
A problem arises in that individual speech recognition systems require database lexicons of voice templates that are orle~ es used only by the system 25 owner and the owner's customers and are unique thereto. For example, a hospital speech recognition system requires a lexicon of voice templates that are dirre~
and distinct from the voice templates recorded in a lexicon of a grocery wholesaler speech recognition system. Thus, each speech recognition system requires the collection of a large number of speech data samples and the manual verification of 30 the speech data samples in order to construct templates required for each specific speech recognition system application. Accordingly, a need exists for an automatic speech recognizer for speech recognition systems and a method of operating the automatic speech recognizer that enables an owner of the system to enter information used in the owner's business application into a speech recognition - 2 - ~ 2 ~ 8 8 0 8 0 system without requiring the collection and verification of large numbers of speech data samples.
Solutions The foregoing problem is solved by apparatus and method arranged for 5 entering data into a speech recognition system comprising: a vocabulary database that contains a plurality of words and a corresponding phoneme string for each of said words; a subword database that stores a plurality of phonemes representing word sounds; means for receiving a new non-spoken word in dictionary form as a series of characters, said new word not previously contained in said vocabulary database;
10 processor means ascertains a new phoneme string based on said new word and enters said new word and new phoneme string in said vocabulary database thereby enabling a subsequent spoken entry of the new word to be recognized based on the presenceof the new word in the vocabulary database.
Brief Description of the Drawin~
FIG. 1 illustrates a speech recognition system embodying the principles of the invention;
FIG. 2 sets forth the details of the processor of the speech recognition system illustrated in FIG. 1 in accordance with the principles of the invention;FIG. 3 sets forth illustrative processor programs and database files used with the speech recognition system set forth in FIGS. 1 and 2; and FIGS. 4 and 5 illustrate flowcharts of the operation of the speech recognition system set forth in FIG. 1 in accordance with the principles of the mventlon.
Detailed Description In an exemplary embodiment of the invention, speech recognition system 1, FIG. 1, is arranged to recognize spoken data, such as alphanumeric data.
Typically, such systems may be used in a wide variety of applications such as the order department of a large store. Such systems may be, although not limited thereto, connected with telephone lines, such as telephone line 31, and are used to transmit a voice prompt message to a calling customer using telephone 3. Speech recognition system I responds to the calling customer by recognizing alphanumeric data spoken by the customer in placing an order with the store. In other applications, speech recognition system I may be used in telephone operator applications to recognize calling party spoken telephone numbers and by stockbroker firms to . ~
Field of the Invention The invention relates to a~p~alus and method for generating word models used in speech recognition systems and in particular to word models used 5 to recognize spoken words.
Back~round and Problem Speech recognition systems are increasingly used in applications to recognize and respond to spoken hlrc~lmation. In a typical application, a department store uses a speech recognition system in telephone sales operations to inform 10 calling CU~C~ of new merch~nrlise and to accept customer spoken orders.
Another application may be used by a stock brokerage firm to respond to a calling customer voice request for a stock quotation by verbally quoting current information relating to the calling customer's account.
Speech recognition systems usually have a cl~t~b~e lexicon storing 15 patterns, such as voice templ~tes or models, representing complete numbers, words and phrases used in various applications. In operation, received verbal hlrc,lmation is cclllpaled with templates to identify the received verbal information as specific numbers, words or phrases. The lexicon templates are created and llpd~t~d in a laboratory by technicians who collect data samples of speech from a large number20 of people. The speech data samples are verified manually and used to construct a template model of a number, word or phrase. The template model is subsequently added to the dflt~b~ce lexicon of the a~r~liate speech recognition system.
A problem arises in that individual speech recognition systems require database lexicons of voice templates that are orle~ es used only by the system 25 owner and the owner's customers and are unique thereto. For example, a hospital speech recognition system requires a lexicon of voice templates that are dirre~
and distinct from the voice templates recorded in a lexicon of a grocery wholesaler speech recognition system. Thus, each speech recognition system requires the collection of a large number of speech data samples and the manual verification of 30 the speech data samples in order to construct templates required for each specific speech recognition system application. Accordingly, a need exists for an automatic speech recognizer for speech recognition systems and a method of operating the automatic speech recognizer that enables an owner of the system to enter information used in the owner's business application into a speech recognition - 2 - ~ 2 ~ 8 8 0 8 0 system without requiring the collection and verification of large numbers of speech data samples.
Solutions The foregoing problem is solved by apparatus and method arranged for 5 entering data into a speech recognition system comprising: a vocabulary database that contains a plurality of words and a corresponding phoneme string for each of said words; a subword database that stores a plurality of phonemes representing word sounds; means for receiving a new non-spoken word in dictionary form as a series of characters, said new word not previously contained in said vocabulary database;
10 processor means ascertains a new phoneme string based on said new word and enters said new word and new phoneme string in said vocabulary database thereby enabling a subsequent spoken entry of the new word to be recognized based on the presenceof the new word in the vocabulary database.
Brief Description of the Drawin~
FIG. 1 illustrates a speech recognition system embodying the principles of the invention;
FIG. 2 sets forth the details of the processor of the speech recognition system illustrated in FIG. 1 in accordance with the principles of the invention;FIG. 3 sets forth illustrative processor programs and database files used with the speech recognition system set forth in FIGS. 1 and 2; and FIGS. 4 and 5 illustrate flowcharts of the operation of the speech recognition system set forth in FIG. 1 in accordance with the principles of the mventlon.
Detailed Description In an exemplary embodiment of the invention, speech recognition system 1, FIG. 1, is arranged to recognize spoken data, such as alphanumeric data.
Typically, such systems may be used in a wide variety of applications such as the order department of a large store. Such systems may be, although not limited thereto, connected with telephone lines, such as telephone line 31, and are used to transmit a voice prompt message to a calling customer using telephone 3. Speech recognition system I responds to the calling customer by recognizing alphanumeric data spoken by the customer in placing an order with the store. In other applications, speech recognition system I may be used in telephone operator applications to recognize calling party spoken telephone numbers and by stockbroker firms to . ~
3 2~ 08D
recognize cuslom~ orders and requests for information. In still other applications, the principles of speech recognition system 1 may be used with a colllpulel to enable the C~ JUh,. to recognize alphanumeric data spoken by a user of the COIll~U~.,l.Speech recognizing system 1 may be, although not limited thereto, a S cc,~ulef 10 such as an AT&T 386 or 486 compuler. Such colllpul~ , FIG. 2, neednot be described in detail for an understanding of the invention and in general have a central processing unit 101, memory unit 102, and an int~o.rf~ce unit 100 eachconnected by address, data and control leads to a data bus 104. Tnt~ ce unit 100serves to couple a data link 1110 inl~ ;ollllected with data input device, such as 10 keyboard 11 with data bus 104, so that alphanumeric data, such as words, numerals and inrolllla~ion, entered into keyboard 11 may be exchanged with central processor unit 101 and memory unit 102. Data bus 104 is also interconnected with data storage device 103 used to store and record database infrastructures, such as tl~t~b~ce infrastructures 1030, 1031 and 1032, by address, data and control leads.
15 Colll~ulel 10, FIG. 1, used as speech recognition system 1, may also have a video termin~l 12 used to display data and hlfollllation.
In operation, FIG. 2, programs, such as generate subword spelling of new word program 1020, aulollla~ic speech recognition (ASR) alg~ hlll program 1021 and keyboard input program 1022, are stored in memor,v unit 102 and20 are loaded into central processor unit 101 at the a~plv~liate time to control operation of speech recognizer system 1 in accor~allce with the principles of the invention.
Database infrastructures dictionary database 1030, vocabulary lexicon d~t~b~e 1031, and subword model d~t~b~e 1032 are stored in data store 103 and are flccessed by central processor unit 102 to read information from and to store 25 information therein.
Dictionary d~t~b~ce 1030, FIG. 3, stores words and a phonetic transcription of each stored word. Typically, the words may be those words found in a standard type of dictionary and are stored in dictionary database 1030 along with a phonetic transcription of the stored word or may be col~u~ed by generate subword30 spelling of new word program 1020. For example, the word "house" may be represented by the associated phonetic transcription "haus". Similarly, the stored word "car" may be represented by the associated phonetic transcription "kaar"
and the word "com~)ulcl" by the associated phonetic transcription "kempyuter".
In addition, letters such as "d" and "w" may be represented by the transcriptions 35 "diy" and "dahbixlyuw".
recognize cuslom~ orders and requests for information. In still other applications, the principles of speech recognition system 1 may be used with a colllpulel to enable the C~ JUh,. to recognize alphanumeric data spoken by a user of the COIll~U~.,l.Speech recognizing system 1 may be, although not limited thereto, a S cc,~ulef 10 such as an AT&T 386 or 486 compuler. Such colllpul~ , FIG. 2, neednot be described in detail for an understanding of the invention and in general have a central processing unit 101, memory unit 102, and an int~o.rf~ce unit 100 eachconnected by address, data and control leads to a data bus 104. Tnt~ ce unit 100serves to couple a data link 1110 inl~ ;ollllected with data input device, such as 10 keyboard 11 with data bus 104, so that alphanumeric data, such as words, numerals and inrolllla~ion, entered into keyboard 11 may be exchanged with central processor unit 101 and memory unit 102. Data bus 104 is also interconnected with data storage device 103 used to store and record database infrastructures, such as tl~t~b~ce infrastructures 1030, 1031 and 1032, by address, data and control leads.
15 Colll~ulel 10, FIG. 1, used as speech recognition system 1, may also have a video termin~l 12 used to display data and hlfollllation.
In operation, FIG. 2, programs, such as generate subword spelling of new word program 1020, aulollla~ic speech recognition (ASR) alg~ hlll program 1021 and keyboard input program 1022, are stored in memor,v unit 102 and20 are loaded into central processor unit 101 at the a~plv~liate time to control operation of speech recognizer system 1 in accor~allce with the principles of the invention.
Database infrastructures dictionary database 1030, vocabulary lexicon d~t~b~e 1031, and subword model d~t~b~e 1032 are stored in data store 103 and are flccessed by central processor unit 102 to read information from and to store 25 information therein.
Dictionary d~t~b~ce 1030, FIG. 3, stores words and a phonetic transcription of each stored word. Typically, the words may be those words found in a standard type of dictionary and are stored in dictionary database 1030 along with a phonetic transcription of the stored word or may be col~u~ed by generate subword30 spelling of new word program 1020. For example, the word "house" may be represented by the associated phonetic transcription "haus". Similarly, the stored word "car" may be represented by the associated phonetic transcription "kaar"
and the word "com~)ulcl" by the associated phonetic transcription "kempyuter".
In addition, letters such as "d" and "w" may be represented by the transcriptions 35 "diy" and "dahbixlyuw".
4 208808~
Vocabulary ASR lexicon d~t~b~se 1031 is used to record and store words and variations thereof with strings of phonetic transcriptions r~lesellting the words and variations of the recorded words. In a typical example, generate subword spelling of new word program 1020 will record the word "car" and the associated 5 phonetic transcription string "kaar" in vocabulary lexicon database 1031. In another example, the variation of the words "rail" and "road" may be recorded as "railroad"
with the associated phonetic transcription string "le~lrowd".
Subword model ~t~b~ce 1032 stores phonemes each identifying a small unit of speech that serves to distinguish one utterance from another. For example, 10 the phoneme "aa" representing a sound as in "cot" is stored in subword model database 1032. Similarly, tl~t~h~se 1032 stores phonemes "ah", "d", "ey"
representing sounds as in "buf", "dad" and "bait", respectively.
A word or numeral, or strings thereof, FIG. 1, may be entered into speech recognition system 1 by typing the word or numeral on keys 110 and 15 keypad 111 of keyboard 11. In other applications, new and additional words and numerals may be added to speech recognition system 1 by entering the words and numericals by tape, disk drives, extra memory or any of a number of dirr~,le.l~ and well-known means of entering data into com~u~ 10 by programs such as keyboard input program 1022. Generate subword spelling of new word program 1032, FIG. 2, 20 controlling the operation of central processor unit 101 in accordance with anembodiment of the invention, responds to entry of words and numerals by accessing dictionary d~t~b~ce 1030. When the entered word is found, for example, "house", FIG. 3, generate subword spelling of new word program 1020 generates a phonetic transcription string "haus" from the dictionary data dictionary phonetic transcriptions 25 identified by the entered word. When a variation of words stored in dictionary d~t~b~ce 1030 is entered into speech recognition system 1, such as a word "railroad", generate subword spelling of new word program 1020 accesses dictionary database 1030 and generates a phonetic transcription string "reylrowd" from dictionary database 1030 phonetic transcriptions "reyl" and "rowd" stored with words "rail" and "road" identified by the entered "railroad". Generate subword spelling of new word program 1020 records each entered word and the associated generated phonetic transcription string in vocabulary lexicon d~t~b~ce 1031. Thus, entered words "car" and "railroad" and their associated phonetic strings "kaar" and "reylrowd" are recorded in vocabulary lexicon database 1031 for use in recognizing 35 spoken words received by speech recognition system 1.
20880~0 Upon receipt of spoken alph~numetic data over telephone line 31, FIG. 1, automatic speech recognition (ASR) algorithm program 1021, FIG. 3, açcesses subword model d~t~b~e 1032 and constructs a model comprising a string of phonemes each identifying a small unit of speech that serves to distinguish one 5 utterance from another and which represent sounds of the received spoken word.Thus, when the spoken word "car" is received, ASR algolilhlll program 1021 accesses subword model ~l~t~h~se 1032 and constructs phonemes "k", "aa" and "r"
as a phoneme string "k aa r" identifying the received spoken word "car". In a like manner, ASR algorithm program 1021 responds to receipt of the spoken word 10 "railroad" by ~ccessing subword model d~t~b~e 1032 and constructing a model comprising the string of phonemes "r ey 1 r ow d" representing the received spoken word "railroad".
After constructing a string phoneme model representing the received spoken word, ASR algorithm program 1021 accesses vocabulary lexicon 15 database 1031 and COl~)al~;S the constructed string phoneme model with ones of vocabulary lexicon d~t~ba~e 1031 recorded phonetic transcription strings. When the constlucted string phoneme model matches a phonetic transcription string recorded in vocabulary lexicon ~l~t~h~e 1031, the spoken word is recognized as the recorded word associated with the matched phonetic transcription string. Thus, when speech 20 recognizer system 1, FIG. 1, receives the spoken word "railroad", ASR algorithm program 1021, FIG. 3, ~ccesses subword model d~t~b~e 1032 and constructs a string phoneme model "r ey 1 r ow d". Program 1021 then accesses vocabulary lexicon ~l~t~b~e 1031 and CO~ ,S the constructed string phoneme model "r ey 1 r ow d" with ones of the vocabulaTy lexicon recorded phonetic transcription 25 strings. The spoken word "railroad" is recognized as the word "railroad" recorded with the phonetic transcription string "reylrowd" when the constructed string phoneme model "r ey 1 r ow d" matches the recorded phonetic transcription string "reylrowd " .
Speech recognizing system 1, FIG. 1, normally has a large number of 30 words stored and recorded in dictionary and vocabulary lexicon databases 1030, 1031, respectively. Additional words may be entered into speech recognizing system 1 and recorded in vocabulary lexicon database 1031 to enable the system to recognize the additional words. Such words may be unique words used in specific applications of speech recognizing system 1. In the present embodiment of the 35 invention, it is ~sllmçd that a special word "dwd" unique to a specific application is to be entered into speech recognizing system 1.
2 ~ 8 ~
The method of operating speech recognizing system 1 for recognizing a spoken word such as "dwd" comprises generating a phonetic transcription string"diydahbixlyuwdiy" from the word "dwd" and recording both the word "dwd" and generated phonetic transcription string "diydahbixlyuwdiy" in vocabulary lexicon5 d~t~b~e 1031, FIG. 3. Upon receiving the spoken word "dwd", the method of operating speech recognizing system 1 ~ccesses subword model database 1032 and constructs a model string of phonemes "d iy d ah b ix 1 y uw d iy" characteristic of the sounds of the spoken word "dwd". The constructed phoneme string model "d iy d ah b ix 1 y uw d iy" is cc.lllpal~d to ones of the lexicon vocabulary recorded 10 phonetic transcription strings and when there is a match of the constructed phoneme string model "d iy d ah b ix l y uw d iy" with the vocabulary lexicon database 1031 recorded phonetic transcription string "diydahbixlyuwdiy", the spoken word is recognized as the word "dwd" recorded with the matched phonetic transcription string "diydahbixlyuwdiy " .
In the method of operation, FIG. 1, a dictionary of standard words and a phonetic transcription of each standard word is initially stored in dictionarycl~t~b~e 1030. Similarly, subword model database 1032 initially stores phonemes,such as "d" and "w", each ide.l~irying a small unit of speech that serves to distinguish one utterance from another. To enter a new word, such as the word "dwd", into 20 speech recognizing system 1, the word "dwd" is typed into keyboard 11 and entered into keyword input program 1022, steps 102200, 102201, FIG. 4. Generate subword spelling of new word program 1020, FIG. 1, responds to the entry of alph~u~ f ~ic data and variations thereof, such as the word "dwd", by accessing dictionary ~l~t~b~ce 1030, steps 102000, 102001, FIG. 4. If the entered word is located in 25 dictionary ~t~b~e 1030, step 102002, generate subword spelling of new word program 1020 selects the word and associated phonetic transcription string, step 102003, and records both the word and associated phonetic transcription string in vocabulary lexicon d~t~b~e 1031, steps 102005, 102006.
When the entered word, for example word variation "dwd", is not found 30 in dictionary tl~t~b~e 1030, step 102002, generate subword spelling of new word program 1020 selects component words, FIGS 3, "d" and "w" from dictionary d~t~b~e 1030 and generates the phonetic transcription string "diydahbixlyuwdiy",step 102004. The word variation "dwd" and generated phonetic transcription string "diydahbixlyuwdiy" is recorded in vocabulary lexicon (l~t~b~ce 1031, steps 102005, 35 102006. If more words are to be entered into speech recognition system 1, step 102007, step 102201 and steps 102000 through 102007 are repeated until 20880~0 all the entered words have been recorded in vocabulary lexicon database 1031 at which time central processor unit 101 exits generate subword spelling of new word program 1020, step 102008.
Speech recognition system 1, FIG. 1, operating in accor~ ce with ASR
5 algoliLLlll program 1021, responds to a spoken word, such as word "dwd", received over telephone line 31, FIG. 5, steps 102100, 102101, 102102, by ~cGes~ing subword model database 1032. In response to the receipt of the spoken word "dwd", ASR
algc,liLl~ program 1021, step 102103, initi~tes a construction of a model comprising a string of phonemes representing the received spoken word "dwd". Construction of 10 the model is in accordance with the sound of the received "dwd" and the sound "d"
results in the selection of "d iy" for the first part of the string model, step 102104.
Since the model is not completed, step 102105, steps 102101 through 102105 are repeated such that the sound of "w" results in the phonemes "d ah b ix 1 y uw"
being selected from subword model d~t~b~ce 1032 and added to the model.
15 The phoneme string model is completed when the last sound "d" results in the phonemes "d iy" being added to the string to complete the phoneme string model "d iy d ah b ix 1 y uw d iy" representation of the spoken word "dwd", step 102105.
Upon the completion of the constructed phoneme string model "d iy d ah b ix 1 y uw d iy" represent~tion of the spoken word "dwd", ASR
20 algc~liLhlll program 1021 ~ccesses vocabulary lexicon d~t~b~e 1031 and COlll~ ,S
the constructed string phoneme model "d iy d ah b ix 1 y uw d iy" with ones of the vocabulary lexicon d~t~b~e recorded phonetic transcription strings, steps 102106, 102107, 102108. If there is a match of the constructed phoneme string model "d iy d ah b ix 1 y uw d iy" with the recorded phonetic transcription string 25 "diydahbixlyuwdiy", step 102109, the spoken word "dwd" is recognized as the word "dwd" recorded in vocabulary lexicon database 1031, step 102110. When there are additional spoken words, step 102111, steps 102101 through 102111 are repeated until all words have been recognized.
Should there be a failure to match, step 102109, and the match attempt 30 is less than a predetermined attempt nulllbel, step 102113, ASR alg~ hm program 1021 increases the repeat attempt number and plolllpls the caller to repeat the spoken word, step 102114. Steps 102101 through 102109 are then repeated.
Should the attempt number exceed the predete~ninçd attempt number, it is determined that an error has occurred, step 102115.
20~80~0 During the matching process, step 102109, a confidence recognition factor is assigned to the match of the constructed phoneme string model with phonetic transcription strings recorded in vocabulary lexicon rl~t~b~e 1031.
When the assigned confidence recognition factor exceeds a pre~leterrnined threshold 5 value, a detennin~ti-n is made that the spoken word associated with the constructed phoneme string model is recognized as the word recorded in vocabulary lexicon cl~t~b~e 1031 and associated with the matched phonetic transcription string.
Vocabulary ASR lexicon d~t~b~se 1031 is used to record and store words and variations thereof with strings of phonetic transcriptions r~lesellting the words and variations of the recorded words. In a typical example, generate subword spelling of new word program 1020 will record the word "car" and the associated 5 phonetic transcription string "kaar" in vocabulary lexicon database 1031. In another example, the variation of the words "rail" and "road" may be recorded as "railroad"
with the associated phonetic transcription string "le~lrowd".
Subword model ~t~b~ce 1032 stores phonemes each identifying a small unit of speech that serves to distinguish one utterance from another. For example, 10 the phoneme "aa" representing a sound as in "cot" is stored in subword model database 1032. Similarly, tl~t~h~se 1032 stores phonemes "ah", "d", "ey"
representing sounds as in "buf", "dad" and "bait", respectively.
A word or numeral, or strings thereof, FIG. 1, may be entered into speech recognition system 1 by typing the word or numeral on keys 110 and 15 keypad 111 of keyboard 11. In other applications, new and additional words and numerals may be added to speech recognition system 1 by entering the words and numericals by tape, disk drives, extra memory or any of a number of dirr~,le.l~ and well-known means of entering data into com~u~ 10 by programs such as keyboard input program 1022. Generate subword spelling of new word program 1032, FIG. 2, 20 controlling the operation of central processor unit 101 in accordance with anembodiment of the invention, responds to entry of words and numerals by accessing dictionary d~t~b~ce 1030. When the entered word is found, for example, "house", FIG. 3, generate subword spelling of new word program 1020 generates a phonetic transcription string "haus" from the dictionary data dictionary phonetic transcriptions 25 identified by the entered word. When a variation of words stored in dictionary d~t~b~ce 1030 is entered into speech recognition system 1, such as a word "railroad", generate subword spelling of new word program 1020 accesses dictionary database 1030 and generates a phonetic transcription string "reylrowd" from dictionary database 1030 phonetic transcriptions "reyl" and "rowd" stored with words "rail" and "road" identified by the entered "railroad". Generate subword spelling of new word program 1020 records each entered word and the associated generated phonetic transcription string in vocabulary lexicon d~t~b~ce 1031. Thus, entered words "car" and "railroad" and their associated phonetic strings "kaar" and "reylrowd" are recorded in vocabulary lexicon database 1031 for use in recognizing 35 spoken words received by speech recognition system 1.
20880~0 Upon receipt of spoken alph~numetic data over telephone line 31, FIG. 1, automatic speech recognition (ASR) algorithm program 1021, FIG. 3, açcesses subword model d~t~b~e 1032 and constructs a model comprising a string of phonemes each identifying a small unit of speech that serves to distinguish one 5 utterance from another and which represent sounds of the received spoken word.Thus, when the spoken word "car" is received, ASR algolilhlll program 1021 accesses subword model ~l~t~h~se 1032 and constructs phonemes "k", "aa" and "r"
as a phoneme string "k aa r" identifying the received spoken word "car". In a like manner, ASR algorithm program 1021 responds to receipt of the spoken word 10 "railroad" by ~ccessing subword model d~t~b~e 1032 and constructing a model comprising the string of phonemes "r ey 1 r ow d" representing the received spoken word "railroad".
After constructing a string phoneme model representing the received spoken word, ASR algorithm program 1021 accesses vocabulary lexicon 15 database 1031 and COl~)al~;S the constructed string phoneme model with ones of vocabulary lexicon d~t~ba~e 1031 recorded phonetic transcription strings. When the constlucted string phoneme model matches a phonetic transcription string recorded in vocabulary lexicon ~l~t~h~e 1031, the spoken word is recognized as the recorded word associated with the matched phonetic transcription string. Thus, when speech 20 recognizer system 1, FIG. 1, receives the spoken word "railroad", ASR algorithm program 1021, FIG. 3, ~ccesses subword model d~t~b~e 1032 and constructs a string phoneme model "r ey 1 r ow d". Program 1021 then accesses vocabulary lexicon ~l~t~b~e 1031 and CO~ ,S the constructed string phoneme model "r ey 1 r ow d" with ones of the vocabulaTy lexicon recorded phonetic transcription 25 strings. The spoken word "railroad" is recognized as the word "railroad" recorded with the phonetic transcription string "reylrowd" when the constructed string phoneme model "r ey 1 r ow d" matches the recorded phonetic transcription string "reylrowd " .
Speech recognizing system 1, FIG. 1, normally has a large number of 30 words stored and recorded in dictionary and vocabulary lexicon databases 1030, 1031, respectively. Additional words may be entered into speech recognizing system 1 and recorded in vocabulary lexicon database 1031 to enable the system to recognize the additional words. Such words may be unique words used in specific applications of speech recognizing system 1. In the present embodiment of the 35 invention, it is ~sllmçd that a special word "dwd" unique to a specific application is to be entered into speech recognizing system 1.
2 ~ 8 ~
The method of operating speech recognizing system 1 for recognizing a spoken word such as "dwd" comprises generating a phonetic transcription string"diydahbixlyuwdiy" from the word "dwd" and recording both the word "dwd" and generated phonetic transcription string "diydahbixlyuwdiy" in vocabulary lexicon5 d~t~b~e 1031, FIG. 3. Upon receiving the spoken word "dwd", the method of operating speech recognizing system 1 ~ccesses subword model database 1032 and constructs a model string of phonemes "d iy d ah b ix 1 y uw d iy" characteristic of the sounds of the spoken word "dwd". The constructed phoneme string model "d iy d ah b ix 1 y uw d iy" is cc.lllpal~d to ones of the lexicon vocabulary recorded 10 phonetic transcription strings and when there is a match of the constructed phoneme string model "d iy d ah b ix l y uw d iy" with the vocabulary lexicon database 1031 recorded phonetic transcription string "diydahbixlyuwdiy", the spoken word is recognized as the word "dwd" recorded with the matched phonetic transcription string "diydahbixlyuwdiy " .
In the method of operation, FIG. 1, a dictionary of standard words and a phonetic transcription of each standard word is initially stored in dictionarycl~t~b~e 1030. Similarly, subword model database 1032 initially stores phonemes,such as "d" and "w", each ide.l~irying a small unit of speech that serves to distinguish one utterance from another. To enter a new word, such as the word "dwd", into 20 speech recognizing system 1, the word "dwd" is typed into keyboard 11 and entered into keyword input program 1022, steps 102200, 102201, FIG. 4. Generate subword spelling of new word program 1020, FIG. 1, responds to the entry of alph~u~ f ~ic data and variations thereof, such as the word "dwd", by accessing dictionary ~l~t~b~ce 1030, steps 102000, 102001, FIG. 4. If the entered word is located in 25 dictionary ~t~b~e 1030, step 102002, generate subword spelling of new word program 1020 selects the word and associated phonetic transcription string, step 102003, and records both the word and associated phonetic transcription string in vocabulary lexicon d~t~b~e 1031, steps 102005, 102006.
When the entered word, for example word variation "dwd", is not found 30 in dictionary tl~t~b~e 1030, step 102002, generate subword spelling of new word program 1020 selects component words, FIGS 3, "d" and "w" from dictionary d~t~b~e 1030 and generates the phonetic transcription string "diydahbixlyuwdiy",step 102004. The word variation "dwd" and generated phonetic transcription string "diydahbixlyuwdiy" is recorded in vocabulary lexicon (l~t~b~ce 1031, steps 102005, 35 102006. If more words are to be entered into speech recognition system 1, step 102007, step 102201 and steps 102000 through 102007 are repeated until 20880~0 all the entered words have been recorded in vocabulary lexicon database 1031 at which time central processor unit 101 exits generate subword spelling of new word program 1020, step 102008.
Speech recognition system 1, FIG. 1, operating in accor~ ce with ASR
5 algoliLLlll program 1021, responds to a spoken word, such as word "dwd", received over telephone line 31, FIG. 5, steps 102100, 102101, 102102, by ~cGes~ing subword model database 1032. In response to the receipt of the spoken word "dwd", ASR
algc,liLl~ program 1021, step 102103, initi~tes a construction of a model comprising a string of phonemes representing the received spoken word "dwd". Construction of 10 the model is in accordance with the sound of the received "dwd" and the sound "d"
results in the selection of "d iy" for the first part of the string model, step 102104.
Since the model is not completed, step 102105, steps 102101 through 102105 are repeated such that the sound of "w" results in the phonemes "d ah b ix 1 y uw"
being selected from subword model d~t~b~ce 1032 and added to the model.
15 The phoneme string model is completed when the last sound "d" results in the phonemes "d iy" being added to the string to complete the phoneme string model "d iy d ah b ix 1 y uw d iy" representation of the spoken word "dwd", step 102105.
Upon the completion of the constructed phoneme string model "d iy d ah b ix 1 y uw d iy" represent~tion of the spoken word "dwd", ASR
20 algc~liLhlll program 1021 ~ccesses vocabulary lexicon d~t~b~e 1031 and COlll~ ,S
the constructed string phoneme model "d iy d ah b ix 1 y uw d iy" with ones of the vocabulary lexicon d~t~b~e recorded phonetic transcription strings, steps 102106, 102107, 102108. If there is a match of the constructed phoneme string model "d iy d ah b ix 1 y uw d iy" with the recorded phonetic transcription string 25 "diydahbixlyuwdiy", step 102109, the spoken word "dwd" is recognized as the word "dwd" recorded in vocabulary lexicon database 1031, step 102110. When there are additional spoken words, step 102111, steps 102101 through 102111 are repeated until all words have been recognized.
Should there be a failure to match, step 102109, and the match attempt 30 is less than a predetermined attempt nulllbel, step 102113, ASR alg~ hm program 1021 increases the repeat attempt number and plolllpls the caller to repeat the spoken word, step 102114. Steps 102101 through 102109 are then repeated.
Should the attempt number exceed the predete~ninçd attempt number, it is determined that an error has occurred, step 102115.
20~80~0 During the matching process, step 102109, a confidence recognition factor is assigned to the match of the constructed phoneme string model with phonetic transcription strings recorded in vocabulary lexicon rl~t~b~e 1031.
When the assigned confidence recognition factor exceeds a pre~leterrnined threshold 5 value, a detennin~ti-n is made that the spoken word associated with the constructed phoneme string model is recognized as the word recorded in vocabulary lexicon cl~t~b~e 1031 and associated with the matched phonetic transcription string.
Claims (12)
1. A speech recognition system comprising:
a vocabulary database that contains a plurality of words and a corresponding phoneme string for each of said words;
a subword database that stores a plurality of phonemes representing word sounds;
means for receiving a new non-spoken word in dictionary form as a series of characters, said new word not previously contained in said vocabulary database;
processor means ascertains a new phoneme string based on said new word and enters said new word and new phoneme string in said vocabulary databasethereby enabling a subsequent spoken entry of the new word to be recognized based on the presence of the new word in the vocabulary database.
a vocabulary database that contains a plurality of words and a corresponding phoneme string for each of said words;
a subword database that stores a plurality of phonemes representing word sounds;
means for receiving a new non-spoken word in dictionary form as a series of characters, said new word not previously contained in said vocabulary database;
processor means ascertains a new phoneme string based on said new word and enters said new word and new phoneme string in said vocabulary databasethereby enabling a subsequent spoken entry of the new word to be recognized based on the presence of the new word in the vocabulary database.
2. The system according to claim 1 wherein said processor means comprises a dictionary database that contains a plurality of words and a corresponding phoneme string for each of said words, said processor means selecting a phoneme string stored in said dictionary database corresponding to said new non-spoken word as the new phoneme string that is stored with said new word in the vocabulary database.
3. The system according to claim 2 wherein said processor means ascertains said new phoneme string without requiring the concurrent entry of the new word in spoken form.
4. The system according to claim 1 wherein said processor means ascertain said new phoneme strain without requiring the concurrent entry of the new word in spoken form.
5. A speech recognition system comprising:
a vocabulary database that contains a plurality of words and a corresponding phoneme string for each of said words;
a subword database that stores a plurality of phonemes representing word sounds;
Processor means for determining if a first spoken word received as an input is one of said words contained on said vocabulary database, said processor means generating a first phoneme string based on phonemes stored in said subword database to represent said first and determining which, if any, of said phoneme strings contained in said vocabulary database matches said first phoneme string;
means for adding sets of new words and corresponding phoneme strings in said vocabulary database without requiring the concurrent spoken entry of such new words, said adding means comprising means for receiving a new non-spoken word indictionary form as a series of characters, said new word not previously contained in said vocabulary database, and said processor means ascertaining a new phoneme string based on said new word and entering said new word and new phoneme string in saidvocabulary database thereby enabling a subsequent spoken entry of the new word in the vocabulary database.
a vocabulary database that contains a plurality of words and a corresponding phoneme string for each of said words;
a subword database that stores a plurality of phonemes representing word sounds;
Processor means for determining if a first spoken word received as an input is one of said words contained on said vocabulary database, said processor means generating a first phoneme string based on phonemes stored in said subword database to represent said first and determining which, if any, of said phoneme strings contained in said vocabulary database matches said first phoneme string;
means for adding sets of new words and corresponding phoneme strings in said vocabulary database without requiring the concurrent spoken entry of such new words, said adding means comprising means for receiving a new non-spoken word indictionary form as a series of characters, said new word not previously contained in said vocabulary database, and said processor means ascertaining a new phoneme string based on said new word and entering said new word and new phoneme string in saidvocabulary database thereby enabling a subsequent spoken entry of the new word in the vocabulary database.
6. The system according to claim 5 wherein said processor means comprises a dictionary database that contains a plurality of words and a corresponding phoneme string for each of said words, said processor means selecting a phoneme string stored in said dictionary database corresponding to said new non-spoken word as the newphoneme string that is stored with said new word in the vocabulary database.
7. A method for recognizing spoken words comprising the steps of:
generating a vocabulary database that contains a plurality of words and a corresponding phoneme string for each of said words;
generating a subword database that stores a plurality of phonemes representing word sounds;
determining by use of a processor means if a first spoken word received as an input is one of said words contained in said vocabulary database, said determining step including generating a first phoneme string based on phonemes stored in said subword database to represent said first words and determining which, if any, of said phoneme strings contained in said vocabulary database matches said first string;
receiving a new non-spoken word in dictionary form as a series of characters, said new word not previously contained in said vocabulary database;
ascertaining by use of said processor means a new phoneme string based on said new word and entering said new word and new phoneme string in said vocabulary database so as to enable a subsequent spoken entry of the new word to be recognized based on the presence of the new word in the vocabulary database.
generating a vocabulary database that contains a plurality of words and a corresponding phoneme string for each of said words;
generating a subword database that stores a plurality of phonemes representing word sounds;
determining by use of a processor means if a first spoken word received as an input is one of said words contained in said vocabulary database, said determining step including generating a first phoneme string based on phonemes stored in said subword database to represent said first words and determining which, if any, of said phoneme strings contained in said vocabulary database matches said first string;
receiving a new non-spoken word in dictionary form as a series of characters, said new word not previously contained in said vocabulary database;
ascertaining by use of said processor means a new phoneme string based on said new word and entering said new word and new phoneme string in said vocabulary database so as to enable a subsequent spoken entry of the new word to be recognized based on the presence of the new word in the vocabulary database.
8. The method according to claim 7 further comprising the steps of generating a dictionary database that contains a plurality of words and a corresponding phoneme string for each of said words, and selecting a phoneme string stored in said dictionary database corresponding to said new non-spoken word as the new phonemestring that is stored with the new word in the vocabulary database.
9. The method according to claim 8 wherein said new phoneme string is ascertained without requiring the concurrent entry of the new word in spoken form.
10. The method according to claim 7 wherein said new phoneme string is ascertained without requiring the concurrent entry of the new word in spoken form.
11. A speech recognition method comprising the steps of:
generating a vocabulary database that contains a plurality of words and a corresponding phoneme string for each of said words:
generating a subword database that stores a plurality of phonemes representing word sounds;
adding sets of new words and corresponding phonemes in said vocabulary database without requiring the concurrent spoken entry of such new words, said adding step including the steps of receiving a new non-spoken word in dictionary form as a series of characters, said new word not contained in said vocabulary database, and ascertaining by use of processor means a new phoneme string based on said new word, and entering said new word and new phoneme string in said vocabulary database toenable a subsequent spoken entry of the new word to be recognized based on the presence of the new word in the vocabulary database.
generating a vocabulary database that contains a plurality of words and a corresponding phoneme string for each of said words:
generating a subword database that stores a plurality of phonemes representing word sounds;
adding sets of new words and corresponding phonemes in said vocabulary database without requiring the concurrent spoken entry of such new words, said adding step including the steps of receiving a new non-spoken word in dictionary form as a series of characters, said new word not contained in said vocabulary database, and ascertaining by use of processor means a new phoneme string based on said new word, and entering said new word and new phoneme string in said vocabulary database toenable a subsequent spoken entry of the new word to be recognized based on the presence of the new word in the vocabulary database.
12. The method according to claim 11 further comprising the steps of generating a dictionary database that contains a plurality of words and a corresponding phoneme string for each of said words, and selecting a phoneme string stored in said dictionary database corresponding to said new non-spoken word as the new phonemestring.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US86239492A | 1992-04-02 | 1992-04-02 | |
US862,394 | 1992-04-02 |
Publications (2)
Publication Number | Publication Date |
---|---|
CA2088080A1 CA2088080A1 (en) | 1993-10-03 |
CA2088080C true CA2088080C (en) | 1997-10-07 |
Family
ID=25338399
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CA002088080A Expired - Fee Related CA2088080C (en) | 1992-04-02 | 1993-01-26 | Automatic speech recognizer |
Country Status (9)
Country | Link |
---|---|
US (1) | US5329608A (en) |
EP (1) | EP0564166B1 (en) |
JP (1) | JPH0612092A (en) |
KR (1) | KR930022267A (en) |
CA (1) | CA2088080C (en) |
DE (1) | DE69327188T2 (en) |
ES (1) | ES2142332T3 (en) |
FI (1) | FI931471A (en) |
HK (1) | HK1004497A1 (en) |
Families Citing this family (196)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6330426B2 (en) | 1994-05-23 | 2001-12-11 | Stephen J. Brown | System and method for remote education using a memory card |
US6968375B1 (en) | 1997-03-28 | 2005-11-22 | Health Hero Network, Inc. | Networked system for interactive communication and remote monitoring of individuals |
US20030212579A1 (en) * | 2002-05-08 | 2003-11-13 | Brown Stephen J. | Remote health management system |
US9215979B2 (en) | 1992-11-17 | 2015-12-22 | Robert Bosch Healthcare Systems, Inc. | Multi-user remote health monitoring system |
US20010011224A1 (en) | 1995-06-07 | 2001-08-02 | Stephen James Brown | Modular microprocessor-based health monitoring system |
US8078407B1 (en) | 1997-03-28 | 2011-12-13 | Health Hero Network, Inc. | System and method for identifying disease-influencing genes |
US6196970B1 (en) | 1999-03-22 | 2001-03-06 | Stephen J. Brown | Research data collection and analysis |
US8626521B2 (en) * | 1997-11-21 | 2014-01-07 | Robert Bosch Healthcare Systems, Inc. | Public health surveillance system |
US7970620B2 (en) * | 1992-11-17 | 2011-06-28 | Health Hero Network, Inc. | Multi-user remote health monitoring system with biometrics support |
US5832448A (en) | 1996-10-16 | 1998-11-03 | Health Hero Network | Multiple patient monitoring system for proactive health management |
WO2001037174A1 (en) | 1992-11-17 | 2001-05-25 | Health Hero Network, Inc. | Method and system for improving adherence with a diet program or other medical regimen |
US5307263A (en) | 1992-11-17 | 1994-04-26 | Raya Systems, Inc. | Modular microprocessor-based health monitoring system |
US5956501A (en) | 1997-01-10 | 1999-09-21 | Health Hero Network, Inc. | Disease simulation system and method |
US8095340B2 (en) | 1992-11-17 | 2012-01-10 | Health Hero Network, Inc. | Home power management system |
US5951300A (en) | 1997-03-10 | 1999-09-14 | Health Hero Network | Online system and method for providing composite entertainment and health information |
US8027809B2 (en) | 1992-11-17 | 2011-09-27 | Health Hero Network, Inc. | Home power management system |
US7624028B1 (en) | 1992-11-17 | 2009-11-24 | Health Hero Network, Inc. | Remote health monitoring and maintenance system |
US8078431B2 (en) | 1992-11-17 | 2011-12-13 | Health Hero Network, Inc. | Home power management system |
EP0692135B1 (en) * | 1993-03-12 | 2000-08-16 | Sri International | Method and apparatus for voice-interactive language instruction |
US7631343B1 (en) | 1993-03-24 | 2009-12-08 | Endgate LLC | Down-line transcription system using automatic tracking and revenue collection |
US5369704A (en) * | 1993-03-24 | 1994-11-29 | Engate Incorporated | Down-line transcription system for manipulating real-time testimony |
US7249026B1 (en) * | 1993-03-24 | 2007-07-24 | Engate Llc | Attorney terminal having outline preparation capabilities for managing trial proceedings |
US5623578A (en) * | 1993-10-28 | 1997-04-22 | Lucent Technologies Inc. | Speech recognition system allows new vocabulary words to be added without requiring spoken samples of the words |
US5970454A (en) * | 1993-12-16 | 1999-10-19 | British Telecommunications Public Limited Company | Synthesizing speech by converting phonemes to digital waveforms |
US8015033B2 (en) | 1994-04-26 | 2011-09-06 | Health Hero Network, Inc. | Treatment regimen compliance and efficacy with feedback |
US5737723A (en) * | 1994-08-29 | 1998-04-07 | Lucent Technologies Inc. | Confusable word detection in speech recognition |
GB2307582A (en) * | 1994-09-07 | 1997-05-28 | Motorola Inc | System for recognizing spoken sounds from continuous speech and method of using same |
US5594834A (en) * | 1994-09-30 | 1997-01-14 | Motorola, Inc. | Method and system for recognizing a boundary between sounds in continuous speech |
US5596679A (en) * | 1994-10-26 | 1997-01-21 | Motorola, Inc. | Method and system for identifying spoken sounds in continuous speech by comparing classifier outputs |
US5638486A (en) * | 1994-10-26 | 1997-06-10 | Motorola, Inc. | Method and system for continuous speech recognition using voting techniques |
ES2143079T3 (en) * | 1994-11-01 | 2000-05-01 | British Telecomm | SPEECH RECOGNITION. |
US5680510A (en) * | 1995-01-26 | 1997-10-21 | Apple Computer, Inc. | System and method for generating and using context dependent sub-syllable models to recognize a tonal language |
US5724481A (en) * | 1995-03-30 | 1998-03-03 | Lucent Technologies Inc. | Method for automatic speech recognition of arbitrary spoken words |
US5822727A (en) * | 1995-03-30 | 1998-10-13 | At&T Corp | Method for automatic speech recognition in telephony |
US5675706A (en) * | 1995-03-31 | 1997-10-07 | Lucent Technologies Inc. | Vocabulary independent discriminative utterance verification for non-keyword rejection in subword based speech recognition |
JP3741156B2 (en) * | 1995-04-07 | 2006-02-01 | ソニー株式会社 | Speech recognition apparatus, speech recognition method, and speech translation apparatus |
US5745875A (en) * | 1995-04-14 | 1998-04-28 | Stenovations, Inc. | Stenographic translation system automatic speech recognition |
US5852801A (en) * | 1995-10-04 | 1998-12-22 | Apple Computer, Inc. | Method and apparatus for automatically invoking a new word module for unrecognized user input |
US5832439A (en) * | 1995-12-14 | 1998-11-03 | U S West, Inc. | Method and system for linguistic command processing in a video server network |
GB9601925D0 (en) * | 1996-01-31 | 1996-04-03 | British Telecomm | Database access |
US6804645B1 (en) * | 1996-04-02 | 2004-10-12 | Siemens Aktiengesellschaft | Dynamic phoneme dictionary for speech recognition |
US5758024A (en) * | 1996-06-25 | 1998-05-26 | Microsoft Corporation | Method and system for encoding pronunciation prefix trees |
US5752230A (en) * | 1996-08-20 | 1998-05-12 | Ncr Corporation | Method and apparatus for identifying names with a speech recognition program |
US6018568A (en) * | 1996-09-25 | 2000-01-25 | At&T Corp. | Voice dialing system |
WO1998014934A1 (en) * | 1996-10-02 | 1998-04-09 | Sri International | Method and system for automatic text-independent grading of pronunciation for language instruction |
US5991712A (en) * | 1996-12-05 | 1999-11-23 | Sun Microsystems, Inc. | Method, apparatus, and product for automatic generation of lexical features for speech recognition systems |
US6151586A (en) * | 1996-12-23 | 2000-11-21 | Health Hero Network, Inc. | Computerized reward system for encouraging participation in a health management program |
US7584108B2 (en) * | 1996-12-23 | 2009-09-01 | Health Hero Network, Inc. | Network media access control system for encouraging patient compliance with a treatment plan |
US6032119A (en) | 1997-01-16 | 2000-02-29 | Health Hero Network, Inc. | Personalized display of health information |
JP3373755B2 (en) | 1997-04-09 | 2003-02-04 | 株式会社鷹山 | Complex despreading processor |
WO1999000758A1 (en) * | 1997-06-26 | 1999-01-07 | Charles Schwab & Co., Inc. | System and method for automatically providing financial services to a user using speech signals |
US6006181A (en) * | 1997-09-12 | 1999-12-21 | Lucent Technologies Inc. | Method and apparatus for continuous speech recognition using a layered, self-adjusting decoder network |
US6404876B1 (en) | 1997-09-25 | 2002-06-11 | Gte Intelligent Network Services Incorporated | System and method for voice activated dialing and routing under open access network control |
US6108627A (en) * | 1997-10-31 | 2000-08-22 | Nortel Networks Corporation | Automatic transcription tool |
US5983177A (en) * | 1997-12-18 | 1999-11-09 | Nortel Networks Corporation | Method and apparatus for obtaining transcriptions from multiple training utterances |
US6243678B1 (en) * | 1998-04-07 | 2001-06-05 | Lucent Technologies Inc. | Method and system for dynamic speech recognition using free-phone scoring |
US6233553B1 (en) * | 1998-09-04 | 2001-05-15 | Matsushita Electric Industrial Co., Ltd. | Method and system for automatically determining phonetic transcriptions associated with spelled words |
US20080004915A1 (en) | 1998-09-25 | 2008-01-03 | Brown Stephen J | Dynamic modeling and scoring risk assessment |
FR2786600B1 (en) * | 1998-11-16 | 2001-04-20 | France Telecom | METHOD FOR SEARCHING BY CONTENT OF TEXTUAL DOCUMENTS USING VOICE RECOGNITION |
JP2002532763A (en) * | 1998-12-17 | 2002-10-02 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | Automatic inquiry system operated by voice |
EP1060471A1 (en) * | 1999-01-05 | 2000-12-20 | Koninklijke Philips Electronics N.V. | Speech recognition device including a sub-word memory |
US20080201168A1 (en) * | 1999-05-03 | 2008-08-21 | Brown Stephen J | Treatment regimen compliance and efficacy with feedback |
US7797730B2 (en) * | 1999-06-24 | 2010-09-14 | Engate Llc | Downline transcription system using automatic tracking and revenue collection |
JP3476008B2 (en) | 1999-09-10 | 2003-12-10 | インターナショナル・ビジネス・マシーンズ・コーポレーション | A method for registering voice information, a method for specifying a recognition character string, a voice recognition device, a storage medium storing a software product for registering voice information, and a software product for specifying a recognition character string are stored. Storage media |
US6442520B1 (en) | 1999-11-08 | 2002-08-27 | Agere Systems Guardian Corp. | Method and apparatus for continuous speech recognition using a layered, self-adjusting decoded network |
JP2001249684A (en) * | 2000-03-02 | 2001-09-14 | Sony Corp | Device and method for recognizing speech, and recording medium |
US8645137B2 (en) | 2000-03-16 | 2014-02-04 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US7047196B2 (en) * | 2000-06-08 | 2006-05-16 | Agiletv Corporation | System and method of voice recognition near a wireline node of a network supporting cable television and/or video delivery |
JP2002123283A (en) * | 2000-10-12 | 2002-04-26 | Nissan Motor Co Ltd | Voice recognition operating device |
US8095370B2 (en) | 2001-02-16 | 2012-01-10 | Agiletv Corporation | Dual compression voice recordation non-repudiation system |
US20030192688A1 (en) * | 2002-04-10 | 2003-10-16 | Thomson Michael A. | Tubing saver rotator and method for using same |
AU2003280474A1 (en) * | 2002-06-28 | 2004-01-19 | Conceptual Speech, Llc | Multi-phoneme streamer and knowledge representation speech recognition system and method |
DE10253786B4 (en) * | 2002-11-19 | 2009-08-06 | Anwaltssozietät BOEHMERT & BOEHMERT GbR (vertretungsberechtigter Gesellschafter: Dr. Carl-Richard Haarmann, 28209 Bremen) | Method for the computer-aided determination of a similarity of an electronically registered first identifier to at least one electronically detected second identifier as well as apparatus and computer program for carrying out the same |
US7399276B1 (en) | 2003-05-08 | 2008-07-15 | Health Hero Network, Inc. | Remote health monitoring system |
US8954325B1 (en) * | 2004-03-22 | 2015-02-10 | Rockstar Consortium Us Lp | Speech recognition in automated information services systems |
US7831549B2 (en) * | 2004-09-17 | 2010-11-09 | Nokia Corporation | Optimization of text-based training set selection for language processing modules |
KR100679042B1 (en) * | 2004-10-27 | 2007-02-06 | 삼성전자주식회사 | Method and apparatus for speech recognition, and navigation system using for the same |
US8677377B2 (en) | 2005-09-08 | 2014-03-18 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US7925677B2 (en) * | 2006-03-16 | 2011-04-12 | Tele Atlas North America, Inc. | Geographic feature name reduction using phonetic algorithms |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US8977255B2 (en) | 2007-04-03 | 2015-03-10 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US8595642B1 (en) | 2007-10-04 | 2013-11-26 | Great Northern Research, LLC | Multiple shell multi faceted graphical user interface |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US8996376B2 (en) | 2008-04-05 | 2015-03-31 | Apple Inc. | Intelligent text-to-speech conversion |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US20100030549A1 (en) | 2008-07-31 | 2010-02-04 | Lee Michael M | Mobile device having human language translation capability with positional feedback |
US9959870B2 (en) | 2008-12-11 | 2018-05-01 | Apple Inc. | Speech recognition involving a mobile device |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US20120311585A1 (en) | 2011-06-03 | 2012-12-06 | Apple Inc. | Organizing task items that represent tasks to perform |
US9431006B2 (en) | 2009-07-02 | 2016-08-30 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US8682667B2 (en) | 2010-02-25 | 2014-03-25 | Apple Inc. | User profiling for selecting user specific voice input processing information |
US10762293B2 (en) | 2010-12-22 | 2020-09-01 | Apple Inc. | Using parts-of-speech tagging and named entity recognition for spelling correction |
US8744856B1 (en) | 2011-02-22 | 2014-06-03 | Carnegie Speech Company | Computer implemented system and method and computer program product for evaluating pronunciation of phonemes in a language |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
GB2513821A (en) * | 2011-06-28 | 2014-11-12 | Andrew Levine | Speech-to-text conversion |
US8781825B2 (en) * | 2011-08-24 | 2014-07-15 | Sensory, Incorporated | Reducing false positives in speech recognition systems |
US8994660B2 (en) | 2011-08-29 | 2015-03-31 | Apple Inc. | Text correction processing |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
US9280610B2 (en) | 2012-05-14 | 2016-03-08 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US9721563B2 (en) | 2012-06-08 | 2017-08-01 | Apple Inc. | Name recognition system |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9576574B2 (en) | 2012-09-10 | 2017-02-21 | Apple Inc. | Context-sensitive handling of interruptions by intelligent digital assistant |
US9547647B2 (en) | 2012-09-19 | 2017-01-17 | Apple Inc. | Voice-based media searching |
BR112015018905B1 (en) | 2013-02-07 | 2022-02-22 | Apple Inc | Voice activation feature operation method, computer readable storage media and electronic device |
US9368114B2 (en) | 2013-03-14 | 2016-06-14 | Apple Inc. | Context-sensitive handling of interruptions |
WO2014144579A1 (en) | 2013-03-15 | 2014-09-18 | Apple Inc. | System and method for updating an adaptive speech recognition model |
KR101759009B1 (en) | 2013-03-15 | 2017-07-17 | 애플 인크. | Training an at least partial voice command system |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
WO2014197334A2 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
WO2014197336A1 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
WO2014197335A1 (en) | 2013-06-08 | 2014-12-11 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
CN105264524B (en) | 2013-06-09 | 2019-08-02 | 苹果公司 | For realizing the equipment, method and graphic user interface of the session continuity of two or more examples across digital assistants |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
CN105265005B (en) | 2013-06-13 | 2019-09-17 | 苹果公司 | System and method for the urgent call initiated by voice command |
JP6163266B2 (en) | 2013-08-06 | 2017-07-12 | アップル インコーポレイテッド | Automatic activation of smart responses based on activation from remote devices |
US8768712B1 (en) * | 2013-12-04 | 2014-07-01 | Google Inc. | Initiating actions based on partial hotwords |
US9953632B2 (en) * | 2014-04-17 | 2018-04-24 | Qualcomm Incorporated | Keyword model generation for detecting user-defined keyword |
US9620105B2 (en) | 2014-05-15 | 2017-04-11 | Apple Inc. | Analyzing audio input for efficient speech and music recognition |
US10592095B2 (en) | 2014-05-23 | 2020-03-17 | Apple Inc. | Instantaneous speaking of content on touch devices |
US9502031B2 (en) | 2014-05-27 | 2016-11-22 | Apple Inc. | Method for supporting dynamic grammars in WFST-based ASR |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US9734193B2 (en) | 2014-05-30 | 2017-08-15 | Apple Inc. | Determining domain salience ranking from ambiguous words in natural speech |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US10289433B2 (en) | 2014-05-30 | 2019-05-14 | Apple Inc. | Domain specific language for encoding assistant dialog |
EP3149728B1 (en) | 2014-05-30 | 2019-01-16 | Apple Inc. | Multi-command single utterance input method |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US9606986B2 (en) | 2014-09-29 | 2017-03-28 | Apple Inc. | Integrated word N-gram and class M-gram language models |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US9711141B2 (en) | 2014-12-09 | 2017-07-18 | Apple Inc. | Disambiguating heteronyms in speech synthesis |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
DK179588B1 (en) | 2016-06-09 | 2019-02-22 | Apple Inc. | Intelligent automated assistant in a home environment |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10586535B2 (en) | 2016-06-10 | 2020-03-10 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
DK179415B1 (en) | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
DK179049B1 (en) | 2016-06-11 | 2017-09-18 | Apple Inc | Data driven natural language event detection and classification |
DK179343B1 (en) | 2016-06-11 | 2018-05-14 | Apple Inc | Intelligent task discovery |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
DK179745B1 (en) | 2017-05-12 | 2019-05-01 | Apple Inc. | SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT |
DK201770431A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
KR102075796B1 (en) * | 2017-11-14 | 2020-03-02 | 주식회사 엘솔루 | Apparatus and method for recognizing speech automatically |
WO2020046269A1 (en) | 2018-08-27 | 2020-03-05 | Google Llc | Algorithmic determination of a story readers discontinuation of reading |
CN112805779A (en) * | 2018-09-04 | 2021-05-14 | 谷歌有限责任公司 | Reading progress estimation based on speech fuzzy matching and confidence interval |
WO2020050822A1 (en) | 2018-09-04 | 2020-03-12 | Google Llc | Detection of story reader progress for pre-caching special effects |
US11328712B2 (en) * | 2019-08-02 | 2022-05-10 | International Business Machines Corporation | Domain specific correction of output from automatic speech recognition |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4277644A (en) * | 1979-07-16 | 1981-07-07 | Bell Telephone Laboratories, Incorporated | Syntactic continuous speech recognizer |
USRE32012E (en) * | 1980-06-09 | 1985-10-22 | At&T Bell Laboratories | Spoken word controlled automatic dialer |
USRE33597E (en) * | 1982-10-15 | 1991-05-28 | Hidden Markov model speech recognition arrangement | |
GB8412424D0 (en) * | 1983-10-26 | 1984-06-20 | Marconi Co Ltd | Speech responsive apparatus |
US4980918A (en) * | 1985-05-09 | 1990-12-25 | International Business Machines Corporation | Speech recognition system with efficient storage and rapid assembly of phonological graphs |
US4748670A (en) * | 1985-05-29 | 1988-05-31 | International Business Machines Corporation | Apparatus and method for determining a likely word sequence from labels generated by an acoustic processor |
US4757541A (en) * | 1985-11-05 | 1988-07-12 | Research Triangle Institute | Audio visual speech recognition |
US4882759A (en) * | 1986-04-18 | 1989-11-21 | International Business Machines Corporation | Synthesizing word baseforms used in speech recognition |
JP2815579B2 (en) * | 1987-03-10 | 1998-10-27 | 富士通株式会社 | Word candidate reduction device in speech recognition |
US5072452A (en) * | 1987-10-30 | 1991-12-10 | International Business Machines Corporation | Automatic determination of labels and Markov word models in a speech recognition system |
JP2955297B2 (en) * | 1988-05-27 | 1999-10-04 | 株式会社東芝 | Speech recognition system |
US5027406A (en) * | 1988-12-06 | 1991-06-25 | Dragon Systems, Inc. | Method for interactive speech recognition and training |
US5033087A (en) * | 1989-03-14 | 1991-07-16 | International Business Machines Corp. | Method and apparatus for the automatic determination of phonological rules as for a continuous speech recognition system |
DE4022511A1 (en) * | 1990-07-14 | 1992-01-16 | Grundig Emv | VOICE CONTROLLED DEVICE FOR ENTERTAINMENT ELECTRONICS, IN PARTICULAR VIDEO RECORDER |
-
1993
- 1993-01-26 CA CA002088080A patent/CA2088080C/en not_active Expired - Fee Related
- 1993-03-25 EP EP93302286A patent/EP0564166B1/en not_active Expired - Lifetime
- 1993-03-25 DE DE69327188T patent/DE69327188T2/en not_active Expired - Lifetime
- 1993-03-25 ES ES93302286T patent/ES2142332T3/en not_active Expired - Lifetime
- 1993-03-26 KR KR1019921004783A patent/KR930022267A/en unknown
- 1993-04-01 FI FI931471A patent/FI931471A/en unknown
- 1993-04-02 JP JP5098442A patent/JPH0612092A/en active Pending
- 1993-08-18 US US08/108,839 patent/US5329608A/en not_active Expired - Lifetime
-
1998
- 1998-05-05 HK HK98103837A patent/HK1004497A1/en not_active IP Right Cessation
Also Published As
Publication number | Publication date |
---|---|
FI931471A (en) | 1993-10-03 |
KR930022267A (en) | 1993-11-23 |
JPH0612092A (en) | 1994-01-21 |
EP0564166A3 (en) | 1995-03-29 |
FI931471A0 (en) | 1993-04-01 |
US5329608A (en) | 1994-07-12 |
DE69327188T2 (en) | 2004-09-02 |
HK1004497A1 (en) | 1998-11-27 |
ES2142332T3 (en) | 2000-04-16 |
DE69327188D1 (en) | 2000-01-13 |
EP0564166B1 (en) | 1999-12-08 |
CA2088080A1 (en) | 1993-10-03 |
EP0564166A2 (en) | 1993-10-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CA2088080C (en) | Automatic speech recognizer | |
US5832063A (en) | Methods and apparatus for performing speaker independent recognition of commands in parallel with speaker dependent recognition of names, words or phrases | |
US5737723A (en) | Confusable word detection in speech recognition | |
US7603279B2 (en) | Grammar update system and method for speech recognition | |
US6269335B1 (en) | Apparatus and methods for identifying homophones among words in a speech recognition system | |
US6915246B2 (en) | Employing speech recognition and capturing customer speech to improve customer service | |
US6073101A (en) | Text independent speaker recognition for transparent command ambiguity resolution and continuous access control | |
US6219407B1 (en) | Apparatus and method for improved digit recognition and caller identification in telephone mail messaging | |
CA2508946C (en) | Method and apparatus for natural language call routing using confidence scores | |
US5802251A (en) | Method and system for reducing perplexity in speech recognition via caller identification | |
US5430827A (en) | Password verification system | |
CN109493850A (en) | Growing Interface | |
US20030115056A1 (en) | Employing speech recognition and key words to improve customer service | |
US9286887B2 (en) | Concise dynamic grammars using N-best selection | |
JP3703991B2 (en) | Method and apparatus for dynamic speech recognition using free speech scoring method | |
US20030220793A1 (en) | Interactive system and method of controlling same | |
US6260014B1 (en) | Specific task composite acoustic models | |
CN109003612B (en) | Voice question-answer verification system and method based on artificial intelligence | |
EP1397797B1 (en) | Speech recognition | |
JPH08116385A (en) | Individual information terminal equipment and voice response system | |
JPH10116307A (en) | Telephone transaction support system and recording medium storing program for making copmuter execute processing in its support system | |
AU657064B2 (en) | Speech recognition system | |
EP1158491A2 (en) | Personal data spoken input and retrieval | |
EP0533338A2 (en) | Interface method and apparatus for information services | |
JPS60118892A (en) | Voice recognition equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
EEER | Examination request | ||
MKLA | Lapsed |