|Publication number||US7970600 B2|
|Application number||US 10/980,646|
|Publication date||Jun 28, 2011|
|Filing date||Nov 3, 2004|
|Priority date||Nov 3, 2004|
|Also published as||US20060095250|
|Publication number||10980646, 980646, US 7970600 B2, US 7970600B2, US-B2-7970600, US7970600 B2, US7970600B2|
|Inventors||John T. Chen, Ming Zhou, Tianlei Wu|
|Original Assignee||Microsoft Corporation|
|Export Citation||BiBTeX, EndNote, RefMan|
|Patent Citations (55), Non-Patent Citations (30), Referenced by (10), Classifications (8), Legal Events (3)|
|External Links: USPTO, USPTO Assignment, Espacenet|
The present invention generally relates to natural language processing. More particularly, the present invention relates to developing a parser for use in natural language processing.
Parsing assigns a structural description, typically in the form of a parse tree or a dependency tree, to a given sentence. A parsed sentence can be used in a variety of natural language processing applications including information retrieval, writing assistance, machine translation, questioning/answering, searching, and the like. Current parsers are trained using statistical or machine learning approaches on hand-annotated corpora. Hand-annotation of corpora requires a large amount of time and expense. Since hand-annotated corpora are expensive to generate, developing a parser based on automatic generation of training data would provide a parser that is less costly to develop and/or maintain.
A computer-implemented method for developing a parser is provided. The method includes accessing a corpus of sentences and parsing the sentences using a pre-existing parser to generate a structural description of each sentence. Subsequently, there can be a deterministic method which does not rely on any human input or labor to transform the resulting structural descriptions into training data. This training data is used directly to train the parser.
The present invention deals with developing a parser for natural language processing. More specifically, the present invention deals with utilizing output from a first parser to train a second parser. However, prior to discussing the present invention in greater detail, one exemplary environment in which the present invention can be used will be discussed.
The invention is operational with numerous other general purpose or special purpose computing system environments or configurations. Examples of well known computing systems, environments, and/or configurations that may be suitable for use with the invention include, but are not limited to, personal computers, server computers, hand-held or laptop devices, multiprocessor systems, microprocessor-based systems, set top boxes, programmable consumer electronics, network PCs, minicomputers, mainframe computers, distributed computing environments that include any of the above systems or devices, and the like.
The invention may be described in the general context of computer-executable instructions, such as program modules, being executed by a computer. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types. The invention may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in both local and remote computer storage media including memory storage devices.
With reference to
Computer 110 typically includes a variety of computer readable media. Computer readable media can be any available media that can be accessed by computer 110 and includes both volatile and nonvolatile media, removable and non-removable media. By way of example, and not limitation, computer readable media may comprise computer storage media and communication media. Computer storage media includes both volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by computer 100. Communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier WAV or other transport mechanism and includes any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, FR, infrared and other wireless media. Combinations of any of the above should also be included within the scope of computer readable media.
The system memory 130 includes computer storage media in the form of volatile and/or nonvolatile memory such as read only memory (ROM) 131 and random access memory (RAM) 132. A basic input/output system 133 (BIOS), containing the basic routines that help to transfer information between elements within computer 110, such as during start-up, is typically stored in ROM 131. RAM 132 typically contains data and/or program modules that are immediately accessible to and/or presently being operated on by processing unit 120. By way of example, and not limitation,
The computer 110 may also include other removable/non-removable volatile/nonvolatile computer storage media. By way of example only,
The drives and their associated computer storage media discussed above and illustrated in
A user may enter commands and information into the computer 110 through input devices such as a keyboard 162, a microphone 163, and a pointing device 161, such as a mouse, trackball or touch pad. Other input devices (not shown) may include a joystick, game pad, satellite dish, scanner, or the like. These and other input devices are often connected to the processing unit 120 through a user input interface 160 that is coupled to the system bus, but may be connected by other interface and bus structures, such as a parallel port, game port or a universal serial bus (USB). A monitor 191 or other type of display device is also connected to the system bus 121 via an interface, such as a video interface 190. In addition to the monitor, computers may also include other peripheral output devices such as speakers 197 and printer 196, which may be connected through an output peripheral interface 190.
The computer 110 may operate in a networked environment using logical connections to one or more remote computers, such as a remote computer 180. The remote computer 180 may be a personal computer, a hand-held device, a server, a router, a network PC, a peer device or other common network node, and typically includes many or all of the elements described above relative to the computer 110. The logical connections depicted in
When used in a LAN networking environment, the computer 110 is connected to the LAN 171 through a network interface or adapter 170. When used in a WAN networking environment, the computer 110 typically includes a modem 172 or other means for establishing communications over the WAN 173, such as the Internet. The modem 172, which may be internal or external, may be connected to the system bus 121 via the user-input interface 160, or other appropriate mechanism. In a networked environment, program modules depicted relative to the computer 110, or portions thereof, may be stored in the remote memory storage device. By way of example, and not limitation,
It should be noted that the present invention can be carried out on a computer system such as that described with respect to
There are several different options for developing a parser. One option is to develop a parser as a rule-based or statistical parser. A rule-based parser is typically based on a human generated grammar. A statistical parser is typically based on a human annotation of a corpus of parse trees. Unlike rule-based parsers, statistical parsers have been shown to be more accurate when parsing sentences from complex domains such as newspaper text. However, the statistical parser generally provides less detail about the corresponding sentence structure. Another option is to develop a full or shallow parser. A full parser explores all parse trees while a shallow parser explores a subset of parse trees. The shallow parser is faster, but the output is less detailed.
In accordance with the present invention, a shallow statistical parser is developed and trained using parsed data that is output from a pre-existing parser. This parser in itself requires less human labor than the more prevalent alternative of having a human hand-parse the data. The shallow parser in question can be referred to as a “skeleton” parser, because output from the parser is restricted to being either subject-verb or verb-object relationships. We define “skeleton relations” as these kinds of relationships. They are valuable in determining importance of words and subject matter of a sentence. Furthermore, limiting the skeleton parser to output only skeleton relations makes it more accurate, because it is trained directly on the output of a pre-existing parser, and pre-existing parsers generally output skeleton relations with higher accuracy than other kinds of relations.
EZParse is a hand-crafted parser with a statistical component. Hand-crafted components include an augmented context-free grammar and a knowledge base with morphological, lexicographic, and semantic information. The statistical component guides a search algorithm that is used by the parser. The input to EZParse is parsed in a cascaded manner with morphological, syntactic chunk, surface syntactic, and semantic information.
Using the parsed data, data is transformed at step 256. Transformation is further described with respect to
Next, at step 286, verb groups are extracted from the portrait output. A verb group includes a head word of a clause and any pre-modifiers. For example, preceding auxiliary verbs and adverbs are included in verb groups. Detecting these verb groups can rely on part-of-speech information as well as a hand-coded list of auxiliary verbs.
At step 288, skeleton relations from the logical form are extracted. The skeleton relation is a word pair consisting of a predicate head word and its corresponding argument head word, and also an associated label that is the relation between the pair. Some relations are extracted verbatim from the EZParse logical form output. Others can be transformed into relations that are defined as skeleton relations. One transformation includes replacing a direct link between a subject and an object in a predicative constructions with two links, wherein one links the subject and the other links the object with a verb. For example,
At step 306, candidate skeleton word pairs are selected based on the parts of speech and the chunking of base noun phrases. Verbs in a sentence outside base noun phrases can be candidate predicate head words. Nouns and adjectives can be candidate argument head words, excluding non-headwords inside base noun phrases. Candidate predicate argument word pairs can also be subject to a restriction that a verb cannot occur between the predicate and argument.
At step 308, a determination is made as to whether each candidate word pair is a skeleton relation. This determination can be made using a so-called maximum entropy (ME) model. The model makes this determination by using features describing the candidate predicate and argument. These features include features from the input sentence as well as features from part-of-speech tagging and base NP chunking from step 304. For example, features include the predicate headword and its part of speech and the argument headword and its part of speech. Additionally, there are features specifying headwords and part-of-speech tags surrounding the candidate predicate and argument, for example, a distance between headwords or a number of chunks to a headword. Other features consider the predicate and argument together, for example, whether the predicate or argument occurs first, the number of punctuations between the predicate and argument and the sequence of parts-of-speech of chunk headword between the predicate and the argument.
The ME model is an exponential model that can learn to find skeleton relations given the output of the part-of-speech tagger and base noun phrase chunker of the skeleton parser. The exponential model that is used is shown below as equation 1:
In equation 1, h is a candidate word pair and its surrounding context, t is a prediction, and f1, . . . , fk are characteristic functions each corresponding to a possible feature value holding in h and value of t. p(h,t) is the probability of h and t as determined by the ME model. It will further be appreciated that alternative components can be used for the ME model, such as different part-of-speech tags, distances between predicate and argument, etc. in accordance with the present invention. Moreover, the ME model may be replaced with other machine learning models, also in accordance with the present invention.
A central aspect of the skeleton parser described above is that the parser is trained on the output of a pre-existing parser, currently embodied by EZParse, a hand-crafted parser. A key advantage of this approach is that training data can be obtained without costly human-annotation. When the skeleton parser is trained with enough (easily-obtainable) data, the resulting parser has increased accuracy over the original EZParse parser. Use of the skeleton parser in applications that need skeleton relations is preferable to use of EZParse because the former is more than five times as fast.
Furthermore, learning more complex features rather than less complex ones from noisy data can increase the skeleton parser's accuracy. For example, using complex part-of-speech tags as features, though noisier, can be more effective than using simple part-of-speech tags. Additionally, a ME model can achieve high accuracy in finding grammatical relations. Also, experiments in selecting which features are to be used in the ME model can enhance the development of the skeleton parser. Features related to the candidate predicate and argument by their relative positions and by feature type, either word or part-of-speech, can serve to increase model accuracy.
Although the present invention has been described with reference to particular embodiments, workers skilled in the art will recognize that changes may be made in form and detail without departing from the spirit and scope of the invention.
|Cited Patent||Filing date||Publication date||Applicant||Title|
|US4811210||Aug 29, 1986||Mar 7, 1989||Texas Instruments Incorporated||A plurality of optical crossbar switches and exchange switches for parallel processor computer|
|US4914590||May 18, 1988||Apr 3, 1990||Emhart Industries, Inc.||Natural language understanding system|
|US5060155 *||Jan 31, 1990||Oct 22, 1991||Bso/Buro Voor Systeemontwikkeling B.V.||Method and system for the representation of multiple analyses in dependency grammar and parser for generating such representation|
|US5068789 *||Sep 15, 1989||Nov 26, 1991||Oce-Nederland B.V.||Method and means for grammatically processing a natural language sentence|
|US5193192||Aug 23, 1990||Mar 9, 1993||Supercomputer Systems Limited Partnership||Vectorized LR parsing of computer programs|
|US5371807 *||Mar 20, 1992||Dec 6, 1994||Digital Equipment Corporation||Method and apparatus for text classification|
|US5649215||Jan 12, 1995||Jul 15, 1997||Richo Company, Ltd.||Language parsing device and method for same|
|US5687384||Oct 12, 1994||Nov 11, 1997||Fujitsu Limited||Parsing system|
|US5696980||Apr 27, 1993||Dec 9, 1997||Sharp Kabushiki Kaisha||Machine translation system utilizing bilingual equivalence statements|
|US5937190||Apr 3, 1995||Aug 10, 1999||Synopsys, Inc.||Architecture and methods for a hardware description language source level analysis and debugging system|
|US6098042||Jan 30, 1998||Aug 1, 2000||International Business Machines Corporation||Homograph filter for speech synthesis system|
|US6182029||Aug 6, 1999||Jan 30, 2001||The Trustees Of Columbia University In The City Of New York||System and method for language extraction and encoding utilizing the parsing of text data in accordance with domain parameters|
|US6353824||Nov 18, 1997||Mar 5, 2002||Apple Computer, Inc.||Method for dynamic presentation of the contents topically rich capsule overviews corresponding to the plurality of documents, resolving co-referentiality in document segments|
|US6446081||Dec 16, 1998||Sep 3, 2002||British Telecommunications Public Limited Company||Data input and retrieval apparatus|
|US6473730||Apr 12, 1999||Oct 29, 2002||The Trustees Of Columbia University In The City Of New York||Method and system for topical segmentation, segment significance and segment function|
|US6675159||Jul 27, 2000||Jan 6, 2004||Science Applic Int Corp||Concept-based search and retrieval system|
|US6681206||Oct 18, 2000||Jan 20, 2004||At&T Corporation||Method for generating morphemes|
|US6714939||Jan 8, 2001||Mar 30, 2004||Softface, Inc.||Creation of structured data from plain text|
|US6795808||Oct 30, 2000||Sep 21, 2004||Koninklijke Philips Electronics N.V.||User interface/entertainment device that simulates personal interaction and charges external database with relevant data|
|US6895430||Jun 21, 2000||May 17, 2005||Eric Schneider||Method and apparatus for integrating resolution services, registration services, and search services|
|US6963831 *||Oct 25, 2000||Nov 8, 2005||International Business Machines Corporation||Including statistical NLU models within a statistical parser|
|US7158930||Aug 15, 2002||Jan 2, 2007||Microsoft Corporation||Method and apparatus for expanding dictionaries during parsing|
|US7386438||Aug 4, 2003||Jun 10, 2008||Google Inc.||Identifying language attributes through probabilistic analysis|
|US7571157||Dec 29, 2004||Aug 4, 2009||Aol Llc||Filtering search results|
|US20010041980||Jun 6, 2001||Nov 15, 2001||Howard John Howard K.||Automatic control of household activity using speech recognition and natural language|
|US20020046018||May 11, 2001||Apr 18, 2002||Daniel Marcu||Discourse parsing and summarization|
|US20020095445||Nov 30, 2000||Jul 18, 2002||Philips Electronics North America Corp.||Content conditioning method and apparatus for internet devices|
|US20020128821||Mar 11, 2002||Sep 12, 2002||Farzad Ehsani||Phrase-based dialogue modeling with particular application to creating recognition grammars for voice-controlled user interfaces|
|US20030036900||Jul 12, 2001||Feb 20, 2003||Weise David Neal||Method and apparatus for improved grammar checking using a stochastic parser|
|US20030046087||Aug 12, 2002||Mar 6, 2003||At&T Corp.||Systems and methods for classifying and representing gestural inputs|
|US20030130837||Jul 31, 2002||Jul 10, 2003||Leonid Batchilo||Computer based summarization of natural language documents|
|US20030182102||Mar 20, 2002||Sep 25, 2003||Simon Corston-Oliver||Sentence realization model for a natural language generation system|
|US20030200077||Apr 19, 2002||Oct 23, 2003||Claudia Leacock||System for rating constructed responses based on concepts and a model answer|
|US20030212543||May 7, 2002||Nov 13, 2003||International Business Machines Corporation||Integrated development tool for building a natural language understanding application|
|US20030233224||Feb 19, 2003||Dec 18, 2003||Insightful Corporation||Method and system for enhanced data searching|
|US20040024739||Jul 1, 2003||Feb 5, 2004||Kanisa Inc.||System and method for implementing a knowledge management system|
|US20040030540||Jul 3, 2003||Feb 12, 2004||Joel Ovil||Method and apparatus for language processing|
|US20040044952||Oct 17, 2001||Mar 4, 2004||Jason Jiang||Information retrieval system|
|US20040059564||Sep 19, 2002||Mar 25, 2004||Ming Zhou||Method and system for retrieving hint sentences using expanded queries|
|US20040059574||Sep 20, 2002||Mar 25, 2004||Motorola, Inc.||Method and apparatus to facilitate correlating symbols to sounds|
|US20040102957||Nov 14, 2003||May 27, 2004||Levin Robert E.||System and method for speech translation using remote devices|
|US20040111253||Dec 10, 2002||Jun 10, 2004||International Business Machines Corporation||System and method for rapid development of natural language understanding using active learning|
|US20040181389||May 28, 2002||Sep 16, 2004||Didier Bourigault||Method and large syntactical analysis system of a corpus, a specialised corpus in particular|
|US20050027512||Aug 30, 2004||Feb 3, 2005||Microsoft Corporation||Ranking parser for a natural language processing system|
|US20050076037||Oct 2, 2003||Apr 7, 2005||Cheng-Chung Shen||Method and apparatus for computerized extracting of scheduling information from a natural language e-mail|
|US20050086592||Oct 15, 2003||Apr 21, 2005||Livia Polanyi||Systems and methods for hybrid text summarization|
|US20050137848 *||Dec 19, 2003||Jun 23, 2005||Xerox Corporation||Systems and methods for normalization of linguisitic structures|
|US20050222837||Apr 6, 2005||Oct 6, 2005||Paul Deane||Lexical association metric for knowledge-free extraction of phrasal terms|
|US20050234707 *||Apr 16, 2004||Oct 20, 2005||International Business Machines Corporation||Chinese character-based parser|
|US20050273314||Jun 7, 2004||Dec 8, 2005||Simpleact Incorporated||Method for processing Chinese natural language sentence|
|US20060074634 *||Oct 6, 2004||Apr 6, 2006||International Business Machines Corporation||Method and apparatus for fast semi-automatic semantic annotation|
|US20060095250||Nov 3, 2004||May 4, 2006||Microsoft Corporation||Parser for natural language processing|
|US20060277028||Jun 1, 2005||Dec 7, 2006||Microsoft Corporation||Training a statistical parser on noisy data by filtering|
|EP0387226A1||Mar 5, 1990||Sep 12, 1990||International Business Machines Corporation||Natural language analyzing apparatus and method|
|EP1209560A1||Nov 21, 2000||May 29, 2002||Natlantech Naamloze Vennootschap||A language modelling system and a fast parsing method|
|1||Adam Berger, Stephen A. Della Pietra and V. Della Pietra, 1996: "A Maximum Entropy Approach to Natural Language Processing" Association for Computational Linguistics, IBM T.J. Watson Research Center, Yorktown Heights, NY, vol. 22, No. 1, Computational Linguistics.|
|2||Adwait Ratnaparkhi. 1996. A maximum entropy model for part-of-speech tagging. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, pp. 133-142, Somerset, NJ.|
|3||Alexander Yeh. 2000. "Comparing two trainable grammatical relations finders" In Proceedings of the 18th International Conference on Computational Linguistics (COLING 2000), pp. 1146-1150, Saarbruecken, Germany.|
|4||Brian Roark and Michiel Bacciani. 2003. Supervised and unsupervised PCFG adaptation to novel domains. In Proceedings of HLT-NAACL 2003, pp. 205-212, Edmonton, Alberta, Canada.|
|5||Chen, Office Action mailed Nov. 1, 2010 for U.S. Appl. No. 11/142,703, filed Jun. 1, 2005.|
|6||Daniel Gildea, "Corpus Variation and Parser Performance" University of California, Berkeley and International Compute Science Institute, 1997.|
|7||Emami et al., "Extract Training on a Neural Syntactic Language Model", EEE International Conference on Acoustics, Speech, and Signal Processing, vol. 1, May 17-21, 2004, 1245-1248.|
|8||Eugene Charniak, "A Maximum-Entropy-Inspired Parser", Brown Laboratory for Linguistic Information Processing, Dept. of Computer Science, Providence, RI.|
|9||Fernando Pereira and Yves Schabes. 1992. "Inside-outside re-estimation from partially bracketed corpora" In Proceedings of the 30th Annual Meeting of the Association for Computational Linguistics, pp. 128-135, Newark, DE.|
|10||Judita Preiss. 2003. "Using grammatical relations to compare parsers" In Proceedings of the 10th Conference of the European Chapter of the Association for Computational Linguistics.|
|11||Mark Johnson and Stefan Riezler. 2000. Exploiting auxiliary distributions in stochastic unification-based grammars. In Proceedings of the 38th Annual Meeting of the Association for Computational Linguistics, Hong Kong, China.|
|12||Mark Steedman et al. 2003a. Example Selection for bootstrapping statistical parses. In Proceedings of the Joint Conference of Human Language Technologies and the Annual Meeting of the North American Chapter of the Association for Computational Linguistics, pp. 236-243, Edmonton, Canada.|
|13||Mark Steedman, Miles Osbourne, Anoop Sarkar, Stephen Clark, Rebecca Hwa, Julia Hockenmaier, Paul Ruhlen, Steven Baker and Jeremiah Crim. 2003. "Boot-strapping statistical parsers from small datasets" In Proceedings of the 11th Conference of the European Chapter of the Association for Computational Linguistics, Budapest, Hungary.|
|14||Michael Collins, 1999, Head-Driven Statistical Models for Natural Language Parsing, Ph.D. thesis, University of Pennsylvania, Daniel Gildea 2001. Corpus variation and parser performance. In Proceedings of the Sixth Conference on Empirical Methods in Natural Language Processing (EMNLP-01), Pittsburgh, PA.|
|15||Michael Collins. 1996. A new statistical parser based on bigram lexical dependencies. In Proceedings of the 34th Annual Meeting of the Association for Computational Linguistics. pp. 184-191.|
|16||Miles Osborne. 2002. "Shallow parsing using noisy and non-stationary training material" Journal of Machine Learning Research, 1(1):695-719.|
|17||Ming Tang et al. 2002. Active Learning for Statistical Natural Language parsing. In Proceedings of the 40th Annual Meeting of the Assoc. for Comp. Linguistics, Philadelphia, PA. 8 pages.|
|18||Mitchell Marcus, Beatrice Santorini, et al. 1993. Building a large annotated corpus for English: the penn treebank. Computational Linguistics, 19(2):313-330.|
|19||Rebecca Hwa, 1999. "Supervised grammar induction using training data with limited constituent information" In Proceedings of the 37th Annual Meeting of the Association for Computational Linguistics, pp. 73-79, College Park, MD.|
|20||Rebecca Hwa. 1999. Sample Selection for statistical parsing. Computational Linguistics, 30(3). 39 pages.|
|21||Robbert Prins and Gertjan van Noord. 2001. "Unsupervised POS-tagging improves parsing accuracy and parsing efficiency". In Proceedings of the Seventh International Workshop on Parsing Technologies, pp. 154-165, Beijing, China.|
|22||Sabine Nicole Buchholz. 2002 Memory-based Grammatical Relations Findings, PhD. thesis, Tilbuerg University. 232 pages.|
|23||Shlomo Argamon, Ido Dagan, and Yuval Krymolowski 1998, A memory-based approach to learning shallow natural language patterns. In Proceedings of the 17th International Conference on Computational Linguistics and the 36th Annual Meeting of the Association for Computational Linguistics, pp. 67-73, Montreal, Canada.|
|24||Tashiro et al. "Efficient Chart Parsing of Speech Recognition Candidates", IEEE International Conference on Acoustics, Speech, and Signal Processing, vol. II, Apr. 19-22, 1994, II/13-II/16.|
|25||U.S. Appl. No. 11/142,703 Office Action mailed Apr. 2, 2010. 18 pages.|
|26||U.S. Appl. No. 11/142,703 Office Action mailed Dec. 20, 2007. 22 pages.|
|27||U.S. Appl. No. 11/142,703 Office Action mailed May 8, 2009. 10 pages.|
|28||U.S. Appl. No. 11/142,703 Office Action mailed Nov. 16, 2009. 11 pages.|
|29||U.S. Appl. No. 11/142,703 Office Action mailed Oct. 10, 2008. 9 pages.|
|30||Xun, E. Zhou et al. 2000. A Unified Statistical Model for the Identification of English Based NP. In the 38th Annual Meeting of the Association for Computational Linguistics, Hong Kong, Oct. 3-6. 8 pages.|
|Citing Patent||Filing date||Publication date||Applicant||Title|
|US8489388 *||Nov 10, 2008||Jul 16, 2013||Apple Inc.||Data detection|
|US8738360||Sep 29, 2008||May 27, 2014||Apple Inc.||Data detection of a character sequence having multiple possible data types|
|US8798990 *||Apr 30, 2013||Aug 5, 2014||At&T Intellectual Property Ii, L.P.||Methods and systems for natural language understanding using human knowledge and collected data|
|US8935151 *||Dec 7, 2011||Jan 13, 2015||Google Inc.||Multi-source transfer of delexicalized dependency parsers|
|US9305544||Jan 12, 2015||Apr 5, 2016||Google Inc.||Multi-source transfer of delexicalized dependency parsers|
|US9454522||May 23, 2014||Sep 27, 2016||Apple Inc.||Detection of data in a sequence of characters|
|US9489371||Jul 12, 2013||Nov 8, 2016||Apple Inc.||Detection of data in a sequence of characters|
|US20100121631 *||Nov 10, 2008||May 13, 2010||Olivier Bonnet||Data detection|
|US20130311170 *||Apr 30, 2013||Nov 21, 2013||At&T Intellectual Property Ii, L.P.||Methods and Systems for Natural Language Understanding Using Human Knowledge and Collected Data|
|WO2016141187A1 *||Mar 3, 2016||Sep 9, 2016||The Allen Institute For Artificial Intelligence||System and methods for generating treebanks for natural language processing by modifying parser operation through introduction of constraints on parse tree structure|
|U.S. Classification||704/9, 704/1, 704/10|
|International Classification||G06F17/20, G06F17/21, G06F17/27|
|Dec 13, 2004||AS||Assignment|
Owner name: MICROSOFT CORPORATION, WASHINGTON
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHEN, JOHN T.;ZHOU, MING;WU, TIANLEI;REEL/FRAME:015451/0416
Effective date: 20041103
|Nov 24, 2014||FPAY||Fee payment|
Year of fee payment: 4
|Dec 9, 2014||AS||Assignment|
Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034543/0001
Effective date: 20141014