|Publication number||US7089541 B2|
|Application number||US 10/058,647|
|Publication date||Aug 8, 2006|
|Filing date||Jan 28, 2002|
|Priority date||Nov 30, 2001|
|Also published as||US20030106049|
|Publication number||058647, 10058647, US 7089541 B2, US 7089541B2, US-B2-7089541, US7089541 B2, US7089541B2|
|Inventors||David M. Ungar|
|Original Assignee||Sun Microsystems, Inc.|
|Export Citation||BiBTeX, EndNote, RefMan|
|Patent Citations (27), Non-Patent Citations (8), Referenced by (19), Classifications (5), Legal Events (4)|
|External Links: USPTO, USPTO Assignment, Espacenet|
This application claims the benefit of U.S. Provisional Application No. 60/334,274, filed Nov. 30, 2001, which is incorporated herein by reference in its entirety.
The invention relates to parsing technology, in particular, to techniques for decomposing a complex parser, such as for a computer programming language, into successive passes of comparatively simple miniparsers that operate on the outputs of respective predecessor miniparsers.
2. Description of the Related Art
In computer technology, a parser is a program, usually part of a compiler, that receives input in the form of sequential source program instructions, interactive online commands, markup tags, or some other defined interface and breaks them up into parts (for example, the nouns (objects), verbs (methods), and their attributes or options) that can then be managed by other programming (for example, other components in a compiler). A parser may also check to see that all input has been provided that is necessary.
Parsers typically translate an input information encoding such as source code text into abstract syntax trees in two steps: first, a lexical analyzer or lexer transforms source code text into a series of tokens or word-like pieces; then a parser converts the tokens into a parse tree. Abstract syntax is a representation of data (e.g., a program being compiled) which is independent of machine-oriented structures and encodings and also of the physical representation of the data. In the case of compilation, the syntax is called concrete syntax and includes all the features visible in the source program such as parentheses and delimiters. The concrete syntax is used when parsing the program or other input, during which it is usually converted into some kind of abstract syntax tree. An abstract syntax tree (AST) is a data structure representing something which has been parsed, often used as a compiler or interpreter's internal representation of a program while it is being optimized and from which code generation is performed. The range of all possible such structures is described by the abstract syntax. A compiler's internal representation of a program will typically be specified by an abstract syntax in terms of categories such as “statement”, “expression” and “identifier”. This is independent of the source syntax (concrete syntax) of the language being compiled (though it may be similar). A parse tree is similar to an abstract syntax tree but it will typically also contain features such as parentheses which are syntactically significant but which are implicit in the structure of the abstract syntax tree.
Although interactive programming environments have found widespread acceptance, most classic parser architectures hail from an era of when computation was a scarce resource. Bottom-up parser generators have succeeded in two important goals: languages such as the BNF (“Backus Normal Form” or “Backus-Naur Form”) provide a concise and elegant notation for the expression of a language's syntax, and parsing algorithms such as LR-1 parsing minimize the time and space required to actually perform a parse. The strengths of bottom-up parser generators were critical in the environments that prevailed in the last millennium, but make little difference in many projects today and in the future. For example, when virtual machines feature an interactive programming environment in which only individual methods are expected to be compiled at a time, and when software is developed on machines with fractional-gigahertz processors and hundreds of megabytes of main memory, a parser can afford decreased performance. Thus, parsing efficiency is not always as important as development time.
Moreover, the strengths of traditional parser architectures come with weaknesses. First, the grammar itself requires a parser. A grammar can provide a formal definition of the syntactic structure of a language which is often given in terms of production rules which specify the order of constituents and their sub-constituents in a sentence or string. Of course, the grammar parser can be generated from a grammar, but some time is needed to get over the bootstrap hump. Next, bottom-up, table-driven parsers can be difficult to modify. This problem is merely a specialized case of a challenge that dogs the heels of all nonprocedural languages; a change to the specification, in this case the grammar, frequently creates unanticipated consequences. With LR parser generators, this issue typically surfaces when a programmer makes a change to the grammar only to discover that she has created unforeseen ambiguities. The hard-won description of the target language's grammar is concise but not malleable. Grammar-driven parsers require implementation effort for the grammar, and bottom-up grammar driven parsers can be brittle.
Even when a traditional, grammar-driven, parser has been tweaked to accept the desired grammar, the parser's output leaves much to be desired. A grammar-driven parse produces a concrete syntax tree, whose topology results from the hierarchical relationships between the grammar's productions. But the grammar is as much a function of what is parsable by a particular algorithm as it is of the target syntax. For example, left- or right-recursion in the grammar can induce a tall, skinny subtree when a short, flat one would be better. Or, if the grammar is incompletely factored as happens all too often, different kinds of tree nodes may redundantly implement the same semantic construct, reflecting its contextual syntactic legality instead of its meaning. Therefore, the choice to employ a grammar-driven parser frequently implies a commitment to write a post-processing system to clean up and reshape the parse tree. Not only does this system add implementation effort to a parser, but it also further impedes malleability. Whenever the grammar is perturbed, the tree postprocessor must also change.
Traditional parsers typically use a top-down recursive-descent parsing algorithm. A traditional such parser would recursively descend the grammar, traversing each token once (modulo look-ahead), in order to build the final parse tree. For example, in a typical implementation of a recursive-descent parser, at every juncture, each possibility must be tried. Because an attempt may fail, each token may actually be examined may times. Each stage of the parse must correctly choose between all possible results that could possibly start with what has already been parsed. But the only data available to make this decision are the tokens lying ahead in the input stream. Recursive descent optimizes performance but makes it harder to generate a correct parse.
A modular parser includes a number of miniparsers. Each miniparser is capable of receiving input and generating an output dependent thereon. In one embodiment, a lexer or a miniparser capable of lexical analysis receives certain code and provides an abstract syntax tree. Another miniparser receives the abstract tree and performs certain operations on the input abstract syntax tree and generates an output abstract syntax tree. Other miniparsers each receive an input syntax tree and generate corresponding output syntax tree. Each miniparser performs operations on a set of syntactical constructs to produce the corresponding output syntax tree, which in turn may become an input syntax tree for a next miniparser.
In one embodiment of the present invention, a multipass parser implementation includes more than one miniparser. Each miniparser is successively operable on a respective abstract syntax tree. The abstract syntax tree corresponds to an input information encoding and includes transformations of predecessor ones of the miniparsers, if any. Respective ones of the miniparsers are limited to particular subsets of syntactic constructs to be parsed in the input information encoding. The multipass parser implementation may be embodied as a computer program product encoded in at least one computer readable medium. The multipass parser implementation may be embodied as or within a software engineering tool.
In another embodiment of the present invention, a method of implementing a parser for an input information encoding includes defining a succession of miniparsers. Each miniparser is operable on a respective parse state resulting from a predecessor one of the miniparsers. Each of the miniparsers recognizes only a particular subset of syntactic constructs to be parsed in the input information encoding. In a further embodiment, the method further includes executing the miniparsers in succession.
In another embodiment of the present invention, a method of parsing an information encoding includes performing plural successive transformations on successive abstract syntax trees, each abstract syntax tree formed as a result of a predecessor one of the transformations. Each of the successive transformations handles only a subset of syntactic constructs to be recognized in the information encoding.
In another embodiment of the present invention, a computer program product is encoded in at least one computer readable medium. The computer program product includes functional encodings of at least two miniparsers. A first one of the miniparsers is executable to transform a first parse tree into a second parse tree, and a second one of the miniparsers is executable to transform the second parse tree into a third parse tree. Each of the at least two miniparsers recognizes only a subset of syntactic constructs to be parsed in an information encoding to which the first, second and third parse trees correspond.
In another embodiment of the present invention, an apparatus includes encoded information and a multipass means. The encoding of information is substantially in accordance with a grammar. The multipass means is for performing a number of successive transformations on the encoded information. Each successive transformation handles only a subset of syntactic constructs in accordance with the grammar.
The foregoing is a summary and thus contains, by necessity, simplifications, generalizations and omissions of detail. Consequently, those skilled in the art will appreciate that the foregoing summary is illustrative only and that it is not intended to be in any way limiting of the invention. Other aspects, inventive features, and advantages of the present invention, as defined solely by the claims, will become apparent in the detailed description set forth below.
The present invention may be better understood, and its numerous objects, features, and advantages made apparent to those skilled in the art by referencing the accompanying drawings. The use of the same reference symbols in different drawings indicates similar or identical items.
The following discussion is intended to provide a detailed description of at least one example of the invention to aid those skilled in the art to understand and practice the invention. For example,
Parse block 130 includes a number of miniparsers. In the example shown in
Parser 110 does not attempt to make a single leap from tokens to fully-parsed trees. Instead, parser 110 decomposes the task of parsing into multiple, simple passes to be performed by miniparsers 132, 134, for example. Each pass is referred to as a miniparse and operates on a respective parse state received from a previous miniparser or received as an initial input of coded information. Each pass takes the output of the previous miniparse pass, if applicable, and produces a new parse tree for a successor miniparser, if applicable. Each parser processes an abstract syntax tree from a previous parser, if applicable, according to a particular subset of syntactic constructs such as comments, delimiters or grouping elements, names, class-level statements, expressions, etc.
Parser 110 performs at least the following two functions: (i) parsing, and (ii) representing the results of the parse. These functions may be represented by the same hierarchy of objects, or they may be represented by two disparate inheritance families of objects, one to parse, and the other to represent the parsed abstract syntax tree. Exemplary parser 110 follows the latter approach. The state for each function has a different lifetime, and the parsing state may only exists while parsing. Also, the parse node objects in the abstract syntax tree can be structured according to their meaning as opposed to their syntax. This division works well with the miniparsers doing the parsing, and the parseNodes objects being name-spaces that hold the parse nodes.
A short example illustrates the multipass approach to parsing. When referring to each successive parse, the following exemplary source file named “main.java” will be used:
Referring again to
Referring again to
Now, each top-level-statement can be visited and parsed. As shown in
Thus, the first several miniparsers of the above description perform the following operations:
As shown in
This modular architecture trades off any extra cycles that may occur to revisit the same portion of the program to be parsed for the simplicity of constructing each miniparser. However, this is not really a significant trade off, if at all, because even a standard, recursive descent parser may scan the same tokens more than once, either to look ahead to choose the correct nonterminal to parse, or to back up and try an alternative when a choice is revealed to be incorrect.
In the embodiments described herein, functional decomposition has been used to modularize parser 110. Moreover, parser 110 nodes are immutable. Once a parse node's creation is complete, none of its state may further change. Instead of rearranging a parse tree it receives, a miniparser produces a new tree. Since subtrees are immutable, the new tree is free to reuse the subtrees of the old one. The functional programming paradigm confers two benefits in this context: ease of debugging and the a priori elimination of bugs. Since a miniparser does not alter its input, both its input and output are available for inspection when debugging a miniparser. And since parse trees are immutable, a class of bugs resulting from mutating shared objects is eliminated. Based on the teachings herein, functional programming is a good match for the intermediate and final results of a multipass parser.
Each miniparser contains mutable states to perform its parse. For example the parenthesis matcher has a stack of umatched opening parentheses that have been seen so far in the parse. Therefore, parser 110 implements a specialization hierarchy of miniparsers that is completely separate from its hierarchy of parse node objects. This dichotomy facilitates the task of decoupling the form of the parse tree from the specifics of the grammar. The miniparsers reflect the grammar, and the parse node hierarchy reflects the semantic content.
The architecture of parser 110 also allows clearer error messages. A bottom-up, generated parser can only output error messages on the mechanical syntactic level, admonishing the user about what sorts of tokens would have been legal. The user is left to try to figure out why. See, for example, the C compiler's obscure “missing semicolon” message. A recursive-descent compiler has only unparsed tokens to look ahead to, and so cannot delimit the extent of the relevant portion of the program. Parser 110 can supply more helpful error messages, since the error messages may be written by hand and may also talk about appropriate regions of code. For example, an error in a package declaration can highlight the entire declaration statement if appropriate.
One embodiment of parser 110 has been implemented in Self, a prototype-based dynamic object-oriented language. The reflective facilities available in Self (as in Smalltalk and Lisp) allowed simplification of the lexer, and object-oriented habits militate towards representation of everything as objects in specialization hierarchies, including tokens. These facets of parser 110 are beneficial but not essential.
One embodiment of parser 110 has been divided into a general-purpose parser kit and a refinement specifically for Java. The Java parser consists of two global-name-space objects (akin to Java packages), each of which belongs to a separate module. The parseKit module contains classes for lexing (the lexer), parsing (the parser), sample objects, and a testing framework. The javaParser module specializes the parseKit to create a parser for the Java language. Its name space overrides corresponding objects in parseKit.
The above description is intended to describe at least one embodiment of the invention, not to define the scope of the invention. Rather, the scope of the invention is defined in the claims that follow this description. Thus, other embodiments of the invention include other variations, modifications, additions, and/or improvements to the above description.
Computer readable media may include, for example and without limitation, any number of the following: magnetic storage media including disk and tape storage media; optical storage media such as disc media (e.g., CD-ROM, CD-R, DVD-ROM, etc.); nonvolatile memory storage media including semiconductor-based memory units such as FLASH memory, EEPROM, EPROM, ROM; ferromagnetic digital memories; holographic media, volatile storage media including registers, buffers or caches, main memory, RAM, etc.; and data transmission media including computer networks, point-to-point telecommunication equipment, and carrier wave transmission media, just to name a few. Other new and various types of computer-readable media may be used to store and/or transmit the software modules discussed herein.
Also, those skilled in the art will recognize that boundaries between logic blocks are merely illustrative and that alternative embodiments may merge logic blocks or circuit elements or impose an alternate decomposition of functionality upon various logic blocks or circuit elements. For example, miniparsers 132 and 134 may be directly resident on platform 100 independent of any overarching parse block 130. Moreover, alternative embodiments may combine multiple instances of a particular component (e.g., more miniparsers within parse block 130.
In one embodiment, system 100 is a computer system such as a personal computer system. Other embodiments may include different types of computer systems. Computer systems are information handling systems which can be designed to give independent computing power to one or more users. Computer systems may be found in many forms including but not limited to mainframes, minicomputers, servers, workstations, personal computers, notepads, personal digital assistants, various wireless devices and embedded systems. A typical computer system includes at least one processing unit, associated memory and a number of input/output (I/O) devices.
A computer system processes information according to a program and produces resultant output information via I/O devices. A program is a list of instructions such as a particular application program and/or an operating system. A computer program is typically stored internally on computer readable storage medium or transmitted to the computer system via a computer readable transmission medium. A computer process typically includes an executing (running) program or portion of a program, current program values and state information, and the resources used by the operating system to manage the execution of the process.
From an operational perspective, each block of
Those skilled in the art will recognize that boundaries between the functionality of the above described operations/modules are merely illustrative. The functionality of multiple operations may be combined into a single operation, and/or the functionality of a single operation may be distributed in additional operations. Moreover, alternative embodiments may include multiple instances of a particular operation, and the order of operations may be altered in various other embodiments. Likewise, those skilled in the art will recognize that boundaries between modules are merely illustrative and alternative embodiments may merge modules or impose an alternative decomposition of functionality of modules. For example, the modules discussed herein may be decomposed into submodules to be executed as multiple computer processes. Moreover, alternative embodiments may combine multiple instances of a particular module or submodule.
It is to be understood that the architectures depicted herein are merely exemplary, and that in fact many other architectures can be implemented which achieve the same functionality. In an abstract, but still definite sense, any arrangement of components to achieve the same functionality is effectively “associated” such that the desired functionality is achieved. Hence, any two components herein combined to achieve a particular functionality can be seen as “associated with” each other such that the desired functionality is achieved, irrespective of architectures or intermedial components. Likewise, any two components so associated can also be viewed as being “operably connected”, or “operably coupled”, to each other to achieve the desired functionality.
The components and devices described herein are used as examples for sake of conceptual clarity. Consequently, as used herein these specific exemplars are intended to be representative of their more general classes. Furthermore, in general, the use of any specific exemplar herein is also intended to be representative of its class and the non-inclusion of any specific devices in any exemplary lists herein should not be taken as indicating that limitation is desired.
Because the above detailed description is exemplary, when “one embodiment” is described, it is an exemplary embodiment. Accordingly, the use of the word “one” in this context is not intended to indicate that one and only one embodiment may have a described feature. Rather, many other embodiments may, and often do, have the described feature of the exemplary “one embodiment.” Thus, as used above, when the invention is described in the context of one embodiment, that one embodiment is one of many possible embodiments of the invention.
Notwithstanding the above caveat regarding the use of the words “one embodiment” in the detailed description, it will be understood by those within the art that if a specific number of an introduced claim element is intended in the below claims, such an intent will be explicitly recited in the claim, and in the absence of such recitation no such limitation is present or intended. For example, in the claims below, when a claim element is described as having “one” feature, it is intended that the element be limited to one and only one of the feature described. Furthermore, when a claim element is described in the claims below as including or comprising “a” feature, it is not intended that the element be limited to one and only one of the feature described. Rather, for example, the claim including “a” feature reads upon an apparatus or method including one or more of the feature in question. That is, because the apparatus or method in question includes a feature, the claim reads on the apparatus or method regardless of whether the apparatus or method includes another such similar feature. This use of the word “a” as a non-limiting, introductory article to a feature of a claim is adopted herein by Applicants as being identical to the interpretation adopted by many courts in the past, notwithstanding any anomalous or precedential case law to the contrary that may be found. Similarly, when a claim element is described in the claims below as including or comprising an aforementioned feature (e.g., “the” feature), it is intended that the element not be limited to one and only one of the feature described merely by the incidental use of the definite article.
Furthermore, the use of introductory phrases such as “at least one” and “one or more” in the claims should not be construed to imply that the introduction of another claim element by the indefinite articles “a” or “an” limits any particular claim containing such introduced claim element to inventions containing only one such element, even when the same claim includes the introductory phrases “one or more” or “at least one” and indefinite articles such as “a” or “an.” The same holds true for the use of definite articles.
While particular embodiments of the present invention have been shown and described, it will be obvious to those skilled in the art that, based upon the teachings herein, various modifications, alternative constructions, and equivalents may be used without departing from the invention claimed herein. Consequently, the appended claims encompass within their scope all such changes, modifications, etc. as are within the true spirit and scope of the invention. Furthermore, it is to be understood that the invention is solely defined by the appended claims. The above description is not intended to present an exhaustive list of embodiments of the invention. Unless expressly stated otherwise, each example presented herein is a non-limiting or nonexclusive example, whether or not the terms non-limiting, nonexclusive or similar terms are contemporaneously expressed with each example. Although an attempt has been made to outline some exemplary embodiments and exemplary variations thereto, other embodiments and/or variations are within the scope of the invention as defined in the claims below.
|Cited Patent||Filing date||Publication date||Applicant||Title|
|US4686623 *||Jun 7, 1985||Aug 11, 1987||International Business Machines Corporation||Parser-based attribute analysis|
|US4887212 *||Oct 29, 1986||Dec 12, 1989||International Business Machines Corporation||Parser for natural language text|
|US4914585 *||May 23, 1988||Apr 3, 1990||Hewlett-Packard Company||Modular complier with a class independent parser and a plurality of class dependent parsers|
|US5060155 *||Jan 31, 1990||Oct 22, 1991||Bso/Buro Voor Systeemontwikkeling B.V.||Method and system for the representation of multiple analyses in dependency grammar and parser for generating such representation|
|US5157759 *||Jun 28, 1990||Oct 20, 1992||At&T Bell Laboratories||Written language parser system|
|US5386570 *||May 24, 1993||Jan 31, 1995||Hewlett-Packard Company||Method for a two pass compiler with the saving parse states from first to second pass|
|US5687378 *||Jun 7, 1995||Nov 11, 1997||Motorola, Inc.||Method and apparatus for dynamically reconfiguring a parser|
|US5812760 *||Jun 25, 1996||Sep 22, 1998||Lsi Logic Corporation||Programmable byte wise MPEG systems layer parser|
|US5878273 *||Jun 7, 1995||Mar 2, 1999||Discovision Associates||System for microprogrammable state machine in video parser disabling portion of processing stages responsive to sequence-- end token generating by token generator responsive to received data|
|US5903756 *||Oct 11, 1996||May 11, 1999||Sun Microsystems, Incorporated||Variable lookahead parser generator|
|US6226291 *||Oct 31, 1997||May 1, 2001||Texas Instruments Incorporated||Transport stream packet parser system|
|US6268874 *||Aug 4, 1998||Jul 31, 2001||S3 Graphics Co., Ltd.||State parser for a multi-stage graphics pipeline|
|US6275791 *||Feb 26, 1999||Aug 14, 2001||David N. Weise||Natural language parser|
|US6292186 *||Nov 6, 1998||Sep 18, 2001||International Business Machines Corporation||Universal information appliance with parser|
|US6307546 *||Dec 30, 1997||Oct 23, 2001||Alcatel Usa Sourcing, L.P.||Telecommunications system craft interface device with parser having object-oriented state machine|
|US6377999 *||May 10, 1999||Apr 23, 2002||Interniche Technologies Inc.||Firmware and software protocol parser|
|US6385769 *||Feb 3, 1999||May 7, 2002||International Business Machines Corporation||Text based object oriented program code with a visual program builder and parser support for predetermined and not predetermined formats|
|US6397263 *||Nov 3, 1993||May 28, 2002||International Business Machines Corporation||String command parser for message based systems|
|US6397383 *||Feb 26, 1999||May 28, 2002||International Business Machines Corp.||Constraint language parser which rejects invalid filter constraint during run time|
|US6493662 *||Feb 11, 1998||Dec 10, 2002||International Business Machines Corporation||Rule-based number parser|
|US6512775 *||Nov 8, 1996||Jan 28, 2003||The Trustees Of Columbia University In The City Of New York||Method and apparatus for a programmable bitstream parser for audiovisual and generic decoding systems|
|US6523172 *||Feb 19, 1999||Feb 18, 2003||Evolutionary Technologies International, Inc.||Parser translator system and method|
|US6611524 *||Jun 30, 1999||Aug 26, 2003||Cisco Technology, Inc.||Programmable data packet parser|
|US6732153 *||May 23, 2000||May 4, 2004||Verizon Laboratories Inc.||Unified message parser apparatus and system for real-time event correlation|
|US6772190 *||Dec 4, 2001||Aug 3, 2004||Dejima, Inc.||Distributed parser of natural language input|
|EP0710027A2 *||Oct 13, 1995||May 1, 1996||General Instrument Corporation Of Delaware||Syntax parser for a video decompression processor|
|WO1996000436A1 *||Jun 26, 1995||Jan 4, 1996||Microsoft Corp||Method and system for bootstrapping statistical processing into a rule-based natural language parser|
|1||*||"Three Models For Descriptions of Languages", Noam Chomsky, MIT, 1956, IRE Trans. Infor, Theory, pp. 113-124.|
|2||Beck, Kent, Smalltalk Best Practice Patterns, 1997 by Prentice Hall PTR, pp. 13-16, pp. 171-189.|
|3||Calingaert, Peter, Assemblers, Compilers, and Program Translation, 1979 Computer Science Press, Inc., pp. 73-232.|
|4||*||Compilers Principles, Techniques and Tools, Aho et al, Chapters 1-5, Sep. 12, 1985.|
|5||*||IBM Technical Bulletin, Mar. 1, 2002, System and Method For JAVA serialization Compatibility using XML, 2 pages.|
|6||*||JAVA!, Tom Richey, Sep. 22, 1995, Chapters 1, 12-14.|
|7||*||Support for Modular Parsing in Software Reengineering, Peake et al, IEEE, Jul. 1997, pp. 58-66.|
|8||*||The Web Server Book, Janathan Magid et al, published Jul. 28, 1995, Chapters 1 and 11.|
|Citing Patent||Filing date||Publication date||Applicant||Title|
|US7415696 *||Dec 2, 2005||Aug 19, 2008||Adobe Systems Incorporated||Extensible help facility for a computer software application|
|US7962904||May 10, 2007||Jun 14, 2011||Microsoft Corporation||Dynamic parser|
|US8296744 *||Oct 3, 2008||Oct 23, 2012||Microsoft Corporation||Tree-based directed graph programming structures for a declarative programming language|
|US8738360 *||Sep 29, 2008||May 27, 2014||Apple Inc.||Data detection of a character sequence having multiple possible data types|
|US8762969||Aug 7, 2008||Jun 24, 2014||Microsoft Corporation||Immutable parsing|
|US8855999||Feb 5, 2014||Oct 7, 2014||Palantir Technologies Inc.||Method and system for generating a parser and parsing complex data|
|US8903717||Feb 21, 2014||Dec 2, 2014||Palantir Technologies Inc.||Method and system for generating a parser and parsing complex data|
|US8924388||Dec 6, 2013||Dec 30, 2014||Palantir Technologies Inc.||Computer-implemented systems and methods for comparing and associating objects|
|US8924389||Dec 24, 2013||Dec 30, 2014||Palantir Technologies Inc.||Computer-implemented systems and methods for comparing and associating objects|
|US8930897||Oct 2, 2013||Jan 6, 2015||Palantir Technologies Inc.||Data integration tool|
|US9081975||Oct 22, 2012||Jul 14, 2015||Palantir Technologies, Inc.||Sharing information between nexuses that use different classification schemes for information access control|
|US20040088425 *||Oct 31, 2002||May 6, 2004||Comverse, Ltd.||Application level gateway based on universal parser|
|US20060080607 *||Dec 2, 2005||Apr 13, 2006||Adobe Systems, Incorporated, A Delaware Corporation||Extensible help facility for a computer software application|
|US20090138850 *||Nov 18, 2008||May 28, 2009||Fujitsu Limited||Processing device for extracting immutable entity of program and processing method|
|US20090306964 *||Sep 29, 2008||Dec 10, 2009||Olivier Bonnet||Data detection|
|US20100088665 *||Oct 3, 2008||Apr 8, 2010||Microsoft Corporation||Tree-based directed graph programming structures for a declarative programming language|
|US20140289715 *||Jun 2, 2014||Sep 25, 2014||Microsoft Corporation||Immutable parsing|
|EP2778913A1 *||Mar 11, 2014||Sep 17, 2014||Palantir Technologies, Inc.||Method and system for generating a parser and parsing complex data|
|EP2778914A1 *||Mar 11, 2014||Sep 17, 2014||Palantir Technologies, Inc.||Method and system for generating a parser and parsing complex data|
|U.S. Classification||717/143, 717/112|
|Jan 28, 2002||AS||Assignment|
Owner name: SUN MICROSYSTEMS, INC., CALIFORNIA
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:UNGER, DAVID M.;REEL/FRAME:012540/0952
Effective date: 20020122
|May 2, 2002||AS||Assignment|
Owner name: SUN MICROSYSTEMS, INC., CALIFORNIA
Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNOR S NAME PREVIOUSLY RECORDED ON REEL 012540, FRAME 0952;ASSIGNOR:UNGAR, DAVID M.;REEL/FRAME:012871/0713
Effective date: 20020122
|Jan 6, 2010||FPAY||Fee payment|
Year of fee payment: 4
|Jan 8, 2014||FPAY||Fee payment|
Year of fee payment: 8