WO2000023918A9 - Modular image query system - Google Patents

Modular image query system

Info

Publication number
WO2000023918A9
WO2000023918A9 PCT/US1999/024094 US9924094W WO0023918A9 WO 2000023918 A9 WO2000023918 A9 WO 2000023918A9 US 9924094 W US9924094 W US 9924094W WO 0023918 A9 WO0023918 A9 WO 0023918A9
Authority
WO
WIPO (PCT)
Prior art keywords
feature
image
modules
database
scoring
Prior art date
Application number
PCT/US1999/024094
Other languages
French (fr)
Other versions
WO2000023918A1 (en
Inventor
Hawley Knox Rising Iii
Original Assignee
Sony Electronics Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sony Electronics Inc filed Critical Sony Electronics Inc
Priority to AU12065/00A priority Critical patent/AU1206500A/en
Publication of WO2000023918A1 publication Critical patent/WO2000023918A1/en
Publication of WO2000023918A9 publication Critical patent/WO2000023918A9/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/5838Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using colour
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99931Database or file accessing
    • Y10S707/99933Query processing, i.e. searching
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99931Database or file accessing
    • Y10S707/99933Query processing, i.e. searching
    • Y10S707/99934Query formulation, input preparation, or translation

Definitions

  • the present invention relates to computer-based systems which provide analysis and query by image content.
  • Conventional image processing techniques include many methods for extracting characteristics or features from an image. For example, techniques are known for extracting color, texture, and component shape characteristics of a given image.
  • the prior art techniques for extracting the color features of an image typically employ one of two methods. First, the user may select a desired color, which is used as the basis for an image color query. Images are matched to the selected color based on the average color of the matched image over the entirety of the image.
  • a second prior art image color analysis technique determines not only the overall color of a desired image, but also the percentage coverage of that color and the compactness of its coverage in a desired image. The percentage color coverage and color compactness are used as additional query conditions in these prior art systems.
  • An example of these conventional techniques is given in E. Binaghi, et al, "Indexing and Fuzzy Logic-Based Retrieval of Color Images", Visual Database Systems, II. IFIP Transactions A-7, pp. 79-92, Elsevier Science Publishers, 1992.
  • Texture features such as granularity, directionality, and tiling features of a given image can be extracted using known techniques.
  • One example of such techniques is found in H. Tamura, et al, "Textural Features Corresponding to Visual Perception", IEEE Proceedings, Vol. SMC-8, No. 6, June 1978, pp.460-473.
  • Still other techniques are known in the prior art for classifying an image based on structure features, which represent shapes found in the image. Using these known techniques, predefined shapes, such as rectangular, triangular, or circular shapes among others, may be compared to an image to determine the presence of such shapes in the image. This known technique may used to query an image database for images having a particular specified shape.
  • One example of a prior art method for image analysis based on shapes is found in G. Taubin and D.B. Cooper, "Recognition and Positioning of Rigid Objects Using Algebraic Moment Invariants", Geometric Methods in Computer Vision, SPIE , Vol. 1570, pp. 175-186, 1992.
  • U.S. Patent No. 5,751,286 describes an image query system and method wherein the visual characteristics of an image such as color, texture, shape, and size are used to develop an image query.
  • the technique described in this patent involves selecting from a plurality of image characteristic selections represented by thumbnail icons corresponding to various image characteristics for a particular image query. As shown in the '286 patent, these image characteristic (feature) selections are submitted to a query by image content (QBIC) engine, which compares the various image characteristic selections with a database of stored images.
  • QBIC image content
  • the described centralized QBIC engine must be capable of handling all of the supported types of image feature processing.
  • the fully supportive QBIC engine has a number of significant drawbacks.
  • image analysis and image query systems demand a high degree of processing power.
  • processing even one of the various types of image characteristics, such as color or texture involves many processor cycles and data storage accesses.
  • An image query system such as the one described in the '286 patent, that supports a plurality of image characteristic analysis methods must therefore be a very complex and expensive system to implement.
  • images for a particular application of such a system may be more appropriately analyzed by a particular image characteristic analysis method and much less efficiently analyzed using other image characteristic analysis methods.
  • an image query system that is configurable for a particular application. Specifically, it would be advantageous to provide an image query system that supported image analysis techniques most appropriate for the types of images encountered in a particular application.
  • Such a configurable image query system should be modular and extensible so that a user need only purchase or program those image analysis methods most appropriate for the particular application and so new image analysis methods may be easily incorporated into an existing system. The prior art does not disclose such a system.
  • the Oracle 8 image data cartridges object interfaces associate specific data with procedures that can operate on that data.
  • the image procedures provide the means by which the images can be copied, format converted, and processed on demand.
  • the Oracle 8 image data cartridges merely support various image and graphic file formats rather than supporting a variety of image content analysis techniques.
  • An image query and storage apparatus and method including a plurality of dynamically linkable feature modules including a plurality of dynamically linkable feature modules.
  • Each of the plurality of feature modules extract a different set of feature information from an input image.
  • the method and apparatus further includes a database coupled to the plurality of feature modules.
  • the database includes storage for the different set of feature information for each of the plurality of feature modules.
  • the method and apparatus support a query by image content of the database of images using the dynamically linked plurality of feature modules.
  • the method and apparatus further includes a plurality of dynamically linkable scoring modules for processing feature specific scoring information generated by the feature modules.
  • Figure 2 illustrates the components of a feature module.
  • Figure 3 illustrates the components of a feature descriptor.
  • Figure 4 illustrates the components of the scoring manager.
  • Figure 5 illustrates the components of a scoring module.
  • Figure 6 illustrates the components of the registry.
  • Figure 7 illustrates the components of the database.
  • Figure 8 illustrates the components of the image feature descriptors as stored in the database.
  • Figure 9 illustrates the steps performed in the process of adding an image to the database.
  • Figure 10 illustrates the steps performed in the process of querying the image database by image content.
  • Figure 11 illustrates the steps performed in the process of registering a feature module or scoring module.
  • Figure 12 illustrates a conventional computer system upon which the present invention may be implemented.
  • the present invention is a method and apparatus for enabling configurable and modular image query.
  • numerous specific details are set forth in order to provide a thorough understanding of the present invention. It will be evident, however, to one of ordinary skill in the art that the present invention may be practiced without these specific details. In other circumstances, well-known structures and devices, and interfaces are shown in block diagram form in order to avoid unnecessarily obscuring the present invention.
  • the present invention is a modular, configurable, and extendable image storage and query system.
  • the invention has the advantage of a modular architecture enabling dynamic installation of modules for performing specific types of feature analysis and scoring on a particular image or database of images.
  • the basic architecture of the present invention is illustrated in Figure 1.
  • User interface component 110 represents software for interfacing with a human user. Using conventional methods, user interface 110 displays various windows, menus, icons, and controls other conventional user input and output functions. User interface 110 provides a means by which a user may provide image input and system configuration selection information input to the modular image storage and query system 100 of the present invention.
  • image input to system 100 includes user-identified images digitized and formatted in conventional ways.
  • the present invention can process images in a bit-mapped format commonly identified by the file name extension ".bmp". It will be apparent to those of ordinary skill in the art that other image file formats may similarly be supported.
  • system 100 receives images and image queries through user interface 110.
  • the images are each analyzed by feature manager 120 using a plurality of dynamically linked feature modules 125.
  • Feature manager 120 uses the feature modules 125 to analyze an input image and to extract a set of image characteristics represented by a plurality of feature descriptors.
  • the feature descriptors for each input image are transferred by feature manager 120 through user interface 110 and storage interface 140 into database 160 for permanent storage.
  • System 100 also allows a user to run a query by image content against the set of images in database 160 represented by feature descriptors.
  • As input to an image query a user provides an input image through user interface 110.
  • the feature manager 120 invokes feature modules 125 to analyze the characteristics of the input image.
  • the feature manager 120 produces a set of feature descriptors for the input image.
  • the feature descriptors for the input image are compared with feature descriptors from the database 160 of stored image data.
  • Feature manager 120 produces a set of scoring data based on the comparison of the input image feature descriptors with the database-resident feature descriptors.
  • This scoring information is passed to scoring manager 130 which uses a set of dynamically linked scoring modules 135 to process the scoring information.
  • scoring manager 130 produces a set of sorted images most closely corresponding to the input image based on the pre-configured set of feature modules 125 and scoring modules 135. This set of sorted images corresponding to the input image may then be displayed to user via user interface 110.
  • the processes of the present invention for adding an image to database 160, for querying an image by its content, and for registering a feature module or scoring module are described in more detail below.
  • user interface 110 communicates with database 160 and a registry 150 through a storage interface 140.
  • Storage interface 140 provides a means for abstracting the implementation detail of registry 150 and database 160 from user interface 110. In this manner, a variety of different database implementations may be used without impacting the user interface 110 implementation.
  • a registry 150 is used for storage of system configuration information and pointers to feature modules 125, scoring modules 135, and image data in database 160.
  • Database 160 is used for the storage of feature descriptors for each image added to system 100. Further details of the structure of registry 150 and database 160 are provided below.
  • Each feature module is responsible for performing all of the processing necessary for analyzing an input image for a particular characteristic or feature.
  • Such particular features include color, texture, or wavelet response.
  • conventional techniques exist for analyzing an image to extract particular features such as color or texture and to produce information that describes the particular characteristics of the image, a
  • feature modules 125 perform all of the processing necessary for analyzing an input image for a particular characteristic. Because all of the feature specific knowledge is retained by the particular feature module 125, the rest of system 100 does not need to be designed with a specific feature analysis method or set of feature analysis methods in mind.
  • the present invention can be configured to operate with a variety of different configurations of feature modules 125.
  • the feature modules 125 are implemented as dynamic link library (DLL) components, which can be dynamically linked to the system 100 using conventional methods.
  • DLL dynamic link library
  • each of the features modules 125 are designed to extract a particular image feature, each of feature modules 125 contain a common set of components as illustrated in Figure 2.
  • Feature modules 125 include an image storage area 410 for storing an image being added to database 160 or an input image upon which an image query will be run. Feature modules 125 also include a query image storage area 420. Area 420 is used for storing images from database 160 which are to be compared with an input image stored in area 410. Feature score storage 430 is used for storage of score information compiled by the feature module during an image query. Feature image analyzer 440 represents a software component or programming code module comprising processing logic for extracting particular image characteristics from an input image and for generating a feature descriptor representing the extracted features.
  • Feature score generator 450 represents a software component or programming code module containing processing logic for generating feature specific scoring information during an image query. This feature specific scoring information is stored in area 430.
  • Feature descriptor 460 is a storage area used by the feature image analyzer 440 for the storage of a feature descriptor generated by the feature module from an input image. The feature descriptor 460 will have a different format depending on the type of feature being extracted from the image and the type of feature extraction methodology used by the particular feature module. One such sample format of a feature descriptor 460 is illustrated in Figure 3.
  • Figure 3 illustrates an embodiment of a structure and method for generating a multi-element feature descriptor, which is capable of implementing the teachings of the present invention. Particularly, Figure 3 illustrates, in block flow diagram format, a method of generating a feature descriptor which is representative of a multi-band image for use in image processing.
  • Image features extracted from the output of spatial filters are often used for image representation.
  • the application of multi-band images to spatial filters enables the construction of feature sets which contain a wide range of spectral and spatial properties.
  • One such type of oriented spatial filter is the steerable filter. Steerable filters obtain information about the response of a filter at any orientation
  • an image [I(x,y)] 900 is applied to the steerable filter [Filter f ⁇ (x,y)] 905 which provides two different matrices for each image, an orientation matrix 110 and an energy matrix 915.
  • the orientation matrix 110 also referred to as an Orientation Map ⁇ (I(x,y)) 910, is derived by computing the dominant orientation at each pixel position (x,y) by using the equation:
  • the energy matrix 915 also referred to as an Energy Map E (I(x,y)) 915, corresponds to the dominant orientation at each pixel position (x,y) in accordance with the equation:
  • E(x,y) E( ⁇ ° )+E(60 ° )+ E(l20 ° )+ 2 ⁇ E 2 ( ⁇ ° )+ E 2 ( ⁇ 0 ° )+ E 2 (l20 ° ) -E(0 ° XE(60 ° )+ E(l20 ° ))- E(60 ° )E(l20 ° )
  • the Orientation Map ⁇ (I(x,y)) 910 and the Energy Map E (I(x,y)) 915 a corresponding histogram or set of histograms is used to represent global information, along with a set of co-occurence matrices which are used to represent local information.
  • the Orientation Map ⁇ (I(x,y)) 910 is represented as a corresponding orientation histogram H ( ⁇ ) 920 and set of orientation co-occurence matrices C ⁇ 925.
  • the Energy Map E (I(x,y)) 915 is represented as a corresponding energy histogram H(E) 930 and set of energy co-occurence matrices CE 935.
  • each image 900 is represented by a corresponding orientation histogram ( ⁇ ) 920, a set of orientation co-occurence matrices C ⁇ 925, a corresponding energy histogram H(E) 930, and a set of energy co-occurence matrices CE 935.
  • a series of descriptors are extracted from each of the corresponding histograms and co-occurence matrices.
  • the descriptors extracted from the orientation histogram ( ⁇ ) 920 of the Orientation Map ⁇ (I(x,y)) 910 are peak descriptors (PD) 940 and statistic descriptors (SD1) 945.
  • the peak descriptors (PD) 940 comprise position, value, and shape data associated with the orientation histogram ( ⁇ ) 920.
  • the statistic descriptors (SD1) 945 indicate mean, standard deviation, third and fourth order moments associated with the orientation histogram H( # ) 920. Select elements within the peak descriptors (PD) 940 are used to classify images into different categories, whereas the statistic descriptors (SD1) 945 are used to describe the shape of the orientation histogram H( ⁇ ) 920.
  • the descriptors extracted from the orientation co-occurence matrices C ⁇ 925 of the Orientation Map ⁇ (I(x,y)) 910 are co-occurence descriptors (COD1) 950.
  • the co-occurence descriptors (COD1) 950 comprise maximum probability, entropy, uniformity, mean, correlation, and difference moments.
  • the co-occurence descriptors (COD1) 950 in the present embodiment are computed in four different orientations (-45 Degrees, O Degrees, 45 Degrees, and 90 Degrees).
  • the descriptors extracted from the energy histogram H(E) 930 of the Energy Map E (I(x,y)) 915 are statistic descriptors (SD2) 955.
  • the statistic descriptors (SD2) 955 indicate mean, standard deviation, third and fourth order moments associated with the energy histogram H(E) 930.
  • the statistic descriptors (SD2) 955 associated with the energy histogram H(E) 930 are used to describe the shape of the orientation energy histogram H(E) 930.
  • the descriptors extracted from the energy co-occurence matrices CE 935 of the Energy Map E (I(x,y)) 915 are co-occurence descriptors (COD2) 960.
  • the co-occurence descriptors (COD2) 960 comprise maximum probability, entropy, uniformity, mean, correlation, and difference moments.
  • the co-occurence descriptors (COD2) 960 in the present embodiment are computed in four different orientations (-45 Degrees, O Degrees, 45 Degrees, and 90 Degrees).
  • each of the descriptors associated with an image is combined in order to form a feature vector or feature descriptor 965.
  • each individual descriptor 970 associated with an image comprises peak descriptors (PD) 940, statistic descriptors (SD1) 945, co-occurence descriptors (COD1) 950, statistic descriptors (SD2) 955, and co-occurence descriptors (COD2) 960, which are combined to form a feature descriptor 965.
  • the feature descriptor 965 is a full representation of each image which may be used for image processing.
  • a feature descriptor 965 is generated for each information band comprising the multi-band image, as such, each information band associated with each multi-band image has a corresponding feature descriptor 965.
  • a multi-band image using the RGB color spectrum would have an individual feature descriptor 965 for each information band or color band (RGB) of the multi-band image.
  • Scoring manager 130 is responsible for receiving a set of scoring information corresponding to scores generated for each individual image by the feature modules 125 and feature manager 120. Because only an individual feature module 125 has the knowledge to process an image for that particular feature, only the feature module 125 can generate a score when that particular feature is compared between two images. Although the feature modules 125 generate scoring information for particular images, the scoring manager 130 in combination with scoring modules 135 must compile this scoring information to produce a composite score for a set of images. In one embodiment of the present invention, the composite score represents a sorting order of a set of query images compared to an input image.
  • the sorting order typically defines the order of the query images from the query images most similar to the input image to the query images least similar to the input image. It will be apparent to those of ordinary skill in the art that many other types of sorting orders or scoring arrangements may equivalently be implemented for a set of image scoring information.
  • the composite score storage area 510 is used for the storage of information defining the composite score of a set of images.
  • the composite score sorter 520 represents software or processing logic for generating sorting information which is passed to user interface 110 for sorting the query images for display to the user.
  • each scoring module 135 will therefore include a composite score generator 530 for generating a composite score according to a particular methodology implemented by that particular scoring module 135.
  • a particular scoring methodology is a simple weighted averaging technique for averaging the individual image scores across the entire scoring domain of the image set.
  • a three-dimensional space may be generated and the individual image scores may be plotted in this three- dimensional space.
  • the composite score generator 530 may then compute the distance between the input image score location in this three-dimensional space and the location of the score of each query image in the three-dimensional space. It will be apparent to one of ordinary skill in the art that many other scoring techniques may be used and implemented on individual scoring modules 135.
  • the scoring modules 135 are implemented as dynamic link library (DLL) components, which can be dynamically linked to the system 100 using conventional methods.
  • DLL dynamic link library
  • Registry 150 and database 160 are used to store system configuration information and information related to a library of searchable images. Referring now to Figure 6, the main components of registry 150 are illustrated. Registry 150 includes general configuration data 610. Because system 100 can include an arbitrary number of dynamically linked feature modules 125 and scoring modules 135, it is necessary to maintain a current count of the number of these modules installed in a system at a particular time. This information is stored in general configuration data 610. Further, other global information useful for any of the components of system 100, such as the number of images in database 160, are stored in general configuration data 610.
  • registry 150 is used for the storage of information for locating and accessing feature modules 125, scoring modules 135, and images in database 160.
  • registry 150 includes image pointers 630, which identify the location of each image in database 160. As each image is added to database 160, using a method described in more detail below, a pointer to the image is added to image pointers 630.
  • One other basic function of registry 150 in the preferred embodiment is the formation of a correspondence between the descriptive name given to a particular feature module and the dynamic link library (DLL) component used to actually implement the feature module.
  • DLL dynamic link library
  • the descriptive feature names 620 are stored in registry 150 along with their corresponding feature module DLL identifiers 640.
  • the descriptive feature names are names of feature extraction methods or image analysis tools displayed to a user for selection through user interface 110.
  • the corresponding feature module DLL identifiers are the DLLs actually invoked when a corresponding descriptive feature name is invoked by the user.
  • the registry 150 is a text file written in the format of a Windows 3.1 .ini file. It will be apparent to those of ordinary skill in the art that other formats for storing registry information in an alternative embodiment may be used. Referring now to Figure 7, the main component of database 160 is illustrated. Database 160 is used for storage of image feature descriptors 710 for each image as generated by the feature modules 125.
  • Image feature descriptors 710 includes a feature descriptor corresponding to each of the feature modules 125 installed in system 100.
  • the detail of image feature descriptors 710 is illustrated in Figure 8.
  • a plurality of feature descriptors corresponding to the installed feature modules 125 are stored for each image in database 160.
  • the number of feature descriptors for each image varies depending on the number of feature modules 125 currently installed in system 100.
  • a header is maintained for each image to assist in maintaining the linked list of feature descriptors for the image.
  • the feature descriptors for each image are stored in image feature descriptors 710, the actual image data represented in a bit-mapped form is stored in image data 720.
  • Image data 720 represents the actual bit-mapped image data that may be used to display the image on a display device.
  • this image data is stored as bit-mapped (.bmp) image data, however it will be apparent to one of ordinary skill in the art that the image data may be represented in any of a variety of conventional image coding techniques.
  • a detail of the image data 720 is illustrated in Figure 8.
  • a pointer to each bit-mapped image data block is maintained in registry 150.
  • the image data for each image in database 160 and the feature descriptors corresponding to the image may be readily located during the process of querying database 160.
  • the database 160 is written in a RIFF format, which is a conventional information format created by IBM Corporation and Microsoft Corporation for manipulation of video and multimedia files.
  • a plurality of databases 160 may be employed in a particular system. Using a plurality of databases 160, a set of images can be partitioned into separate databases and thereby searched more quickly.
  • a database identifier for each of the plurality of installed databases 160 is stored in registry 150.
  • the system In a multiple database 160 configuration, the system must provide a database identifier during the process of adding an image to a database or during the process of querying an image against one or more databases of the system.
  • a user provides a new image for entry into database 160 through user interface 110.
  • the user identifies a data file containing the raw image data of the new image.
  • This image data is retrieved from the identified file in processing block 200 illustrated in Figure 9.
  • the retrieved image data is passed to feature manager 120.
  • Feature manager 120 begins a loop in which the image data is passed to each of the previously installed feature modules 125. This loop process is illustrated in Figure 9 at bubble 205 and box 210.
  • feature manager 120 By invoking each of the installed feature modules in loop 205, feature manager 120 causes each of the feature modules 125 to produce a corresponding feature descriptor generated as a result of each feature module's analysis of the input image data.
  • feature manager 120 begins loop 215 illustrated in Figure 9.
  • each of the feature modules 125 contain all of the information and processing methods for analyzing an input image for a particular feature. As such, it is possible for the feature descriptor generated by each feature module to be of varying length or content. Also, because the feature specific information is retained in feature modules 125, feature manager 120 cannot know the length or content of the feature descriptor generated by each feature module 125. For this reason, feature manager 120 executes loop 215 to request from each feature module 125 the length of the feature descriptor generated by each feature module (processing blocks 220).
  • the image data 720 for the new image is added to database 160 in processing block 225.
  • a data block of an arbitrary size may be requested and obtained from database 160.
  • the data blocks necessary for storage of the image feature descriptors 710 are prepared.
  • the header 730 for the new image is written to database 160.
  • a loop 230 is initiated for writing each of the feature descriptors generated by feature modules 125 to database 160. Because the length of each feature descriptor is known from the processing preformed in blocks 220, the free data blocks of the appropriate length can be obtained from database 160.
  • each of the feature descriptors of the new image are written to the feature descriptor blocks 740 allocated for the new image.
  • the database 160 nor the feature manager 120 are aware of the format or content of the feature descriptor written to feature descriptor blocks 740. Only the particular feature module 125 which generated the particular feature descriptor can decode and interpret the particular feature descriptor stored in database 160.
  • processing block 240 illustrated in Figure 9 is executed to complete the "add image" process of the preferred embodiment of the present invention.
  • the registry 150 is updated to reflect the presence of a new image.
  • the general configuration data 610 is updated to advance the number of images counter by one.
  • the image pointers 630 are updated to reflect the presence and location of the new image in database 160. It will be apparent to one of ordinary skill in the art that other incidental data may need to be updated to reflect the addition of a new image to database 160.
  • the new image and its corresponding feature descriptors are stored in database 160 and available for query by a user.
  • a user may provide or identify to system 100 through user interface 110 an image representing an input image to a query request.
  • system 100 will analyze the features of the input image and then search database 160 for images resident in database 160 (i.e. query images) having features most similar to the input image provided by the user.
  • This query process is described in more detail in Figure 10.
  • an input image is identified by a user through user interface 110 and retrieved from the specified location.
  • the input image data is provided by user interface 110 to feature manager 120.
  • Feature manager 120 begins loop 305 in which the input image is sequentially passed to each of the feature modules 125.
  • Each feature module analyzes the input image for each particular feature corresponding to each installed feature module 125 (processing blocks 310).
  • each feature module 125 generates a feature descriptor corresponding to the input image.
  • a set of feature descriptors as generated by each feature module 125 will have been created.
  • This set of feature descriptors for the input image represents the feature information which will be compared with corresponding feature information for each of the images in the database (denoted query image feature descriptors).
  • the query image feature descriptors are retrieved from database 160.
  • the feature descriptors for all of the images resident in database 160 are retrieved for comparison with the input image feature descriptors.
  • the user may specify a subset of the images resident in database 160 for the purpose of restricting the query to a smaller domain of images.
  • a subset of database 160 resident images may be specified using textual descriptive information or other information corresponding to a particular classification of the images within database 160.
  • a set of query images from multiple databases may be specified.
  • the query image feature descriptors for a set of images from database 160 that will be used as a query domain are retrieved in processing block 315.
  • feature manager 120 passes the input image feature descriptors and the query image feature descriptors to each of the feature modules 125.
  • the input image feature descriptors are stored in image storage 410 and the query image feature descriptors are stored in query image storage 420.
  • each feature module 125 may be stored on each feature module 125 or alternatively stored in a central location, such as within feature manager 120 or in another component for processing by a particular feature module.
  • feature manager 120 invokes each of the feature modules 125 and passes to each feature module the input image feature descriptor and the query image feature descriptor for the corresponding feature module.
  • Each feature module 125 sequentially processes the feature descriptors and compares the input image feature descriptor with the query image feature descriptor. As a result of this comparison, a score is generated by the feature module to indicate the level of similarity or dissimilarity between the input image and the query image.
  • the feature score generator for 50 as shown is Figure 2 compares the input feature descriptor and the query feature descriptor for the particular feature and generates a score.
  • the score for the particular feature comparison is stored in feature score storage 430 as shown in Figure 2.
  • processing passes from decision block 330 to processing block 315 where the feature descriptors for the next query image are obtained from database 160 and the new query image feature descriptors are processed through each of the feature modules 125 via loop 320.
  • loop 335 is executed.
  • feature manager 120 sequentially cycles through each of the feature modules 125 to obtain the scoring information generated as a result of the feature module's comparison of the input image to each of the query images. In the preferred embodiment, this scoring information is obtained from feature score storage 430 shown in Figure 2.
  • processing passes to loop 345 shown in Figure 10.
  • the scoring information obtained by feature manager 120 from each of the feature modules 125 is passed through user interface 110 to scoring manager 130. It will be apparent to one of ordinary skill in the art that the scoring information in an alternative embodiment may equivalently be passed directly from feature manager 120 to a scoring manager 130.
  • Scoring manager 130 initiates loop 345.
  • scoring manager 130 passes to scoring modules 135 the scoring information generated by each of the feature modules 125 for each of the query images from database 160. Using this information, scoring modules 135 compute the similarities between the input image and each of the query images (processing blocks 350).
  • scoring modules 135 which can each implement one of the variety of scoring techniques.
  • Each of these installed scoring techniques as implemented on scoring modules 135 are invoked by scoring manager 130 sequentially as loop 345 is executed.
  • the scoring modules 135 generate the similarities in scoring information between the input image and each of the query images (processing blocks 350). Once the similarities between the input image and the query images is generated in processing blocks 350, scoring manager 130 enters loop 355 where each of the scoring modules 135 are again invoked sequentially to compute the final scores which rank the query images in an order corresponding to their similarity to the input image (processing blocks 360). Having ordered the query images according to their similarity with the input image, the most similar images may then be displayed to the user through user interface 110 at processing block 365.
  • processing block 1110 the user invokes an "Add Feature" or an "Add Scoring Method” command through user interface 110. This command directs the user through a series of dialog boxes and input screens to obtain the information necessary for registering the new feature module or scoring module.
  • processing block 1120 the user specifies a descriptive name of a particular feature or feature extraction method. The user also specifies a descriptive name corresponding to the scoring method to be added to system 100. The descriptive names of the new feature or scoring method are displayed in various user interface 110 menus or information windows. A dynamic link library (DLL) module corresponding to the descriptive name of the feature or scoring method is also provided by the user in processing block 1120.
  • DLL dynamic link library
  • the DLL corresponds to the executable software necessary for implementing the feature extraction method or scoring method.
  • the user specifies the location of the DLL, typically by specifying the path or file name corresponding to the DLL in a particular file system.
  • the user-entered descriptive name and DLL identifier is added to the registry 150 in processing block 1130.
  • the user specified DLL is linked to the linked list of feature modules or scoring modules in processing block 1140.
  • a feature module 125 or scoring module 135 may be dynamically linked with system 100 and thereafter be employed for analyzing input images.
  • FIG 12 illustrates a typical data processing system upon which one embodiment of the present invention is implemented. It will be apparent to those of ordinary skill in the art, however that other alternative systems of various system architectures may also be used.
  • the data processing system illustrated in Figure 12 includes a bus or other internal communication means 801 for communicating information, and a processor 802 coupled to the bus 801 for processing information.
  • the system further comprises a random access memory (RAM) or other volatile storage device 804 (referred to as main memory), coupled to bus 801 for storing information and instructions to be executed by processor 802.
  • Main memory 804 also may be used for storing temporary variables or other intermediate information during execution of instructions by processor 802.
  • the system also comprises a read only memory (ROM) and /or static storage device 806 coupled to bus 801 for storing static information and instructions for processor 802, and a data storage device 807 such as a magnetic disk or optical disk and its corresponding disk drive.
  • Mass data storage device 807 is coupled to bus 801 for storing information and instructions.
  • the system may further be coupled to a display device 821, such as a cathode ray tube (CRT) or a liquid crystal display (LCD) coupled to bus 801 through bus 803 for displaying information to a computer user.
  • An alphanumeric input device 822 including alphanumeric and other keys, may also be coupled to bus 801 through bus 803 for communicating information and command selections to processor 802.
  • cursor control 823 such as a mouse, a trackball, stylus, or cursor direction keys coupled to bus 801 through bus 803 for communicating direction information and command selections to processor 802, and for controlling cursor movement on display device 821.
  • cursor control 823 such as a mouse, a trackball, stylus, or cursor direction keys coupled to bus 801 through bus 803 for communicating direction information and command selections to processor 802, and for controlling cursor movement on display device 821.
  • a hard copy device 824 which may be used for printing instructions, data, or other information on a medium such as paper, film, or similar types of media.
  • a communication device 826 is coupled to bus 801 through bus 803 for use in accessing other nodes of a distributed system via a network.
  • This communication device 826 may include any of a number of commercially available networking peripheral devices such as those used for coupling to an Ethernet, token ring, Internet, or wide area network. Note that any or all of the components of the system illustrated in Figure 12 and associated hardware may be used in various embodiments of the present invention; however, it will be appreciated by those of ordinary skill in the art that any configuration of the system may be used for various purposes according to the particular implementation. In one embodiment of the present invention, the data processing system illustrated in Figure 1 is an
  • Processor 102 may be one of the X86 compatible microprocessors such as the PENTIUM® brand microprocessors manufactured by INTEL® Corporation of Santa Clara, California upon which a conventional operating system such as the Windows 95 brand operating system developed by Microsoft Corporation of Redmond, Washington is executed.
  • control logic or software implementing the present invention can be stored in main memory 804, mass storage device 807, or other storage medium locally accessible to processor 802.
  • Other storage media may include floppy disk drives, memory cards, flash memory, or CD-ROM drives.
  • This control logic or software may also be resident on an article of manufacture comprising a computer readable medium 808 having computer readable program code embodied therein and being readable by the mass storage device 807 and for causing the processor 802 to coordinate accesses to a storage system in accordance with the teachings herein.

Abstract

An image query and storage apparatus (100) and method including a plurality of dynamically linkable feature modules is disclosed. Each of the plurality of feature modules (125) extract a different set of feature information from an input image. The method and apparatus further includes a database (160) coupled to the plurality of feature modules. The database includes storage for the different set of feature information for each of the plurality of feature modules. The method and apparatus support a query by image content of the database of images using the dynamically linked plurality of feature modules. The method and apparatus further includes a plurality of dynamically linkable scoring modules (135) for processing feature specific scoring information generated by the feature modules.

Description

MODULAR IMAGE QUERY SYSTEM
FIELD OF THE INVENTION
The present invention relates to computer-based systems which provide analysis and query by image content.
DESCRIPTION OF RELATED ART
Conventional computer systems have been employed to analyze visual images. These visual images include, for example, photographic stills, digitally rendered graphics, video clips, and any other monochrome or color images suitable for representation in a digital system. One goal of these image analysis or image processing systems is to generate information about the characteristics of an image so the image can be classified or used to query an image database.
Conventional image processing techniques include many methods for extracting characteristics or features from an image. For example, techniques are known for extracting color, texture, and component shape characteristics of a given image. The prior art techniques for extracting the color features of an image typically employ one of two methods. First, the user may select a desired color, which is used as the basis for an image color query. Images are matched to the selected color based on the average color of the matched image over the entirety of the image. A second prior art image color analysis technique determines not only the overall color of a desired image, but also the percentage coverage of that color and the compactness of its coverage in a desired image. The percentage color coverage and color compactness are used as additional query conditions in these prior art systems. An example of these conventional techniques is given in E. Binaghi, et al, "Indexing and Fuzzy Logic-Based Retrieval of Color Images", Visual Database Systems, II. IFIP Transactions A-7, pp. 79-92, Elsevier Science Publishers, 1992.
Other prior art image analysis techniques are known for extracting texture features of an image. Texture features such as granularity, directionality, and tiling features of a given image can be extracted using known techniques. One example of such techniques is found in H. Tamura, et al, "Textural Features Corresponding to Visual Perception", IEEE Proceedings, Vol. SMC-8, No. 6, June 1978, pp.460-473.
Still other techniques are known in the prior art for classifying an image based on structure features, which represent shapes found in the image. Using these known techniques, predefined shapes, such as rectangular, triangular, or circular shapes among others, may be compared to an image to determine the presence of such shapes in the image. This known technique may used to query an image database for images having a particular specified shape. One example of a prior art method for image analysis based on shapes is found in G. Taubin and D.B. Cooper, "Recognition and Positioning of Rigid Objects Using Algebraic Moment Invariants", Geometric Methods in Computer Vision, SPIE , Vol. 1570, pp. 175-186, 1992.
Other prior art systems have sought to combine a plurality of image color analysis techniques into a single system. For example, U.S. Patent No. 5,751,286 describes an image query system and method wherein the visual characteristics of an image such as color, texture, shape, and size are used to develop an image query. The technique described in this patent involves selecting from a plurality of image characteristic selections represented by thumbnail icons corresponding to various image characteristics for a particular image query. As shown in the '286 patent, these image characteristic (feature) selections are submitted to a query by image content (QBIC) engine, which compares the various image characteristic selections with a database of stored images. Although the '286 patent describes the technique for processing various types of image characteristics, the described centralized QBIC engine must be capable of handling all of the supported types of image feature processing. As will be discussed in more detail below, the fully supportive QBIC engine has a number of significant drawbacks.
As evident from the prior art describing image-processing techniques, image analysis and image query systems demand a high degree of processing power. In fact, processing even one of the various types of image characteristics, such as color or texture, involves many processor cycles and data storage accesses. An image query system, such as the one described in the '286 patent, that supports a plurality of image characteristic analysis methods must therefore be a very complex and expensive system to implement. On the other hand, images for a particular application of such a system may be more appropriately analyzed by a particular image characteristic analysis method and much less efficiently analyzed using other image characteristic analysis methods. Thus, it would be advantageous to enable the configuration of an image query system for a particular application. Unfortunately, the prior art, as represented by the techniques illustrated in the '286 patent, do not enable such a specific configuration given that the QBIC engine is built to handle a full range of image analysis techniques. One problem with this approach is that a user is forced to purchase or program a full-service system even though many of the supported techniques may be underutilized. Further, the system cannot be easily augmented if a new image analysis technique is developed.
It would be advantageous to implement an image query system that is configurable for a particular application. Specifically, it would be advantageous to provide an image query system that supported image analysis techniques most appropriate for the types of images encountered in a particular application. Such a configurable image query system should be modular and extensible so that a user need only purchase or program those image analysis methods most appropriate for the particular application and so new image analysis methods may be easily incorporated into an existing system. The prior art does not disclose such a system.
Some conventional products purport to provide image analysis modularity. Oracle Corporation of Redwood Shores, California developed the image data cartridge component of the Oracle 8 Database. The Oracle 8 image data cartridges object interfaces associate specific data with procedures that can operate on that data. The image procedures provide the means by which the images can be copied, format converted, and processed on demand. In reality, the Oracle 8 image data cartridges merely support various image and graphic file formats rather than supporting a variety of image content analysis techniques.
Thus, a configurable modular image query system supporting modular feature extraction components and modular scoring components is needed.
SUMMARY OF THE INVENTION
An image query and storage apparatus and method including a plurality of dynamically linkable feature modules is disclosed. Each of the plurality of feature modules extract a different set of feature information from an input image. The method and apparatus further includes a database coupled to the plurality of feature modules. The database includes storage for the different set of feature information for each of the plurality of feature modules. The method and apparatus support a query by image content of the database of images using the dynamically linked plurality of feature modules. The method and apparatus further includes a plurality of dynamically linkable scoring modules for processing feature specific scoring information generated by the feature modules.
BRIEF DESCRIPTION OF THE DRAWINGS
The features and advantages of the present invention will be apparent in the drawings identified below followed by the detailed description of the preferred embodiment.
Figure 2 illustrates the components of a feature module.
Figure 3 illustrates the components of a feature descriptor.
Figure 4 illustrates the components of the scoring manager.
Figure 5 illustrates the components of a scoring module. Figure 6 illustrates the components of the registry.
Figure 7 illustrates the components of the database.
Figure 8 illustrates the components of the image feature descriptors as stored in the database.
Figure 9 illustrates the steps performed in the process of adding an image to the database.
Figure 10 illustrates the steps performed in the process of querying the image database by image content.
Figure 11 illustrates the steps performed in the process of registering a feature module or scoring module.
Figure 12 illustrates a conventional computer system upon which the present invention may be implemented.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
The present invention is a method and apparatus for enabling configurable and modular image query. In the following description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the present invention. It will be evident, however, to one of ordinary skill in the art that the present invention may be practiced without these specific details. In other circumstances, well-known structures and devices, and interfaces are shown in block diagram form in order to avoid unnecessarily obscuring the present invention.
The present invention is a modular, configurable, and extendable image storage and query system. The invention has the advantage of a modular architecture enabling dynamic installation of modules for performing specific types of feature analysis and scoring on a particular image or database of images. The basic architecture of the present invention is illustrated in Figure 1.
Referring to Figure 1, the software architecture of the present invention is illustrated. Each of the software components of the present invention illustrated in Figure 1 can be implemented on a conventional computer system such as the sample system illustrated in Figure 12 and described below. User interface component 110 represents software for interfacing with a human user. Using conventional methods, user interface 110 displays various windows, menus, icons, and controls other conventional user input and output functions. User interface 110 provides a means by which a user may provide image input and system configuration selection information input to the modular image storage and query system 100 of the present invention. In general, image input to system 100 includes user-identified images digitized and formatted in conventional ways. For example, the present invention can process images in a bit-mapped format commonly identified by the file name extension ".bmp". It will be apparent to those of ordinary skill in the art that other image file formats may similarly be supported.
At its most basic level, system 100 receives images and image queries through user interface 110. The images are each analyzed by feature manager 120 using a plurality of dynamically linked feature modules 125. Feature manager 120 uses the feature modules 125 to analyze an input image and to extract a set of image characteristics represented by a plurality of feature descriptors. The feature descriptors for each input image are transferred by feature manager 120 through user interface 110 and storage interface 140 into database 160 for permanent storage. System 100 also allows a user to run a query by image content against the set of images in database 160 represented by feature descriptors. As input to an image query, a user provides an input image through user interface 110. The feature manager 120 invokes feature modules 125 to analyze the characteristics of the input image. Again, the feature manager 120 produces a set of feature descriptors for the input image. Once the input image has been analyzed, the feature descriptors for the input image are compared with feature descriptors from the database 160 of stored image data. Feature manager 120 produces a set of scoring data based on the comparison of the input image feature descriptors with the database-resident feature descriptors. This scoring information is passed to scoring manager 130 which uses a set of dynamically linked scoring modules 135 to process the scoring information. As a result of this scoring processing, scoring manager 130 produces a set of sorted images most closely corresponding to the input image based on the pre-configured set of feature modules 125 and scoring modules 135. This set of sorted images corresponding to the input image may then be displayed to user via user interface 110. The processes of the present invention for adding an image to database 160, for querying an image by its content, and for registering a feature module or scoring module are described in more detail below.
Again referring to Figure 1, user interface 110 communicates with database 160 and a registry 150 through a storage interface 140. Storage interface 140 provides a means for abstracting the implementation detail of registry 150 and database 160 from user interface 110. In this manner, a variety of different database implementations may be used without impacting the user interface 110 implementation. In the preferred embodiment of the present invention, a registry 150 is used for storage of system configuration information and pointers to feature modules 125, scoring modules 135, and image data in database 160. Database 160 is used for the storage of feature descriptors for each image added to system 100. Further details of the structure of registry 150 and database 160 are provided below.
Referring now to Figure 2, a more detailed diagram illustrates the components present in each of the feature modules 125. Each feature module is responsible for performing all of the processing necessary for analyzing an input image for a particular characteristic or feature. Such particular features include color, texture, or wavelet response. As indicated in the background section of this patent application, conventional techniques exist for analyzing an image to extract particular features such as color or texture and to produce information that describes the particular characteristics of the image, a In general, feature modules 125 perform all of the processing necessary for analyzing an input image for a particular characteristic. Because all of the feature specific knowledge is retained by the particular feature module 125, the rest of system 100 does not need to be designed with a specific feature analysis method or set of feature analysis methods in mind. By isolating feature specific information into feature modules 125, the present invention can be configured to operate with a variety of different configurations of feature modules 125. In the preferred embodiment, the feature modules 125 are implemented as dynamic link library (DLL) components, which can be dynamically linked to the system 100 using conventional methods. Although each of the features modules 125 are designed to extract a particular image feature, each of feature modules 125 contain a common set of components as illustrated in Figure 2.
Referring now to Figure 2, the common components included in each of the feature modules 125 is illustrated. Feature modules 125 include an image storage area 410 for storing an image being added to database 160 or an input image upon which an image query will be run. Feature modules 125 also include a query image storage area 420. Area 420 is used for storing images from database 160 which are to be compared with an input image stored in area 410. Feature score storage 430 is used for storage of score information compiled by the feature module during an image query. Feature image analyzer 440 represents a software component or programming code module comprising processing logic for extracting particular image characteristics from an input image and for generating a feature descriptor representing the extracted features. Feature score generator 450 represents a software component or programming code module containing processing logic for generating feature specific scoring information during an image query. This feature specific scoring information is stored in area 430. Feature descriptor 460 is a storage area used by the feature image analyzer 440 for the storage of a feature descriptor generated by the feature module from an input image. The feature descriptor 460 will have a different format depending on the type of feature being extracted from the image and the type of feature extraction methodology used by the particular feature module. One such sample format of a feature descriptor 460 is illustrated in Figure 3.
Referring now to Figure 3, an example of a feature descriptor format and a feature extraction methodology is illustrated. Figure 3 illustrates an embodiment of a structure and method for generating a multi-element feature descriptor, which is capable of implementing the teachings of the present invention. Particularly, Figure 3 illustrates, in block flow diagram format, a method of generating a feature descriptor which is representative of a multi-band image for use in image processing.
Image features extracted from the output of spatial filters are often used for image representation. The application of multi-band images to spatial filters enables the construction of feature sets which contain a wide range of spectral and spatial properties. One such type of oriented spatial filter is the steerable filter. Steerable filters obtain information about the response of a filter at any orientation
using a small set of basis filters. In one embodiment, x2 exp X is chosen as 2 * σ2 ) the kernel of the steerable filter. Accordingly, for this kernel of information, a steerable filter at an arbitrary orientation θ can be synthesized using a linear combination of three basis filters according to hθ(x,y) = kl(θ)h0(x,y)+ k2(θ)hω(x,y) + k3(θ)h120(x,y), where
Figure imgf000011_0001
and *,(0)= l + 2cos20
&2(0) = l - cos20 +V3sin20 k3 (ø) = l = cos20 - V3sin 2Θ .
As illustrated in the embodiment of Figure 3, an image [I(x,y)] 900 is applied to the steerable filter [Filter f θ (x,y)] 905 which provides two different matrices for each image, an orientation matrix 110 and an energy matrix 915. The orientation matrix 110, also referred to as an Orientation Map Θ (I(x,y)) 910, is derived by computing the dominant orientation at each pixel position (x,y) by using the equation:
Figure imgf000012_0001
Whereas, the energy matrix 915, also referred to as an Energy Map E (I(x,y)) 915, corresponds to the dominant orientation at each pixel position (x,y) in accordance with the equation:
E(x,y) = E(θ° )+E(60°)+ E(l20°)+ 2^E2°)+ E2(β0°)+ E2(l20°) -E(0°XE(60°)+ E(l20°))- E(60°)E(l20°)
Accordingly, for each matrix or map, the Orientation Map Θ (I(x,y)) 910 and the Energy Map E (I(x,y)) 915, a corresponding histogram or set of histograms is used to represent global information, along with a set of co-occurence matrices which are used to represent local information. As such, the Orientation Map Θ (I(x,y)) 910 is represented as a corresponding orientation histogram H (θ) 920 and set of orientation co-occurence matrices CΘ 925. Similarly, the Energy Map E (I(x,y)) 915 is represented as a corresponding energy histogram H(E) 930 and set of energy co-occurence matrices CE 935. Therefore, each image 900 is represented by a corresponding orientation histogram ( θ) 920, a set of orientation co-occurence matrices C Θ 925, a corresponding energy histogram H(E) 930, and a set of energy co-occurence matrices CE 935. Next, a series of descriptors are extracted from each of the corresponding histograms and co-occurence matrices.
The descriptors extracted from the orientation histogram ( θ ) 920 of the Orientation Map Θ (I(x,y)) 910 are peak descriptors (PD) 940 and statistic descriptors (SD1) 945. The peak descriptors (PD) 940 comprise position, value, and shape data associated with the orientation histogram ( θ ) 920. The statistic descriptors (SD1) 945 indicate mean, standard deviation, third and fourth order moments associated with the orientation histogram H( # ) 920. Select elements within the peak descriptors (PD) 940 are used to classify images into different categories, whereas the statistic descriptors (SD1) 945 are used to describe the shape of the orientation histogram H( θ ) 920.
The descriptors extracted from the orientation co-occurence matrices C Θ 925 of the Orientation Map Θ (I(x,y)) 910 are co-occurence descriptors (COD1) 950. The co-occurence descriptors (COD1) 950 comprise maximum probability, entropy, uniformity, mean, correlation, and difference moments. The co-occurence descriptors (COD1) 950 in the present embodiment are computed in four different orientations (-45 Degrees, O Degrees, 45 Degrees, and 90 Degrees).
Correspondingly, the descriptors extracted from the energy histogram H(E) 930 of the Energy Map E (I(x,y)) 915 are statistic descriptors (SD2) 955. The statistic descriptors (SD2) 955 indicate mean, standard deviation, third and fourth order moments associated with the energy histogram H(E) 930. The statistic descriptors (SD2) 955 associated with the energy histogram H(E) 930 are used to describe the shape of the orientation energy histogram H(E) 930.
Likewise, the descriptors extracted from the energy co-occurence matrices CE 935 of the Energy Map E (I(x,y)) 915 are co-occurence descriptors (COD2) 960. The co-occurence descriptors (COD2) 960 comprise maximum probability, entropy, uniformity, mean, correlation, and difference moments. The co-occurence descriptors (COD2) 960 in the present embodiment are computed in four different orientations (-45 Degrees, O Degrees, 45 Degrees, and 90 Degrees).
Each of the descriptors associated with an image is combined in order to form a feature vector or feature descriptor 965. As such, in this example, each individual descriptor 970 associated with an image comprises peak descriptors (PD) 940, statistic descriptors (SD1) 945, co-occurence descriptors (COD1) 950, statistic descriptors (SD2) 955, and co-occurence descriptors (COD2) 960, which are combined to form a feature descriptor 965. As such, the feature descriptor 965 is a full representation of each image which may be used for image processing. For multi-band applications, a feature descriptor 965 is generated for each information band comprising the multi-band image, as such, each information band associated with each multi-band image has a corresponding feature descriptor 965. For instance, a multi-band image using the RGB color spectrum would have an individual feature descriptor 965 for each information band or color band (RGB) of the multi-band image.
Referring now to Figure 4, the basic components of scoring manager 130 are illustrated. Scoring manager 130 is responsible for receiving a set of scoring information corresponding to scores generated for each individual image by the feature modules 125 and feature manager 120. Because only an individual feature module 125 has the knowledge to process an image for that particular feature, only the feature module 125 can generate a score when that particular feature is compared between two images. Although the feature modules 125 generate scoring information for particular images, the scoring manager 130 in combination with scoring modules 135 must compile this scoring information to produce a composite score for a set of images. In one embodiment of the present invention, the composite score represents a sorting order of a set of query images compared to an input image. The sorting order typically defines the order of the query images from the query images most similar to the input image to the query images least similar to the input image. It will be apparent to those of ordinary skill in the art that many other types of sorting orders or scoring arrangements may equivalently be implemented for a set of image scoring information. As shown in Figure 4, the composite score storage area 510 is used for the storage of information defining the composite score of a set of images. The composite score sorter 520 represents software or processing logic for generating sorting information which is passed to user interface 110 for sorting the query images for display to the user.
Because there can be many different ways of scoring an image query, the architecture of the present invention provides a means for dynamically linking scoring modules to the system 100 for implementing any one or more of the variety of scoring methods. In general, it is a responsibility of each scoring module 135 to generate a composite score using a particular scoring method. As shown in Figure 5, each scoring module 135 will therefore include a composite score generator 530 for generating a composite score according to a particular methodology implemented by that particular scoring module 135. One example of a particular scoring methodology is a simple weighted averaging technique for averaging the individual image scores across the entire scoring domain of the image set. In another example of a particular scoring methodology, a three-dimensional space may be generated and the individual image scores may be plotted in this three- dimensional space. The composite score generator 530 may then compute the distance between the input image score location in this three-dimensional space and the location of the score of each query image in the three-dimensional space. It will be apparent to one of ordinary skill in the art that many other scoring techniques may be used and implemented on individual scoring modules 135. In the preferred embodiment, the scoring modules 135 are implemented as dynamic link library (DLL) components, which can be dynamically linked to the system 100 using conventional methods.
Referring again to Figure 1, user interface 110 connects the feature manager 120 and scoring manager 130 segments to registry 150 and database 160 through storage interface 140. Registry 150 and database 160 are used to store system configuration information and information related to a library of searchable images. Referring now to Figure 6, the main components of registry 150 are illustrated. Registry 150 includes general configuration data 610. Because system 100 can include an arbitrary number of dynamically linked feature modules 125 and scoring modules 135, it is necessary to maintain a current count of the number of these modules installed in a system at a particular time. This information is stored in general configuration data 610. Further, other global information useful for any of the components of system 100, such as the number of images in database 160, are stored in general configuration data 610. It will be apparent to one of ordinary skill in the art that many items of such global information may be stored in the general configuration data 160 of a particular system 100. In general, registry 150 is used for the storage of information for locating and accessing feature modules 125, scoring modules 135, and images in database 160. As shown in Figure 6, registry 150 includes image pointers 630, which identify the location of each image in database 160. As each image is added to database 160, using a method described in more detail below, a pointer to the image is added to image pointers 630. One other basic function of registry 150 in the preferred embodiment, is the formation of a correspondence between the descriptive name given to a particular feature module and the dynamic link library (DLL) component used to actually implement the feature module. The descriptive feature names 620 are stored in registry 150 along with their corresponding feature module DLL identifiers 640. The descriptive feature names are names of feature extraction methods or image analysis tools displayed to a user for selection through user interface 110. The corresponding feature module DLL identifiers are the DLLs actually invoked when a corresponding descriptive feature name is invoked by the user. In the preferred embodiment of the present invention, the registry 150 is a text file written in the format of a Windows 3.1 .ini file. It will be apparent to those of ordinary skill in the art that other formats for storing registry information in an alternative embodiment may be used. Referring now to Figure 7, the main component of database 160 is illustrated. Database 160 is used for storage of image feature descriptors 710 for each image as generated by the feature modules 125. Image feature descriptors 710 includes a feature descriptor corresponding to each of the feature modules 125 installed in system 100. The detail of image feature descriptors 710 is illustrated in Figure 8. As shown in Figure 8, a plurality of feature descriptors corresponding to the installed feature modules 125, are stored for each image in database 160. The number of feature descriptors for each image varies depending on the number of feature modules 125 currently installed in system 100. A header is maintained for each image to assist in maintaining the linked list of feature descriptors for the image. As the feature descriptors for each image are stored in image feature descriptors 710, the actual image data represented in a bit-mapped form is stored in image data 720. Image data 720 represents the actual bit-mapped image data that may be used to display the image on a display device. In the preferred embodiment, this image data is stored as bit-mapped (.bmp) image data, however it will be apparent to one of ordinary skill in the art that the image data may be represented in any of a variety of conventional image coding techniques. A detail of the image data 720 is illustrated in Figure 8. A pointer to each bit-mapped image data block is maintained in registry 150. In this manner, the image data for each image in database 160 and the feature descriptors corresponding to the image may be readily located during the process of querying database 160. In the preferred embodiment of the present invention, the database 160 is written in a RIFF format, which is a conventional information format created by IBM Corporation and Microsoft Corporation for manipulation of video and multimedia files.
It will also be apparent to one of ordinary skill in the art that a plurality of databases 160 may be employed in a particular system. Using a plurality of databases 160, a set of images can be partitioned into separate databases and thereby searched more quickly. In support of multiple databases 160, a database identifier for each of the plurality of installed databases 160 is stored in registry 150. In a multiple database 160 configuration, the system must provide a database identifier during the process of adding an image to a database or during the process of querying an image against one or more databases of the system.
Referring now to Figure 9, the process used in the preferred embodiment of the present invention for adding an image to database 160 is illustrated. In the "add image" process of the present invention, a user provides a new image for entry into database 160 through user interface 110. Initially, the user identifies a data file containing the raw image data of the new image. This image data is retrieved from the identified file in processing block 200 illustrated in Figure 9. The retrieved image data is passed to feature manager 120. Feature manager 120 begins a loop in which the image data is passed to each of the previously installed feature modules 125. This loop process is illustrated in Figure 9 at bubble 205 and box 210. By invoking each of the installed feature modules in loop 205, feature manager 120 causes each of the feature modules 125 to produce a corresponding feature descriptor generated as a result of each feature module's analysis of the input image data. Once feature manager 120 has invoked each of the feature modules 125 and a corresponding feature descriptor for each module has been generated, feature manager 120 begins loop 215 illustrated in Figure 9. As described above, each of the feature modules 125 contain all of the information and processing methods for analyzing an input image for a particular feature. As such, it is possible for the feature descriptor generated by each feature module to be of varying length or content. Also, because the feature specific information is retained in feature modules 125, feature manager 120 cannot know the length or content of the feature descriptor generated by each feature module 125. For this reason, feature manager 120 executes loop 215 to request from each feature module 125 the length of the feature descriptor generated by each feature module (processing blocks 220).
After feature manager 120 has completed loop 215, the image data 720 for the new image is added to database 160 in processing block 225. Using well-known techniques, a data block of an arbitrary size may be requested and obtained from database 160. Once the image data 720 for the new image is written to database 160, the data blocks necessary for storage of the image feature descriptors 710 are prepared. First, the header 730 for the new image is written to database 160. Next, a loop 230 is initiated for writing each of the feature descriptors generated by feature modules 125 to database 160. Because the length of each feature descriptor is known from the processing preformed in blocks 220, the free data blocks of the appropriate length can be obtained from database 160. In loop 230 and processing blocks 235, each of the feature descriptors of the new image are written to the feature descriptor blocks 740 allocated for the new image. As described above, neither the database 160 nor the feature manager 120 are aware of the format or content of the feature descriptor written to feature descriptor blocks 740. Only the particular feature module 125 which generated the particular feature descriptor can decode and interpret the particular feature descriptor stored in database 160. Once each of the feature descriptors are written to database 160, processing block 240 illustrated in Figure 9 is executed to complete the "add image" process of the preferred embodiment of the present invention. In processing block 240, the registry 150 is updated to reflect the presence of a new image. For example, the general configuration data 610 is updated to advance the number of images counter by one. Additionally, the image pointers 630 are updated to reflect the presence and location of the new image in database 160. It will be apparent to one of ordinary skill in the art that other incidental data may need to be updated to reflect the addition of a new image to database 160. At the completion of this process, the new image and its corresponding feature descriptors are stored in database 160 and available for query by a user.
Referring now to Figure 10, the process used by the present invention for querying an image by content is illustrated. In the present invention, a user may provide or identify to system 100 through user interface 110 an image representing an input image to a query request. In general, system 100 will analyze the features of the input image and then search database 160 for images resident in database 160 (i.e. query images) having features most similar to the input image provided by the user. This query process is described in more detail in Figure 10. In processing block 300, an input image is identified by a user through user interface 110 and retrieved from the specified location. The input image data is provided by user interface 110 to feature manager 120. Feature manager 120 begins loop 305 in which the input image is sequentially passed to each of the feature modules 125. Each feature module analyzes the input image for each particular feature corresponding to each installed feature module 125 (processing blocks 310). In processing blocks 310, each feature module 125 generates a feature descriptor corresponding to the input image. Once loop 305 has been completed, a set of feature descriptors as generated by each feature module 125 will have been created. This set of feature descriptors for the input image (denoted the input image feature descriptors) represents the feature information which will be compared with corresponding feature information for each of the images in the database (denoted query image feature descriptors). In processing block 315, the query image feature descriptors are retrieved from database 160. In one embodiment of the present invention, the feature descriptors for all of the images resident in database 160 are retrieved for comparison with the input image feature descriptors. In another embodiment of the present invention, the user may specify a subset of the images resident in database 160 for the purpose of restricting the query to a smaller domain of images. For example, a subset of database 160 resident images may be specified using textual descriptive information or other information corresponding to a particular classification of the images within database 160. In another embodiment, a set of query images from multiple databases may be specified. In any case, the query image feature descriptors for a set of images from database 160 that will be used as a query domain are retrieved in processing block 315. It is now necessary to compare the input image feature descriptors to each of the query image feature descriptors. Because only the feature modules 125 can decode and interpret the feature descriptor contents, the input image feature descriptors and the query image feature descriptors must be passed to the corresponding feature module of the set of feature modules 125. Thus, in loop 320, feature manager 120 passes the input image feature descriptors and the query image feature descriptors to each of the feature modules 125. Referring again to Figure 2, the input image feature descriptors are stored in image storage 410 and the query image feature descriptors are stored in query image storage 420. It will be apparent to one of ordinary skill in the art that these feature descriptors may be stored on each feature module 125 or alternatively stored in a central location, such as within feature manager 120 or in another component for processing by a particular feature module. In processing blocks 325, feature manager 120 invokes each of the feature modules 125 and passes to each feature module the input image feature descriptor and the query image feature descriptor for the corresponding feature module. Each feature module 125 sequentially processes the feature descriptors and compares the input image feature descriptor with the query image feature descriptor. As a result of this comparison, a score is generated by the feature module to indicate the level of similarity or dissimilarity between the input image and the query image. The feature score generator for 50 as shown is Figure 2 compares the input feature descriptor and the query feature descriptor for the particular feature and generates a score. The score for the particular feature comparison is stored in feature score storage 430 as shown in Figure 2. Once loop 320 has cycled through each of the feature modules 125, the decision block 330 is executed. In decision block 330, the next query image feature descriptors are retrieved from database 160. If all of the query images have been processed from database 160, processing passes from decision block 330 to loop 335 shown in Figure 10. If there are still more images in database 160 (or the query domain) to be processed for the query, processing passes from decision block 330 to processing block 315 where the feature descriptors for the next query image are obtained from database 160 and the new query image feature descriptors are processed through each of the feature modules 125 via loop 320. When each of the query images from database 160 have been compared with the input image for each of the features modules 125 and corresponding scores have been generated, loop 335 is executed. In loop 335, feature manager 120 sequentially cycles through each of the feature modules 125 to obtain the scoring information generated as a result of the feature module's comparison of the input image to each of the query images. In the preferred embodiment, this scoring information is obtained from feature score storage 430 shown in Figure 2. Once feature manager 120 has obtained scoring information from each of the feature modules 125 (processing blocks 340), processing passes to loop 345 shown in Figure 10. In loop 345, the scoring information obtained by feature manager 120 from each of the feature modules 125 is passed through user interface 110 to scoring manager 130. It will be apparent to one of ordinary skill in the art that the scoring information in an alternative embodiment may equivalently be passed directly from feature manager 120 to a scoring manager 130. Scoring manager 130 initiates loop 345. In loop 345, scoring manager 130 passes to scoring modules 135 the scoring information generated by each of the feature modules 125 for each of the query images from database 160. Using this information, scoring modules 135 compute the similarities between the input image and each of the query images (processing blocks 350). Because a variety of different scoring techniques may be used for a particular implementation, it is convenient to provide modular, and dynamically linkable scoring modules 135 which can each implement one of the variety of scoring techniques. Each of these installed scoring techniques as implemented on scoring modules 135 are invoked by scoring manager 130 sequentially as loop 345 is executed. The scoring modules 135 generate the similarities in scoring information between the input image and each of the query images (processing blocks 350). Once the similarities between the input image and the query images is generated in processing blocks 350, scoring manager 130 enters loop 355 where each of the scoring modules 135 are again invoked sequentially to compute the final scores which rank the query images in an order corresponding to their similarity to the input image (processing blocks 360). Having ordered the query images according to their similarity with the input image, the most similar images may then be displayed to the user through user interface 110 at processing block 365.
Referring now to Figure 11, the steps performed in the present invention for registering a feature module or scoring module are illustrated. In processing block 1110, the user invokes an "Add Feature" or an "Add Scoring Method" command through user interface 110. This command directs the user through a series of dialog boxes and input screens to obtain the information necessary for registering the new feature module or scoring module. In processing block 1120, the user specifies a descriptive name of a particular feature or feature extraction method. The user also specifies a descriptive name corresponding to the scoring method to be added to system 100. The descriptive names of the new feature or scoring method are displayed in various user interface 110 menus or information windows. A dynamic link library (DLL) module corresponding to the descriptive name of the feature or scoring method is also provided by the user in processing block 1120. The DLL corresponds to the executable software necessary for implementing the feature extraction method or scoring method. The user specifies the location of the DLL, typically by specifying the path or file name corresponding to the DLL in a particular file system. The user-entered descriptive name and DLL identifier is added to the registry 150 in processing block 1130. Finally, the user specified DLL is linked to the linked list of feature modules or scoring modules in processing block 1140. In this manner, a feature module 125 or scoring module 135 may be dynamically linked with system 100 and thereafter be employed for analyzing input images.
Figure 12 illustrates a typical data processing system upon which one embodiment of the present invention is implemented. It will be apparent to those of ordinary skill in the art, however that other alternative systems of various system architectures may also be used. The data processing system illustrated in Figure 12 includes a bus or other internal communication means 801 for communicating information, and a processor 802 coupled to the bus 801 for processing information. The system further comprises a random access memory (RAM) or other volatile storage device 804 (referred to as main memory), coupled to bus 801 for storing information and instructions to be executed by processor 802. Main memory 804 also may be used for storing temporary variables or other intermediate information during execution of instructions by processor 802. The system also comprises a read only memory (ROM) and /or static storage device 806 coupled to bus 801 for storing static information and instructions for processor 802, and a data storage device 807 such as a magnetic disk or optical disk and its corresponding disk drive. Mass data storage device 807 is coupled to bus 801 for storing information and instructions. The system may further be coupled to a display device 821, such as a cathode ray tube (CRT) or a liquid crystal display (LCD) coupled to bus 801 through bus 803 for displaying information to a computer user. An alphanumeric input device 822, including alphanumeric and other keys, may also be coupled to bus 801 through bus 803 for communicating information and command selections to processor 802. An additional user input device is cursor control 823, such as a mouse, a trackball, stylus, or cursor direction keys coupled to bus 801 through bus 803 for communicating direction information and command selections to processor 802, and for controlling cursor movement on display device 821. Another device which may optionally be coupled to bus 801 through bus 803 is a hard copy device 824 which may be used for printing instructions, data, or other information on a medium such as paper, film, or similar types of media. In the preferred embodiment, a communication device 826 is coupled to bus 801 through bus 803 for use in accessing other nodes of a distributed system via a network. This communication device 826 may include any of a number of commercially available networking peripheral devices such as those used for coupling to an Ethernet, token ring, Internet, or wide area network. Note that any or all of the components of the system illustrated in Figure 12 and associated hardware may be used in various embodiments of the present invention; however, it will be appreciated by those of ordinary skill in the art that any configuration of the system may be used for various purposes according to the particular implementation. In one embodiment of the present invention, the data processing system illustrated in Figure 1 is an
IBM® compatible personal computer or a Sun® SPARC Workstation. Processor 102 may be one of the X86 compatible microprocessors such as the PENTIUM® brand microprocessors manufactured by INTEL® Corporation of Santa Clara, California upon which a conventional operating system such as the Windows 95 brand operating system developed by Microsoft Corporation of Redmond, Washington is executed.
The control logic or software implementing the present invention can be stored in main memory 804, mass storage device 807, or other storage medium locally accessible to processor 802. Other storage media may include floppy disk drives, memory cards, flash memory, or CD-ROM drives. It will be apparent to those of ordinary skill in the art that the methods and processes described herein can be implemented as software stored in main memory 804 or read only memory 806 and executed by processor 802. This control logic or software may also be resident on an article of manufacture comprising a computer readable medium 808 having computer readable program code embodied therein and being readable by the mass storage device 807 and for causing the processor 802 to coordinate accesses to a storage system in accordance with the teachings herein.
Thus, a configurable and modular image query system is described. Although the present has been described with reference to specific exemplary embodiments, it will be evident that various modifications and alterations may be made to these embodiments without departing from the broader spirit and scope of the invention as set forth in the claims below. Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense.

Claims

CLAIMSWhat is claimed is:
1. An image query and storage apparatus comprising: a plurality of dynamically linkable feature modules (125), each of said plurality of feature modules extracting a different set of feature information from an input image; and a database (160) coupled to said plurality of feature modules, said database including storage for said different set of feature information for each of said plurality of feature modules.
2. The apparatus as claimed in Claim 1 wherein the plurality of feature modules further include logic for generating feature specific scoring information.
3. The apparatus as claimed in Claim 1 further including a user interface coupled to the plurality of feature modules and the database for providing the input image to the plurality of feature modules.
4. The apparatus as claimed in Claim 1 further including a registry coupled to the database for storing registration information related to the plurality of feature modules.
5. The apparatus as claimed in Claim 2 further including a plurality of dynamically linkable scoring modules, each of said plurality of scoring modules generating a different set of scoring information from the feature specific scoring information generated by the feature modules.
PCT/US1999/024094 1998-10-19 1999-10-15 Modular image query system WO2000023918A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU12065/00A AU1206500A (en) 1998-10-19 1999-10-15 Modular image query system

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US09/175,155 1998-10-19
US09/175,155 US6445834B1 (en) 1998-10-19 1998-10-19 Modular image query system

Publications (2)

Publication Number Publication Date
WO2000023918A1 WO2000023918A1 (en) 2000-04-27
WO2000023918A9 true WO2000023918A9 (en) 2000-09-08

Family

ID=22639156

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1999/024094 WO2000023918A1 (en) 1998-10-19 1999-10-15 Modular image query system

Country Status (3)

Country Link
US (1) US6445834B1 (en)
AU (1) AU1206500A (en)
WO (1) WO2000023918A1 (en)

Families Citing this family (94)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7016916B1 (en) * 1999-02-01 2006-03-21 Lg Electronics Inc. Method of searching multimedia data
US6411724B1 (en) * 1999-07-02 2002-06-25 Koninklijke Philips Electronics N.V. Using meta-descriptors to represent multimedia information
GB2388761B (en) * 1999-07-09 2004-01-21 Korea Electronics Telecomm Texture description method and texture-based retrieval method in frequency domain
US6718067B1 (en) * 1999-07-30 2004-04-06 Sony Corporation Method of manipulating illumination effects associated with an image
US7058647B1 (en) 1999-08-31 2006-06-06 Charles E. Hill & Associates Electronic presentation generation system and method
US6647141B1 (en) * 1999-11-18 2003-11-11 Sony Corporation User interface of query by color statistics
JP4433236B2 (en) * 1999-12-03 2010-03-17 ソニー株式会社 Information processing apparatus, information processing method, and program recording medium
US6647156B1 (en) * 2000-06-14 2003-11-11 Chipdata, Inc. Data retrieval method and apparatus
JP2002007432A (en) * 2000-06-23 2002-01-11 Ntt Docomo Inc Information retrieval system
US7433881B1 (en) * 2000-08-31 2008-10-07 Charles E. Hill & Associates, Inc. System and method for handling multi-resolution graphics files
US7680324B2 (en) 2000-11-06 2010-03-16 Evryx Technologies, Inc. Use of image-derived information as search criteria for internet and other search engines
US7565008B2 (en) 2000-11-06 2009-07-21 Evryx Technologies, Inc. Data capture and identification system and process
US9310892B2 (en) 2000-11-06 2016-04-12 Nant Holdings Ip, Llc Object information derived from object images
US7899243B2 (en) 2000-11-06 2011-03-01 Evryx Technologies, Inc. Image capture and identification system and process
US8224078B2 (en) 2000-11-06 2012-07-17 Nant Holdings Ip, Llc Image capture and identification system and process
US6629104B1 (en) * 2000-11-22 2003-09-30 Eastman Kodak Company Method for adding personalized metadata to a collection of digital images
KR100788643B1 (en) * 2001-01-09 2007-12-26 삼성전자주식회사 Searching method of image based on combination of color and texture
US20030063575A1 (en) * 2001-09-28 2003-04-03 Fuji Photo Film Co., Ltd. Order processing apparatus, order processing system and image photographing device
JP2003281153A (en) * 2002-03-27 2003-10-03 Minolta Co Ltd Server and terminal device for information retrieval system
US7240075B1 (en) * 2002-09-24 2007-07-03 Exphand, Inc. Interactive generating query related to telestrator data designating at least a portion of the still image frame and data identifying a user is generated from the user designating a selected region on the display screen, transmitting the query to the remote information system
JP2004151979A (en) * 2002-10-30 2004-05-27 Olympus Corp System for automated preparation of index for electronic catalog
AU2003903728A0 (en) * 2003-07-21 2003-07-31 Hao Hang Zheng Method and system for performing combined image classification storage and rapid retrieval on a computer database
US7672986B2 (en) * 2004-02-12 2010-03-02 Microsoft Corporation Managing graphic databases
JP2005267609A (en) * 2004-02-20 2005-09-29 Fuji Photo Film Co Ltd Digital picture book system, and method and program for searching picture book
US7624123B2 (en) * 2004-02-26 2009-11-24 Ati Technologies, Inc. Image processing system and method
JP2005277981A (en) * 2004-03-26 2005-10-06 Seiko Epson Corp Target image selection for image processing
US7590310B2 (en) 2004-05-05 2009-09-15 Facet Technology Corp. Methods and apparatus for automated true object-based image analysis and retrieval
US8563133B2 (en) * 2004-06-08 2013-10-22 Sandisk Corporation Compositions and methods for modulation of nanostructure energy levels
US7487072B2 (en) 2004-08-04 2009-02-03 International Business Machines Corporation Method and system for querying multimedia data where adjusting the conversion of the current portion of the multimedia data signal based on the comparing at least one set of confidence values to the threshold
US7929056B2 (en) * 2004-09-13 2011-04-19 Hewlett-Packard Development Company, L.P. User interface with tiling of video sources, widescreen modes or calibration settings
US7813552B2 (en) * 2004-09-23 2010-10-12 Mitsubishi Denki Kabushiki Kaisha Methods of representing and analysing images
US8600989B2 (en) 2004-10-01 2013-12-03 Ricoh Co., Ltd. Method and system for image matching in a mixed media environment
US8521737B2 (en) * 2004-10-01 2013-08-27 Ricoh Co., Ltd. Method and system for multi-tier image matching in a mixed media environment
US9405751B2 (en) 2005-08-23 2016-08-02 Ricoh Co., Ltd. Database for mixed media document system
US7702673B2 (en) 2004-10-01 2010-04-20 Ricoh Co., Ltd. System and methods for creation and use of a mixed media environment
US8838591B2 (en) 2005-08-23 2014-09-16 Ricoh Co., Ltd. Embedding hot spots in electronic documents
US8825682B2 (en) 2006-07-31 2014-09-02 Ricoh Co., Ltd. Architecture for mixed media reality retrieval of locations and registration of images
US8868555B2 (en) 2006-07-31 2014-10-21 Ricoh Co., Ltd. Computation of a recongnizability score (quality predictor) for image retrieval
US7812986B2 (en) 2005-08-23 2010-10-12 Ricoh Co. Ltd. System and methods for use of voice mail and email in a mixed media environment
US8965145B2 (en) 2006-07-31 2015-02-24 Ricoh Co., Ltd. Mixed media reality recognition using multiple specialized indexes
US9384619B2 (en) 2006-07-31 2016-07-05 Ricoh Co., Ltd. Searching media content for objects specified using identifiers
US8510283B2 (en) 2006-07-31 2013-08-13 Ricoh Co., Ltd. Automatic adaption of an image recognition system to image capture devices
US9171202B2 (en) 2005-08-23 2015-10-27 Ricoh Co., Ltd. Data organization and access for mixed media document system
US8156116B2 (en) 2006-07-31 2012-04-10 Ricoh Co., Ltd Dynamic presentation of targeted information in a mixed media reality recognition system
US8856108B2 (en) 2006-07-31 2014-10-07 Ricoh Co., Ltd. Combining results of image retrieval processes
US8176054B2 (en) 2007-07-12 2012-05-08 Ricoh Co. Ltd Retrieving electronic documents by converting them to synthetic text
US8989431B1 (en) 2007-07-11 2015-03-24 Ricoh Co., Ltd. Ad hoc paper-based networking with mixed media reality
US9373029B2 (en) 2007-07-11 2016-06-21 Ricoh Co., Ltd. Invisible junction feature recognition for document security or annotation
US8949287B2 (en) 2005-08-23 2015-02-03 Ricoh Co., Ltd. Embedding hot spots in imaged documents
US9530050B1 (en) 2007-07-11 2016-12-27 Ricoh Co., Ltd. Document annotation sharing
US7882057B1 (en) * 2004-10-04 2011-02-01 Trilogy Development Group, Inc. Complex configuration processing using configuration sub-models
US7574409B2 (en) * 2004-11-04 2009-08-11 Vericept Corporation Method, apparatus, and system for clustering and classification
KR100708130B1 (en) * 2005-05-04 2007-04-17 삼성전자주식회사 Apparatus and method for extracting moving image
US9020955B2 (en) * 2005-05-12 2015-04-28 Hewlett-Packard Development Company, L.P. Compositional balance driven content retrieval
JP4542013B2 (en) * 2005-09-29 2010-09-08 富士フイルム株式会社 Print order system, program, program storage medium, and print order server
US7644373B2 (en) 2006-01-23 2010-01-05 Microsoft Corporation User interface for viewing clusters of images
US7836050B2 (en) * 2006-01-25 2010-11-16 Microsoft Corporation Ranking content based on relevance and quality
US7814040B1 (en) 2006-01-31 2010-10-12 The Research Foundation Of State University Of New York System and method for image annotation and multi-modal image retrieval using probabilistic semantic models
TWI403912B (en) * 2006-06-08 2013-08-01 Univ Nat Chiao Tung Method and system of image retrieval
US8676810B2 (en) 2006-07-31 2014-03-18 Ricoh Co., Ltd. Multiple index mixed media reality recognition using unequal priority indexes
US8201076B2 (en) * 2006-07-31 2012-06-12 Ricoh Co., Ltd. Capturing symbolic information from documents upon printing
US9020966B2 (en) 2006-07-31 2015-04-28 Ricoh Co., Ltd. Client device for interacting with a mixed media reality recognition system
US9063952B2 (en) 2006-07-31 2015-06-23 Ricoh Co., Ltd. Mixed media reality recognition with image tracking
US9176984B2 (en) 2006-07-31 2015-11-03 Ricoh Co., Ltd Mixed media reality retrieval of differentially-weighted links
US8489987B2 (en) 2006-07-31 2013-07-16 Ricoh Co., Ltd. Monitoring and analyzing creation and usage of visual content using image and hotspot interaction
US20080082559A1 (en) * 2006-09-28 2008-04-03 Gm Global Technology Operations, Inc. Method of linking information to an electronically enabled manufactured part archive
US7707208B2 (en) * 2006-10-10 2010-04-27 Microsoft Corporation Identifying sight for a location
US7657504B2 (en) * 2006-10-10 2010-02-02 Microsoft Corporation User interface for displaying images of sights
US8862582B2 (en) * 2007-11-15 2014-10-14 At&T Intellectual Property I, L.P. System and method of organizing images
US8200025B2 (en) 2007-12-07 2012-06-12 University Of Ottawa Image classification and search
US8433140B2 (en) * 2009-11-02 2013-04-30 Microsoft Corporation Image metadata propagation
US9710491B2 (en) * 2009-11-02 2017-07-18 Microsoft Technology Licensing, Llc Content-based image search
US20110106798A1 (en) * 2009-11-02 2011-05-05 Microsoft Corporation Search Result Enhancement Through Image Duplicate Detection
US8706812B2 (en) 2010-04-07 2014-04-22 On24, Inc. Communication console with component aggregation
US11438410B2 (en) 2010-04-07 2022-09-06 On24, Inc. Communication console with component aggregation
KR101116434B1 (en) * 2010-04-14 2012-03-07 엔에이치엔(주) System and method for supporting query using image
US9026034B2 (en) 2010-05-04 2015-05-05 Project Oda, Inc. Automatic detection of broadcast programming
SG189840A1 (en) 2010-10-19 2013-06-28 3M Innovative Properties Co Computer-aided assignment of ratings to digital samples of a manufactured web product
EP2638384A4 (en) 2010-11-12 2017-04-19 3M Innovative Properties Company Rapid processing and detection of non-uniformities in web-based materials
KR101675785B1 (en) * 2010-11-15 2016-11-14 삼성전자주식회사 Method and apparatus for image searching using feature point
US8639034B2 (en) * 2010-11-19 2014-01-28 Ricoh Co., Ltd. Multimedia information retrieval system with progressive feature selection and submission
US9870376B2 (en) * 2011-04-01 2018-01-16 Excalibur Ip, Llc Method and system for concept summarization
US8732739B2 (en) 2011-07-18 2014-05-20 Viggle Inc. System and method for tracking and rewarding media and entertainment usage including substantially real time rewards
US9058331B2 (en) 2011-07-27 2015-06-16 Ricoh Co., Ltd. Generating a conversation in a social network based on visual search results
US11068532B2 (en) 2011-09-21 2021-07-20 Horsetooth Ventures, LLC Interactive image display and selection system
US9734167B2 (en) 2011-09-21 2017-08-15 Horsetooth Ventures, LLC Interactive image display and selection system
US9183466B2 (en) 2013-06-15 2015-11-10 Purdue Research Foundation Correlating videos and sentences
ES2530687B1 (en) * 2013-09-04 2016-08-19 Shot & Shop. S.L. Method implemented by computer for image recovery by content and computer program of the same
US11429781B1 (en) 2013-10-22 2022-08-30 On24, Inc. System and method of annotating presentation timeline with questions, comments and notes using simple user inputs in mobile devices
US10785325B1 (en) 2014-09-03 2020-09-22 On24, Inc. Audience binning system and method for webcasting and on-line presentations
US10212240B2 (en) * 2015-04-22 2019-02-19 Samsung Electronics Co., Ltd. Method for tracking content and electronic device using the same
US20170263031A1 (en) * 2016-03-09 2017-09-14 Trendage, Inc. Body visualization system
US11188822B2 (en) 2017-10-05 2021-11-30 On24, Inc. Attendee engagement determining system and method
US11281723B2 (en) * 2017-10-05 2022-03-22 On24, Inc. Widget recommendation for an online event using co-occurrence matrix

Family Cites Families (58)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS59183458A (en) 1983-04-01 1984-10-18 Hitachi Ltd Picture retrieving system
JPS60196856A (en) 1984-03-20 1985-10-05 Olympus Optical Co Ltd Picture retrieval registering system
US4646250A (en) 1984-10-18 1987-02-24 International Business Machines Corp. Data entry screen
JPS6274181A (en) 1985-09-27 1987-04-04 Sony Corp Character recognizing device
US4829453A (en) 1987-03-05 1989-05-09 Sharp Kabushiki Kaisha Apparatus for cataloging and retrieving image data
US4739407A (en) * 1986-03-27 1988-04-19 Hekimian Laboratories, Inc. Frequency modulation television receiver with tracking filter
JP2735187B2 (en) 1987-03-17 1998-04-02 株式会社東芝 Information search method
US4944023A (en) 1987-05-19 1990-07-24 Ricoh Company, Ltd. Method of describing image information
CA1318977C (en) 1987-07-22 1993-06-08 Kazuhito Hori Image recognition system
JPH087553B2 (en) 1988-10-27 1996-01-29 インターナショナル・ビジネス・マシーンズ・コーポレーション Color image quantization method and apparatus
JP2993673B2 (en) 1989-01-27 1999-12-20 株式会社日立製作所 Electronic file device
US5220648A (en) 1989-05-31 1993-06-15 Kabushiki Kaisha Toshiba High-speed search system for image data storage
US5179652A (en) 1989-12-13 1993-01-12 Anthony I. Rozmanith Method and apparatus for storing, transmitting and retrieving graphical and tabular data
US5012334B1 (en) 1990-01-29 1997-05-13 Grass Valley Group Video image bank for storing and retrieving video image sequences
GB9012273D0 (en) 1990-06-01 1990-07-18 Rediffusion Simulation Ltd Image generator
JP3141245B2 (en) 1990-11-01 2001-03-05 ソニー株式会社 How to display images
US5202828A (en) 1991-05-15 1993-04-13 Apple Computer, Inc. User interface system having programmable user interface elements
JPH04343178A (en) 1991-05-20 1992-11-30 Sony Corp Image processor
CA2069976A1 (en) 1991-05-31 1992-12-01 Shin-Ichi Nakazawa Retrieval system for graphic information
US5381158A (en) 1991-07-12 1995-01-10 Kabushiki Kaisha Toshiba Information retrieval apparatus
US5249056A (en) 1991-07-16 1993-09-28 Sony Corporation Of America Apparatus for generating video signals from film
US5481620A (en) 1991-09-27 1996-01-02 E. I. Du Pont De Nemours And Company Adaptive vision system
US5421008A (en) 1991-11-08 1995-05-30 International Business Machines Corporation System for interactive graphical construction of a data base query and storing of the query object links as an object
WO1993011484A2 (en) * 1991-11-27 1993-06-10 Telefonaktiebolaget Lm Ericsson Software structure for telecommunication switching systems
US5544284A (en) 1992-02-11 1996-08-06 Eastman Kodak Company Sequential product code quantization of digital color image
GB2267633B (en) 1992-06-04 1996-03-06 Sony Broadcast & Communication Analyzing image data processing operations
JP3252381B2 (en) 1992-09-08 2002-02-04 ソニー株式会社 Pattern recognition device
GB2271261A (en) 1992-10-02 1994-04-06 Canon Res Ct Europe Ltd Processing image data
US5448631A (en) * 1992-10-13 1995-09-05 U S West Advanced Technologies, Inc. Apparatus for handling features in a telephone network
US5664080A (en) 1992-10-20 1997-09-02 International Business Machines Corporation System and method for generating a universal palette and mapping an original color space to the universal palette
US5579471A (en) 1992-11-09 1996-11-26 International Business Machines Corporation Image query system and method
US6182069B1 (en) * 1992-11-09 2001-01-30 International Business Machines Corporation Video query system and method
US5404396A (en) * 1993-08-27 1995-04-04 Telefonaktiebolaget Lm Ericsson Feature interaction manager
JP3234064B2 (en) 1993-09-02 2001-12-04 キヤノン株式会社 Image retrieval method and apparatus
JP3422026B2 (en) 1993-10-04 2003-06-30 ソニー株式会社 Audio player
US5704013A (en) 1994-09-16 1997-12-30 Sony Corporation Map determination method and apparatus
US5659626A (en) 1994-10-20 1997-08-19 Calspan Corporation Fingerprint identification system
US5793888A (en) 1994-11-14 1998-08-11 Massachusetts Institute Of Technology Machine learning apparatus and method for image searching
US5530869A (en) * 1994-11-30 1996-06-25 Salle; Donald A. System and method for locating and implementing commands without invoking a menu hierarchy
US5729471A (en) 1995-03-31 1998-03-17 The Regents Of The University Of California Machine dynamic selection of one video camera/image of a scene from multiple video cameras/images of the scene in accordance with a particular perspective on the scene, an object in the scene, or an event in the scene
US5767893A (en) 1995-10-11 1998-06-16 International Business Machines Corporation Method and apparatus for content based downloading of video programs
US5794249A (en) 1995-12-21 1998-08-11 Hewlett-Packard Company Audio/video retrieval system that uses keyword indexing of digital recordings to display a list of the recorded text files, keywords and time stamps associated with the system
US5915250A (en) * 1996-03-29 1999-06-22 Virage, Inc. Threshold-based comparison
US5913205A (en) 1996-03-29 1999-06-15 Virage, Inc. Query optimization for visual information retrieval system
US5983237A (en) * 1996-03-29 1999-11-09 Virage, Inc. Visual dictionary
US5893095A (en) 1996-03-29 1999-04-06 Virage, Inc. Similarity engine for content-based retrieval of images
US5911139A (en) 1996-03-29 1999-06-08 Virage, Inc. Visual image database search engine which allows for different schema
US5835099A (en) 1996-06-26 1998-11-10 Xerox Corporation Representing a region of a color image using a space-color separable model
US6035057A (en) 1997-03-10 2000-03-07 Hoffman; Efrem H. Hierarchical data matrix pattern recognition and identification system
US6016487A (en) 1997-03-26 2000-01-18 National Research Council Of Canada Method of searching three-dimensional images
US5966434A (en) * 1997-03-31 1999-10-12 Telcordia Technologies, Inc. System and method for managing feature interaction of telephone services
US6120299A (en) * 1997-06-06 2000-09-19 Educational Testing Service System and method for interactive scoring of standardized test responses
US6269368B1 (en) * 1997-10-17 2001-07-31 Textwise Llc Information retrieval using dynamic evidence combination
US5987456A (en) * 1997-10-28 1999-11-16 University Of Masschusetts Image retrieval by syntactic characterization of appearance
US6035055A (en) * 1997-11-03 2000-03-07 Hewlett-Packard Company Digital image management system in a distributed data access network system
US6016494A (en) * 1997-11-21 2000-01-18 International Business Machines Corporation Expanding web documents by merging with linked documents
US6175829B1 (en) * 1998-04-22 2001-01-16 Nec Usa, Inc. Method and apparatus for facilitating query reformulation
US6256596B1 (en) * 1998-06-30 2001-07-03 Autodesk, Inc. Extensible framework for capturing feature information in a neutral format

Also Published As

Publication number Publication date
US6445834B1 (en) 2002-09-03
WO2000023918A1 (en) 2000-04-27
AU1206500A (en) 2000-05-08

Similar Documents

Publication Publication Date Title
US6445834B1 (en) Modular image query system
US7015931B1 (en) Method and apparatus for representing and searching for color images
KR100491801B1 (en) Method of and apparatus for identifying subsets of interrelated image objects from a set of image objects
Taycher et al. Image digestion and relevance feedback in the ImageRover WWW search engine
Samet et al. Marco: Map retrieval by content
EP0722145B1 (en) Information retrieval system and method of operation
CA2307264C (en) An interactive framework for understanding user's perception of multimedia data
US6804684B2 (en) Method for associating semantic information with multiple images in an image database environment
US6072904A (en) Fast image retrieval using multi-scale edge representation of images
JP3522146B2 (en) Image matching method and image inquiry device
Park et al. Color image retrieval using hybrid graph representation
US20030145277A1 (en) Interactively comparing records in a database
JPH11224346A (en) Sorting method for document image
JP2001202523A (en) Method and device for image processing
JPH10154149A (en) Similar object retrieval method and device therefor
Lee et al. Indexing for complex queries on a query-by-content image database
Leung et al. Benchmarking for content-based visual information search
Takahashi et al. Image retrieval using spatial intensity features
Bali et al. Clustering Technique Approach to Detect the Special Patterns for Medical Video Mining
Barros et al. Indexing multispectral images for content-based retrieval
Barros et al. System for indexing multispectral satellite images for efficient content-based retrieval
JPH08287086A (en) Method and device for emphasizing and displaying image in order of adaptability
Ali Content-based image classification and retrieval: A rule-based system using rough sets framework
US6671402B1 (en) Representing an image with weighted joint histogram
King et al. A feature-based image retrieval database for the fashion, textile, and clothing industry in Hong Kong

Legal Events

Date Code Title Description
ENP Entry into the national phase

Ref country code: AU

Ref document number: 2000 12065

Kind code of ref document: A

Format of ref document f/p: F

AK Designated states

Kind code of ref document: A1

Designated state(s): AE AL AM AT AU AZ BA BB BG BR BY CA CH CN CR CU CZ DE DK DM EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
AK Designated states

Kind code of ref document: C2

Designated state(s): AE AL AM AT AU AZ BA BB BG BR BY CA CH CN CR CU CZ DE DK DM EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: C2

Designated state(s): GH GM KE LS MW SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

COP Corrected version of pamphlet

Free format text: PAGES 1/11-11/11, DRAWINGS, REPLACED BY NEW PAGES 1/10-10/10; DUE TO LATE TRANSMITTAL BY THE RECEIVING OFFICE

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

122 Ep: pct application non-entry in european phase