WO2001011489A9 - Retrieving images by defining an image content comprising objects - Google Patents
Retrieving images by defining an image content comprising objectsInfo
- Publication number
- WO2001011489A9 WO2001011489A9 PCT/US2000/021735 US0021735W WO0111489A9 WO 2001011489 A9 WO2001011489 A9 WO 2001011489A9 US 0021735 W US0021735 W US 0021735W WO 0111489 A9 WO0111489 A9 WO 0111489A9
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- objects
- images
- database
- filtered
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/53—Querying
- G06F16/532—Query formulation, e.g. graphical querying
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/20—ICT specially adapted for the handling or processing of medical images for handling medical images, e.g. DICOM, HL7 or PACS
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/40—ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S707/00—Data processing: database and file management or data structures
- Y10S707/99931—Database or file accessing
- Y10S707/99933—Query processing, i.e. searching
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S707/00—Data processing: database and file management or data structures
- Y10S707/99931—Database or file accessing
- Y10S707/99933—Query processing, i.e. searching
- Y10S707/99936—Pattern matching access
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S707/00—Data processing: database and file management or data structures
- Y10S707/99941—Database schema or data structure
- Y10S707/99944—Object-oriented database structure
- Y10S707/99945—Object-oriented database structure processing
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S707/00—Data processing: database and file management or data structures
- Y10S707/99941—Database schema or data structure
- Y10S707/99948—Application of database or data structure, e.g. distributed, multimedia, or image
Definitions
- the invention relates to management of an image database and the classification and retrieval of images therefrom.
- images are digitized and stored in a database in association with one or more keywords which describe their content or character.
- a database can be searched linguistically for particular keywords, and images which are associated with these keywords are retrieved in response.
- one or more "query images” are utilized, and images from the database which are in some sense similar to the query image are located and retrieved.
- the pixel values of the query image and the images in the database are processed to produce a set of parameters indicative of color distribution, pixel intensity variation across the image, as well as other characteristics of the image as a whole (color-shape-texture approach).
- These parameters are calculated using various image filtering and processing techniques so as to produce a vector of feature parameters which is indicative of the image itself.
- image imposing techniques such as, for example, tiling (when the image is artificially divided into a multiple sections using overlaying grid).
- the comparison process involves comparing feature vectors from images in the database with a query feature vector, and images from the database having similar feature vectors are retrieved.
- a system of this nature is described in U.S. Patent No. 5,644,765 to Shimura et al., the disclosure of which is hereby incorporated by reference in its entirety, as well as compiled book of the latest achievements in image retrieval Image Description and Retrieval, editor Enrico Vicario (University of Florence, Italy) and published in 1999 by Plenum Press .
- the invention comprises methods and systems for processing, comparing, classifying and retrieving images.
- the invention comprises a method of identifying similarities between first and second imaged structures present in one or more digital images.
- the method includes processing one or more digital images so as to define at least first group of objects (could be only one object) and second group of objects (could be only one object), assigning a first group of objects characterization parameter set to the first group of objects, and assigning a second object characterization parameter set to the second group of objects.
- the method further comprises calculating a similarity index based on the first group of objects characterization parameter set and the second object characterization parameter set.
- the invention comprises a method of classifying and retrieving digital images having content similar to the content of a query image.
- the invention comprises filtering a query image with a modifiable filtering function to produce a user selected filtered query image, substantially continuously displaying the filtered query image as the filtering functions are applied to produce the filtered query image, defining or detecting manually or automatically portions of query image, comparing at least a portion of the filtered query image to at least portions of images in an image database, and selecting one or more images in the image database or portions thereof which are similar to the at least a portion of the filtered query image or images.
- Image classification and retrieval systems are also provided.
- a retrieval system comprises an image database comprising a plurality of images, at least one query image stored in a memory, and a user definable filter set receiving both the query image and receiving images from the image database as inputs.
- the filter set is configured to provide a filtered query image and filtered database images as outputs to an object definition module in data communication with the filter.
- the object definition module is configured to define objects within the filtered query image and the filtered database images.
- the system also includes an object parameterization module configured to calculate a parameter set associated with at least one object in the filtered query image and a parameter set associated with at least one object in the filtered database images.
- a parameter set comparison module receives the parameter sets as inputs, and is configured to compare the content of the parameter sets.
- such a classification system comprises an image database comprising a plurality of images clustered into classes of images, at least one query image stored in a memory, and a user definable filter set receiving both the query image and receiving images from the image database as inputs.
- the filter set is configured to provide a filtered query image and filtered database images as outputs to an object definition module in data communication with the filter.
- the object definition module is configured to define objects within the filtered query image and the filtered database images.
- the system also includes an object parameterization module configured to calculate a parameter set associated with at least one object in the filtered query image and a parameter set associated with at least one object in the filtered database images.
- a parameter set comparison module receives the parameter sets as inputs, and is configured to compare the content of the parameter sets and then to assign query image or portion thereof as belonging to a class.
- Another system in accordance with the invention includes an image filter having a raw image as an input and a filtered image as an output and an image display device receiving the filtered image output from the image filter.
- An input device provides input parameters for modifying a filtering function performed by the image filter, whereby an image appearing on the image display device is updated substantially continuously as the filtering function is modified.
- the system also includes a database of images and an image comparison module configured to compare the filtered image output with images in the database and to select images in the database which are similar to the filtered image output. Filtering and image enhancement processes can be automatic and use system calculated default parameters.
- the comparison and/or classification process is weighted.
- a method of comparing an object or an object cluster in a query image with objects in images in an image database comprises defining a plurality of parameters characterizing the objects, calculating a parameter vector comprising said plurality of parameters for objects in the image database so as to produce a set of parameter vectors, deriving a set of weights from the calculated set of parameter vectors, wherein each weight is associated with a corresponding one of the plurality of parameters.
- a parameter vector comprising the plurality of parameters for an object or cluster of objects in said query image is calculated, and a weighted comparison and/or classification between the object or cluster of objects in the query image and the objects, or cluster of objects or image in the image database using the parameter vectors and the derived weights is performed.
- FIG. 1 is a flowchart of a method of image retrieval in one embodiment of the invention.
- FIG. 2 is a block diagram of an image retrieval system according to the invention which may be utilized to carry out the method of Figure 1.
- FIG. 3 is a conceptual schematic of parameter sets associated with objects segmented from an image which may be created by the object parameterization module of Figure 2.
- FIG. 4 is a flowchart of one embodiment of an object parameterization process which may be implemented in the object parameterization module of Figure 2.
- FIG. 5 is a screen display of user configured look up table filter functions according to one embodiment of the invention and which may be generated by the system of Figure 2.
- FIG. 6 is a screen display of user configured sharpening filter functions according to one embodiment of the invention and which may be generated by the system of Figure 2.
- FIG. 7 is a screen display of user configured general and edge enhancement filter functions according to one embodiment of the invention and which may be generated by the system of Figure 2.
- FIG. 8 is a screen display of user configured object definition according to one embodiment of the invention and which may be generated by the system of Figure 2.
- FIG. 9 is a screen display of user configured object searching and comparison according to one embodiment of the invention and which may be generated by the system of Figure 2.
- a user of the system wishes to find images in an image database which contain a certain defined object or a specific content represented by presence or absence of a certain objects (areas) such as, for example, tumors and lesions in medicine, aircrafts in aerospace images, defective cracks in quality control applications and etc.
- the search may be for images containing a chair, sunset, mountain, or the like. This may be the case in the advertising or commercial arts fields, for example, where large searchable image files are kept for use in the production of artwork, posters, etc.
- Other applications may also benefit from a robust object recognition system.
- images such as X-ray films, CAT scans, ultrasound and MRI images generally contain specific internal objects of interest such as blocked blood vessels, tumors, and other structures.
- diagnosis and treatment would be facilitated if a physician could search an image database for similar structures found in other patients so as to obtain valuable information regarding diagnosis, treatment, and outcome for other patients showing similar objects under X-ray or MRI imaging.
- the proposed system may also be used to guide and monitor tumor ablation therapy as well as to monitor tumor response to other therapies.
- Another important industrial application of the presented invention is educational and training application when a user compare his/her results of an image filtered to a database of know findings clinically proven and described as to its absolute accuracy.
- a method of image comparison begins at block 12, where a starting or query image is selected.
- the query image will typically be provided by a user of the system and will comprise an image which contains one or more structures or objects of interest.
- the structure of interest in the image may not be well defined or distinct relative to the background.
- the object boundaries may be poorly delineated, or it may have significant internal features present that are not immediately apparent in the image.
- the system performs image filtering at block 14.
- the filtering performed is under the control of the system user.
- the system may also perform filtering automatically using default filter functions or filter functions previously defined and stored by a user.
- a wide variety of well known image filtering techniques may be made available to the user. Many image filtering techniques which may be used in embodiments of the invention are described at pages 151-346 of The Image Processing Handbook, 2d Edition, John C. Russ, author, and published in 1995 by CRC Press and A spatial clustering procedure for multi-image data, R.M Haralick, I. Dinstein, authors, and published in 1975 by CGI Proc, which is hereby incorporated by reference into this application in its entirety.
- filters which are utilized in one embodiment of the invention are set forth below with reference to Figures 4-6. These filters may enhance edges, enhance the appearance of pixels in particular brightness ranges, stretch contrast in selected pixel brightness ranges, reduce noise, or perform any of a wide variety of pixel processing functions. It will be appreciated that the filtering performed at block 14 may comprise the sequential application of several individual pixel filtering functions. Advantageously, filtering performed in block 14 can result in the enhancement of features which are characteristic of objects of interest or objects within a certain class, etc., but which do not appear in other objects or in the image background. Following the filtering of block 14, objects within the filtered image are defined at block 16. Once again, this process may be performed under the control of the user, or performed automatically by the system.
- this process involves evaluating pixel values so as to classify them as either an object pixel or a background pixel.
- the object definition process of block 16 may be done using many well known techniques, some of which are described at pages 347-405 of The Image Processing Handbook mentioned above. Example object definition protocols provided in one embodiment of the invention are described in more detail with reference to Figure 7.
- each defined object is separately numerically characterized by a set of parameters which are calculated from the pixel locations and brightness values of each defined object.
- the numerical parameters are measures of the object's shape, size, brightness, contrast, energy, density, relief, color, texture and other calculated characteristics.
- the values present in the parameter sets are similar for objects of the same type. Example parameters which may advantageously be used in embodiments of the invention are described below with reference to Figure 3.
- a template for comparison is defined by the user.
- the template may be a single defined object, or may be a group or cluster of defined objects in a region of the image.
- similarities between the template and other objects or sets of objects are calculated. If the template is a single object, this may be done by comparing the parameter set assigned to the template object with the parameter sets assigned to other objects.
- There are several well known ways of evaluating the similarity between two parameter vectors For example, Euclidean or Minkowski line metrics may be used. If the parameter set is represented as a bit string, the Hamming distance may be used as the similarity measure.
- multi-dimensional non-binary parameter sets are associated with the objects, and as stated above, a comparison may be performed between not only individual parameter sets but also between parameter set groups associated with clusters of a plurality of objects.
- a more complicated formula has been developed and may be used, based on ideas set forth in Voronin, Yu. A., Theory of Classification and Its Applications 1985, published in Russia by Nauka. This formula is set forth fully below.
- the comparison involves not only a comparison of the objects themselves, but also the spatial relationship between them.
- a query image may comprise a digital image of an area of skin pigmentation.
- a physician may be interested in evaluating the likelihood that the pigmentation in the image is a melanoma.
- the digital image is filtered and an image area associated with the pigmentation is defined as an object within the image.
- Other images of skin pigmentation which are stored in an image database are also filtered and areas of skin pigmentation are defined as objects, advantageously using the same filters and object definition functions. These objects in the database are then also parameterized.
- the query parameter set is compared to the parameter sets associated with the database objects, and images of skin pigmentation which are similar are identified.
- the pigmentation area of the stored images have been previously characterized as being melanoma or not.
- the physician may be alerted that the possibility of melanoma for the query image is high.
- the filtering and object definition procedures enhance those aspects of skin pigmentation images which are closely associated with the presence of a melanoma.
- the parameter set itself may be tailored to the class of objects being analyzed. This may be done by assigning different weights to the different parameters of the parameter set during the comparison. For the melanoma example, a high weight may be assigned to parameters which are indicative of an irregular boundary or surface, while a lower weight may be assigned to a parameter associated with the total area of the object.
- An image acquisition device 26 is used to initially create images for storage in an image database 24 and/or for routing to a query image selection module 28 of the system.
- the image acquisition device may be a source of images of any type, including photographs, ultrasound images, X-ray or MRI images, a CRT display or trace, or any other data source having an output, which is definable as a collection of digital values.
- the image acquisition device may, for example, be a digital camera or ultrasound workstation.
- the image acquisition device may produce the image directly.
- the system may also import previously created images from one or more image sources.
- the image acquisition device may be an external digital image source for such systems like the Internet or Telnet, for example, or niche specific system such as PACS or RIS.
- the image data array processed by the system will be a two-dimensional array of pixels wherein each pixel is assigned an associated scalar or vector value.
- each pixel is associated with a brightness value, typically eight bits, defining a gray scale from zero (black) to 255 (white) or to 4096 for 16-bit formats.
- a three component vector of data values may be associated with each pixel.
- the query image selection module may, under the control of a user, select a query image from the image acquisition device, or may retrieve an image from the image database 24.
- the system also comprises a display 30 which provides a visual output of one or more images to the user of the system.
- the query image itself will typically be displayed to the user with the display device 30.
- This display of the query image may further be performed after image filtering by the filter module 32 and object definition by the object definition module 34. If no filtering or object segmentation has yet been implemented by the user with these modules, the unprocessed query image will be displayed to the user.
- a user input device 36 such as a keyboard, touchpad, or mouse
- the user may control the filter module 32 so as to implement the filtering described above with reference to block 14 of Figure 1. It is one aspect of some embodiments of the invention that the image continues to be displayed as the filtering is implemented.
- the visual impact of the filter application on the image is displayed to the user.
- the user may also control the implementation of object definition by the object definition module 34.
- Pixel brightness thresholds and other features of the object definition procedure may be modified by the user with the input device 36.
- the image may be displayed after object definition so that the user can observe visually the contours and internal features of objects defined in the image. If the object definition technique is modified by the user, the display of the image may be accordingly updated so that the user can evaluate the effects of the alteration graphically on the display.
- the user may allow the system to perform object definition automatically, without requiring any additional user input.
- the above described display updates may be performed after this automatic object definition as well.
- the user may also control aspects of parameter calculation via the user input device 36.
- cranial X-ray images may all be processed with the same filter set and object definition functions prior to parameterization. This helps ensure that compatible images and objects therein are parameterized for comparison. Of course, care must be taken that the sources of the images are themselves compatible.
- Overall brightness, dimensional variations, and other differences between, for example, different microscopes used to obtain the query image and images in the database 24 should be compensated for either prior to or as part of the processing procedures, known as dimension and/or brightness calibration.
- user defined macros of filter and object definitions and detection functions may be stored in a macro database 35 for future use on additional images.
- the user-friendliness of the system is improved by this feature because images from similar sources can be processed in the same way without requiring the user to remember and manually re-select the same set of filtering and object definition functions when processing similar images in the future.
- the user may operate on an image using either individual filter and object definition functions stored in the macro database or user defined groups of individual filter and object definition functions stored in the macro database 35.
- the object definition module 34 is connected to an object parameterization module 38, which receives the pixel values and contour coordinates of the objects defined in the image. This module then calculates the parameter sets described above with reference to block 18 of Figure 1 using the input pixel values. The calculated parameter sets may be stored in an index database 40 for future use. During the image searching and retrieval process, one or more parameter sets associated with a current template will be forwarded to a parameter set comparison module 42 along with parameter sets associated with other objects in the image or other objects in images stored in the image database 24. Objects or object clusters that are similar to the template, are then also displayed to the user on the display 30.
- any given image may have associated with it several different parameter sets, with each parameter set associated with a detected object in that image.
- the image database 24 may store a plurality of images 46, 48, each of which includes a plurality of defined objects 50a-d and 52a-b. Each object is associated with a parameter set 54a-f, which is stored in the index database 40.
- the parameter set includes a computation of the object area by a formula which counts the number of pixels defined as part of object "A" and multiplies that number by a calibration coefficient as follows:
- the area parameter may be calculated instead by the formula:
- X, Y are the coordinates of the periphery pixels of the object.
- Other advantageous object characterization parameters include the length of the perimeter, and the maximum and minimum diameters of the object through the center of gravity of the object. These may be calculated with the formulas:
- the maximum and minimum Ferret diameters of the object may also be included as part of the parameter set, namely:
- Parameters which relate to pixel intensities within the object are also advantageous to include in the object characterization parameter set. These may include optical density, which may be calculated as:
- If] is the brightness (i.e. 0-255) of pixel ij
- I m ⁇ is the maximum pixel brightness in the area/image.
- More complicated intensity functions which parameterize the texture of the object may be utilized as well.
- This parameter belongs to a textural class of parameters and is a measure of the average difference between a pixel values in the object and the values of its surrounding pixels.
- ⁇ (Njj) Njj, although the function may comprise multiplication by a constant, or may involve a more complicated function of the number of nearest neighbors or pixel position within the object.
- I intensity
- i, j e A intensity
- N is a renormalizing constant
- Figure 4 illustrates a flowchart of the parameter set generation process which may be performed by the object parameterization module 38 of Figure 2.
- the base or fundamental parameters are calculated. These are the parameters that use raw pixel positions or intensities as inputs. Examples include area (Equation 1), perimeter (Equation 3), integrated intensity (Equation 10), etc.
- Another set of parameters referred to herein as "secondary" parameters are also calculated.
- the system may receive user input (by entering information into a dialog box with a mouse and/or keyboard, for example) regarding secondary parameter definitions not already utilized by the system.
- the system calculates both the user defined and standard secondary parameters, and at block 58 the parameters thus calculated are formatted into a feature vector and output to either or both the index database 40 and the comparison and statistics system 42 of Figure 2.
- FIGs 5 through 9 a specific implementation of the invention is illustrated by example screen displays which illustrate aspects of user control (via the input devices 36 of Figure 2) and visualization (via the display 30 of Figure 2) of the filtering and object definition processes.
- this embodiment of the invention is implemented in software on a general purpose computer.
- a wide variety of data processing system environments may be utilized in conjunction with the present invention.
- the invention is implemented in software coded in C/C++ programming languages and running on a Pentium series personal computer with, for example, 32 Mbytes of RAM and a 640 MB hard drive.
- the personal computer in this implementation will typically be connected to an image database through a local or wide area network, or via Internet/Telnet client-server system.
- the personal computer runs a standard web browser, which display a communicating application and accesses image databases and image analysis and computer-aided detection software hosted on a remote Internet server.
- Intranet version of the application is also envisioned and implemented.
- the system works as a part of PACS, for example, using LAN and HIS as a hosting system.
- original images 60a and 60b are displayed to the user of the system in respective portions of the display.
- the upper display 60a comprises a close up of a suspected malignancy in a mammogram.
- the lower display 60b is a bone density image utilized in evaluating osteoporosis.
- On another portion 62 of the screen is a display of a filter protocol.
- This portion 62 of the screen display shown one of the computationally simplest filtering techniques under user control in this embodiment, which is look-up-table (LUT) filtering. With this filter, each input pixel brightness value is mapped onto an output pixel brightness value. If pixel brightness ranges from a value of 0 (black) to 255 (white), each value from 0 to 255 is mapped to a new value defined by the LUT being used.
- LUT look-up-table
- the user is provided with a visual indication 64 of the look-up table form being applied, with input pixel values on the horizontal axis and output pixel values on the vertical axis.
- the user may define the nature of the look-up- table filter being applied.
- the user may define both a table form and a table function.
- the form may be selected between linear (no effect on pixel values), triangular, and sawtooth (also referred to as notch).
- the triangular form is illustrated in Figure 5.
- the user may be provided with a slidebar 66 or other input method for selecting the number of periods in the input brightness range.
- the user may also import a previously used user defined LUT if desired.
- the look-up-table form may also be varied by additional user defined functions. These functions may include negative inversion, multiplication or division by a constant, binarization, brightness shifting, contrast stretching, and the like. For each of these functions, the user may control via slidebars or other user manipulatable displays the constants and thresholds utilized by the system for these functions. Histogram based look-up table filtering may also be provided, such as histogram equalization and histogram based piecewise contrast stretching. After the user defines the desired LUT filter, they may apply it to the image by selecting the "APPLY" button 68. The look-up-table defined by the user is then applied to the image or a selected portion thereof.
- second display 70a and 70b of the image is provided following application of the three period triangular LUT filter. If the user modifies the LUT filter function, the image display 70a, 70b is updated to show the visual result of the new filter function when the user clicks the APPLY button 68. Thus, the user may view a substantially continuously updated filtered image as the filter functions used are modified.
- filtered image 70a regions of suspected malignancy are enhanced with respect to the background following LUT application.
- the filtered image 70b the bone density variations present in the central bone segment are enhanced and pronounced.
- convolution filters In addition to LUT filtering, convolution filters, frequency domain filters, and other filter types may be utilized to further enhance and define significant features of imaged objects.
- additional filter types may be selected with checkboxes 78, 80.
- Filter parameters such as filter box size are user controllable via slidebars 82, 84.
- APPLY buttons 86, 88 initiate the filter operation and display update to show the filtered image or image region.
- the bone image 60b is filtered with a 3x3 edge detection filter which produces the filtered image 87 having enhanced pixels along edges in the image.
- a region of interest 89 in an image of blood cells in bodily fluids where a shading filter was used to compensate for a background brightness variation across the image.
- This type of filter belongs to a class of Laplacian filters.
- the filter is a linear filter in the frequency domain.
- the 3x3 kernel is understood to mean that central pixel brightness value is multiplied by 4. As a result of this filtering, the sharpness of small details (not to exceed 3x3 ) of the image is increased.
- This type of filter belongs to a class of Laplacian filters. Functionality is similar to the 3x3 kernel type filter. As a result of this filtering, the sharpness of small details (not to exceed 5x5 ) of the image is increased.
- This filter performs convolution transformation of the image through a user defined multiplication factor. As a result, all details of a user defined size are sharpened.
- the size of processed image detail may be defined through available editing submenu windows for X and Y dimensions.
- This filter performs convolution transformation of the image and belongs to a spatial domain filters.
- the filtering is performed through a user defined multiplication Factor and automatically calculated special parameter.
- This parameter is a ratio of a current pixel value to Mean Square Deviation of a pixel value calculated for the given size of the pixel aperture (or filter box).
- the size of the processed image detail may be defined through available for editing submenu windows for X and Y dimensions.
- This edge enhancement filter belongs to a non-linear range filter. User defines the size of the filter box. This filter provides two regimes, selected by the user. If the default regime Strong is changed by the user to regime Weak, the filter will change the processing method to avoid images noise impact in certain high frequencies.
- This edge detection filter belongs to modified Laplacian omnidirectional edge detection convolution filters. User defines the size of the filter box. This filter performs edge detection of the image through a user defined Factor. The Factor is used for convolution mask values calculations 7. Dilation filters
- Both filters belong to morphological class and are inversive to each other. The first one should be used for image light elements dilation, the second one - for dark elements dilation. If the default regime Strong is changed by the user to regime Weak, both filters will change the processing method to avoid images noise impact in certain high frequencies. In general:
- This filter represents a convolution transformation of modified Gaussian type. It belongs to a class of linear filters in frequency domain. The size of pixel box or aperture is defined by the user for X and Y dimensions. The filter is used often for certain frequencies noise reduction. In general:
- This filter belongs to a non-linear edge-detection class.
- the filter uses a technique with partial derivatives replacement with their estimates. It is known in image processing as a Sobel filter. The size of the pixel box or aperture defined by the user for X and Y dimensions. This filter performs convolution transformation of the image through a user defined amplification Factor. The user also is provided with the ability to set a binarization Threshold if a correspondent check-box is marked.
- the threshold serves as a modification to the classic Sobel filter and enables the user to find right flexibility for the edge detection process. If the threshold is used the outcome of transformation will be a binary image.
- the default but modifiable masks are:
- This filter belongs to a smoothing class filter.
- the size of the pixel box or aperture is defined by the user for X and Y dimensions.
- the filter is modified from a classical type shading correction filter by enabling the user with shifting capability. If check-box Shift is marked the user will be able to change the default value of the shift to a custom one. This filter is very handy for elimination of a negative lighting impact which sometimes occurs during the image acquisition process.
- the default size of the kernel is 9x9.
- the convolution mask contains default typically used weights values. Push-button activates the customization regime when the user is able to modify dimensions of the mask and then modify default weights in the convolution mask.
- Moving median (or sometimes referred as rank ) filter produces as an output the median, replacing a pixel ( rather than the mean ), of the pixel values in a square pixel box centered around that pixel.
- the filter is a non-linear type filter with the filtration window dimensions of 3x3. Usually used to eliminate very small details of the image sized at 1-2 pixels.
- This filter is similar to the filters described above, but with the filtration window dimensions set by the user. The size of eliminated details depend on the size of the set filtration window.
- This filter is similar to median type filters described above. However it provides rectangular filtration window controlled by the user and performs transformation in a two pass algorithm
- object definition (corresponding to module 34 of Figure 2) is illustrated in Figure 8.
- the user implements manual or semi-automatic object definition.
- slidebars allow the user to select a brightness range of pixels. All pixels outside this range are considered background.
- An object is thus defined as a connected set of pixels having brightness values in the user defined range. Background pixels may be reassigned a zero brightness value.
- the thresholds are calculated automatically by the system from the image histogram. In this mode, the system may allow the user to set up multiple thresholds by setting their values manually or by choosing their sequential numbers from the automatically calculated table of thresholds.
- the image is displayed as the object definition function is applied.
- the object definition function is applied.
- Object definition may be performed in two modes with three regimes each mode: manual, semi-automatic and automatic.
- the user selects the regimes to use.
- the first mode is one-layer segmentation, when the resultant objects belong to the same layer— basically everything that does not belong to the background belongs to the object layer.
- manual regime of this mode the user through the menu defines two thresholds.
- the threshold is calculated by the system after the original pixel histogram ("brightness histogram") is automatically smoothed.
- the user has a choice of threshold values calculated by the system, but only one selected threshold is used for the segmentation process.
- the number of pixels that defines the segmented object is determined by hard-coded percentages of pixels with given values.
- the percentages are controlled by the user through menu selection of "Small, Medium or Large” percent indicators.
- the system calculates a set of thresholds based on pixel code histogram distributions. For such calculations an algorithm of the histogram local min-max identification is implemented. The method used for segmentation represents some modification of a technique that assumes a distinct and dominant peak for the background that is symmetric about its maximum. The technique can benefit from smoothing as described above. Once the maximum value in the histogram is found the algorithm then searches on the non-object pixel side of that maximum to find a % point. After smoothing the system uses as a threshold a displacement to the left of the maximum that is equal to the displacement to the right where the defined % is found because of the assumed symmetry.
- This technique is adapted easily where we have light objects on a dark, dominant background or dark objects on a light, dominant background.
- An additional variation on this symmetry is used when an estimate of the sample standard deviation based on one side of the dominant peak exceeds some given value. Then the selected threshold is modified to "+” or "-". The choice of "+” or "-" depends on which direction from maximum is being defined as the object/background threshold.
- the second mode is multi-layer segmentation, where the resultant objects belong to different layers.
- This mode produces not only different layers of objects but also objects of different type—nesting and pyramidal.
- manual and semi-automatic regimes of this mode are similar in principle to one-layer mode segmentation, the automatic regime uses a different algorithm of threshold calculation.
- This algorithm uses statistical sequences of brightness slopes in conjunction with estimation of asymmetry for the gray-level histogram.
- thresholds are calculated the user may execute segmentation in batch regime automatically or he may modify automatic thresholding by moving, deleting or adding thresholds.
- the display system allows switching on and off the selected object layers so that the user may export or view all or partial results of object quantification.
- the image is of poorer quality the system allows for guided segmentation based on selection of a "seed point" by the user.
- Figure 9 shows a display of the original image 104 after filtering and object segmentation, as well as the template 106 selected for comparison to objects in the remainder of the image.
- the template 106 is a three object cluster.
- seven displays 108a-g which display in rank order the seven objects of the image most similar to the template object.
- Also displayed at 110 is a list of the parameters used in the comparison and the weights assigned to them for the comparison process. These weights may be manually set, or they may be set via a statistical process which is described in further detail below.
- a parameter difference vector may be computed which has as each element the difference between the parameter values divided by the maximum difference observed between the template object and all objects being compared to the template.
- P is a parameter-vector
- ft is the index of template object
- j 1,...,L
- L is all objects that the template object is being compared to
- j is the index of specific object being compared to template object.
- a numerical similarity may then be computed using either a modified form of Euclidean or Minkowski line metrics as set forth below:
- the spatial relationship between selected objects of the template to other objects in the template may be numerically characterized and effectively added as one or more additional subvectors of the object parameter vector.
- the overall similarity between a multi-object template and object clusters in the image database may, in some embodiments of the invention be calculated as follows:
- ⁇ is a thresholds and/or tolerances vector
- ⁇ s is a weights vector
- This formula combines not only parametric similarity but spatial similarity also.
- spatial similarity the closeness of the position and pattern fit for objects of the template and objects of the database are numerically evaluated.
- the mathematical method for parameterizing these spatial relationships may, for example, use some simple Euclidean distances between objects for primitive cases and up to pattern fit calculations based on second, third, or fourth moments of inertia for comparable components in complex cases. Metrics of similarity could be of different types from relative similarity to normalized or absolute metrics of similarities.
- the comparison calculation involves the mathematical generation of a value which characterizes how "similar" two vectors or matrices of numbers without further reference to the meaning associated with those numbers.
- a wide variety of mathematical techniques are available to perform such a numerical characterization, and different approaches may be more suitable than others in different contexts.
- the specific formalism used to mathematically define and quantify similarity between number sets may vary widely in different embodiments of the invention and different techniques may be appropriate depending on the application.
- the weight assigned to a given parameter during this comparison process may be manually set by the user or set using a statistical method.
- the statistical method is especially useful when the database of images includes a large number of objects which have been characterized as having or not having a characteristic trait, such as an area of skin pigmentation is either melanoma or not melanoma, or which have been characterized numerically as more similar or less similar to a "model" object.
- this data can be analyzed to determine how strongly different parameters of the parameter set values correlate with the presence or absence of the specific trait.
- the weight used for a given parameter in the comparison process may thus be derived from the values of the parameter vectors associated with the detected objects in the image database.
- a system is represented as a totality of factors.
- the mathematical simulation tools are correlation, regression, and multifactor analyses, where the coefficients of pairwise and multiple correlation are computed and a linear or non-linear regression is obtained.
- the data for a specific model experiment are represented as a matrix whose columns stand for factors describing the system and the rows for the experiments (values of these factors).
- the factor Y for which the regression is obtained, is referred to as the system response.
- Responses are integral indicators but theoretically, any factor can be a response. All the factors describing the system can be successively analyzed.).
- the coefficients of the regression equation and the covariances help to "redistribute" the multiple determination coefficient among the factors; in other words the "impact” of every factor to response variations is determined.
- the specific impact indicator of the factor is the fraction to which a response depending on a totality of factors in the model changes due to this factor. This specific impact indicator may then be used as the appropriate weight to assign to that factor (i.e. parameter of the parameter set associated with the objects).
- the impact of a specific factor is described by a specific impact indicator which is computed by the following algorithm:
- ⁇ is the specific impact indicator of the j-th factor
- k is the number of factors studied simultaneously
- bj is the j-th multiple regression coefficient which is computed by the formula
- R is the coefficient of multiple determination computed by the formula
- n is the number of observations, which cannot be below (2 * K);
- XQ is the value of the system response in the i-th observation,
- CQJ is the covariance coefficient of the system response indicator and the j-th factor. It is given by the relation
- the specific contribution indicator is obtained mainly from the coefficient of multiple determination, which is computed by the formula
- D 2 is the response variance.
- the specific impact of the j-th factor on the determination coefficient depends only on the ratio of addends in this formula. This implies that the addend whose magnitude is the largest is associated with the largest specific impact. Since the regression coefficients may have different signs, their magnitudes have to be taken in the totals. For this reason, the coefficients ⁇ of the specific impact are bound to be positive. However, it is important that the direction in which the factor acts by the computed ⁇ is dictated by the sign of the regression coefficient. If this sign is positive, the impact on the response variable is positive and if it is not, the increase of the factor results in a reduction of the response function. The influence of the background factors, which are not represented in the data, is computed by the formula
- a rearrangement of the initial data matrix at every experimental step makes it possible to investigate successively the dynamics of the significance of the impact the factors have on all system indicators that become responses successively.
- This method increases the statistical significance of the results obtained from the algorithm for the recomputation of the initial data matrix.
- the algorithm embodies serial repeatability of the experiments by fixing the factors at certain levels. If the experiment is passive, the rows of the initial matrix are chosen in a special way so that, in every computation, rows with the closest values of factors (indicators) influencing the response are grouped together.
- the dynamics of the specific contributions is computed by using the principle of data elimination.
- the computation of the dynamics of the insignificant information is gradually eliminated.
- the value of ⁇ does not change remarkably until the significant information is rejected.
- a dramatic reduction of ⁇ is associated with a threshold with which this elimination of useful information occurs.
- the algorithm of this operation is an iterative ⁇ recomputation by formula (23) and a rejection of information exceeding the threshold computed.
- the significance of the result and of the information eliminated is increased by recomputing the initial data matrix into a series-averaged matrix, the series being, for instance, the totality of matrix rows grouped around the closest values of the factor in the case of a passive factorial experiment.
- the series may also consist of repeated changes of the indicator with the others fixed at a specified level.
- Xi j is the value of the i-th series in which the factor X ⁇ is observed and for which the critical (rejection) threshold is determined after the elimination of data with a threshold of H
- nj is the number of observations in the i-th series
- m is the number of values of the X ⁇ which exceed h and (0 ⁇ m ⁇ nj )
- N is the number of observation series (rows of the N * (K + 1) matrix of the initial information, where K is the number of factors investigated simultaneously.)
- the invention thus provides image searching and comparison based in a much more direct way on image content and meaning than has been previously available.
- using the described method of weights calculations for targeting similarities between a multi-component template and a database of images in medical fields is much more mathematically justified and sound than neural network techniques used for the same purposes. That is important to understand because template matching may be used in such applications to decrease the difficulty of database creation and search, and improve early cancer diagnostics, early melanoma detection, etc.
Abstract
Description
Claims
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AU65338/00A AU6533800A (en) | 1999-08-09 | 2000-08-09 | Object based image retrieval |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/370,366 | 1999-08-09 | ||
US09/370,366 US6941323B1 (en) | 1999-08-09 | 1999-08-09 | System and method for image comparison and retrieval by enhancing, defining, and parameterizing objects in images |
Publications (3)
Publication Number | Publication Date |
---|---|
WO2001011489A2 WO2001011489A2 (en) | 2001-02-15 |
WO2001011489A9 true WO2001011489A9 (en) | 2002-07-18 |
WO2001011489A3 WO2001011489A3 (en) | 2004-04-22 |
Family
ID=23459327
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2000/021735 WO2001011489A2 (en) | 1999-08-09 | 2000-08-09 | Retrieving images by defining an image content comprising objects |
Country Status (3)
Country | Link |
---|---|
US (3) | US6941323B1 (en) |
AU (1) | AU6533800A (en) |
WO (1) | WO2001011489A2 (en) |
Families Citing this family (153)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6445884B1 (en) * | 1995-06-22 | 2002-09-03 | 3Dv Systems, Ltd. | Camera with through-the-lens lighting |
AU4910500A (en) * | 1999-05-19 | 2000-12-12 | Neurosearch A/S | Inhibitors of proton-gated cation channels and their use in the treatment of ischaemic disorders |
US6993170B2 (en) * | 1999-06-23 | 2006-01-31 | Icoria, Inc. | Method for quantitative analysis of blood vessel structure |
GB2352076B (en) * | 1999-07-15 | 2003-12-17 | Mitsubishi Electric Inf Tech | Method and apparatus for representing and searching for an object in an image |
US6941323B1 (en) * | 1999-08-09 | 2005-09-06 | Almen Laboratories, Inc. | System and method for image comparison and retrieval by enhancing, defining, and parameterizing objects in images |
US6901156B2 (en) * | 2000-02-04 | 2005-05-31 | Arch Development Corporation | Method, system and computer readable medium for an intelligent search workstation for computer assisted interpretation of medical images |
JP2003123086A (en) * | 2001-10-11 | 2003-04-25 | Sony Corp | Information processor and method, and information processing program |
JP2003132090A (en) * | 2001-10-26 | 2003-05-09 | Olympus Optical Co Ltd | Similar data retrieval device and method therefor |
US7352886B1 (en) * | 2001-11-20 | 2008-04-01 | Icad, Inc. | Error handling in a high throughput computer-aided detection environment |
US20030103663A1 (en) * | 2001-11-23 | 2003-06-05 | University Of Chicago | Computerized scheme for distinguishing between benign and malignant nodules in thoracic computed tomography scans by use of similar images |
AU2002348833A1 (en) * | 2001-11-30 | 2003-06-10 | Koninklijke Philips Electronics N.V. | Medical viewing system and method for enhancing structures in noisy images |
WO2003049032A2 (en) * | 2001-12-07 | 2003-06-12 | Koninklijke Philips Electronics N.V. | Medical viewing system and method for spatially enhancing structures in noisy images |
EP1470702A4 (en) | 2002-01-11 | 2005-08-03 | Portrait Innovations Inc | Systems and methods for producing portraits |
DE10213284B4 (en) * | 2002-03-25 | 2007-11-08 | Siemens Ag | Method for imaging in medical imaging |
US9710852B1 (en) | 2002-05-30 | 2017-07-18 | Consumerinfo.Com, Inc. | Credit report timeline user interface |
US9400589B1 (en) | 2002-05-30 | 2016-07-26 | Consumerinfo.Com, Inc. | Circular rotational interface for display of consumer credit information |
US20040076345A1 (en) * | 2002-09-18 | 2004-04-22 | Olszak Artur G. | Method for referencing image data |
US7194119B2 (en) * | 2002-11-21 | 2007-03-20 | Siemens Aktiengesellschaft | Method and system for retrieving a medical picture |
US7228006B2 (en) * | 2002-11-25 | 2007-06-05 | Eastman Kodak Company | Method and system for detecting a geometrically transformed copy of an image |
US20040120558A1 (en) * | 2002-12-18 | 2004-06-24 | Sabol John M | Computer assisted data reconciliation method and apparatus |
US7490085B2 (en) * | 2002-12-18 | 2009-02-10 | Ge Medical Systems Global Technology Company, Llc | Computer-assisted data processing system and method incorporating automated learning |
US20040122719A1 (en) * | 2002-12-18 | 2004-06-24 | Sabol John M. | Medical resource processing system and method utilizing multiple resource type data |
US20040122703A1 (en) * | 2002-12-19 | 2004-06-24 | Walker Matthew J. | Medical data operating model development system and method |
US7315833B2 (en) * | 2003-01-16 | 2008-01-01 | Rosetta Holdings, Llc | Graphical internet search system and methods |
US20040179735A1 (en) * | 2003-03-13 | 2004-09-16 | Aruna Kumar | Method and apparatus for characterizing objects within an image |
JP2004326491A (en) * | 2003-04-25 | 2004-11-18 | Canon Inc | Image processing method |
JP2004334339A (en) * | 2003-04-30 | 2004-11-25 | Canon Inc | Information processor, information processing method, and storage medium, and program |
JP2004361987A (en) * | 2003-05-30 | 2004-12-24 | Seiko Epson Corp | Image retrieval system, image classification system, image retrieval program, image classification program, image retrieval method, and image classification method |
US20050008262A1 (en) * | 2003-06-03 | 2005-01-13 | Konica Minolta Medical & Graphic, Inc. | Medical image system, and medical image processing method |
US7353510B2 (en) * | 2003-07-11 | 2008-04-01 | Computer Associates Think, Inc. | System and method for comparing objects |
US20050027666A1 (en) * | 2003-07-15 | 2005-02-03 | Vente, Inc | Interactive online research system and method |
US20050041845A1 (en) * | 2003-08-20 | 2005-02-24 | Payne Randall Kenneth | Medical imaging system with tissue-selective image sharpening |
US8442280B2 (en) * | 2004-01-21 | 2013-05-14 | Edda Technology, Inc. | Method and system for intelligent qualitative and quantitative analysis of digital radiography softcopy reading |
US20050209983A1 (en) * | 2004-03-18 | 2005-09-22 | Macpherson Deborah L | Context driven topologies |
US7664326B2 (en) * | 2004-07-09 | 2010-02-16 | Aloka Co., Ltd | Method and apparatus of image processing to detect and enhance edges |
US7578790B2 (en) * | 2004-07-20 | 2009-08-25 | Boston Scientific Scimed, Inc. | Systems and methods for detecting and presenting textural information from medical images |
US20060173318A1 (en) * | 2004-07-20 | 2006-08-03 | Scimed Life Systems Inc. | Systems and methods for detecting and presenting textural information from medical images |
US20060036147A1 (en) * | 2004-07-20 | 2006-02-16 | Scimed Life Systems, Inc. | Systems and methods for detecting and presenting textural information from medical images |
US20060171574A1 (en) * | 2004-11-12 | 2006-08-03 | Delmonego Brian | Graphical healthcare order processing system and method |
NZ556655A (en) * | 2005-01-19 | 2010-10-29 | Dermaspect Llc | Devices and methods for identifying and monitoring changes of a suspect area on a patient |
US20060242143A1 (en) * | 2005-02-17 | 2006-10-26 | Esham Matthew P | System for processing medical image representative data from multiple clinical imaging devices |
US8305427B2 (en) * | 2005-03-22 | 2012-11-06 | Olympus Corporation | Image processor and endoscope apparatus |
US20070076929A1 (en) * | 2005-10-05 | 2007-04-05 | General Electric Company | System and method for automatic post processing image generation |
JP2009511163A (en) | 2005-10-14 | 2009-03-19 | アプライド リサーチ アソシエイツ エヌゼット リミテッド | Method and apparatus for observing surface features |
DE102006001090B4 (en) * | 2006-01-09 | 2011-11-17 | Siemens Ag | Operating method and imaging device |
US7526116B2 (en) * | 2006-01-19 | 2009-04-28 | Luigi Armogida | Automated microscopic sperm identification |
US20070211930A1 (en) * | 2006-03-09 | 2007-09-13 | Terry Dolwick | Attribute based image enhancement and display for medical imaging applications |
US8243999B2 (en) | 2006-05-03 | 2012-08-14 | Ut-Battelle, Llc | Method and system for the diagnosis of disease using retinal image content and an archive of diagnosed human patient data |
TWI403912B (en) * | 2006-06-08 | 2013-08-01 | Univ Nat Chiao Tung | Method and system of image retrieval |
US7840060B2 (en) * | 2006-06-12 | 2010-11-23 | D&S Consultants, Inc. | System and method for machine learning using a similarity inverse matrix |
US20080046424A1 (en) * | 2006-06-23 | 2008-02-21 | Horton Richard B | System and method of selecting images according to color content |
US8121360B2 (en) * | 2006-07-31 | 2012-02-21 | Siemens Medical Solutions Usa, Inc. | Computer aided detection and decision support |
JP2008059088A (en) * | 2006-08-29 | 2008-03-13 | Toshiba Corp | Image display system and image display device |
US7835043B2 (en) * | 2006-09-06 | 2010-11-16 | Hewlett-Packard Development Company, L.P. | Imaging device and calibration method therefor |
US20080123953A1 (en) * | 2006-11-29 | 2008-05-29 | Gateway Inc. | Digital camera with histogram zoom |
US7689610B2 (en) * | 2006-12-01 | 2010-03-30 | Computer Associates Think, Inc. | Automated grouping of messages provided to an application using string similarity analysis |
US7917911B2 (en) * | 2006-12-01 | 2011-03-29 | Computer Associates Think, Inc. | Automated grouping of messages provided to an application using execution path similarity analysis |
US20100079481A1 (en) * | 2007-01-25 | 2010-04-01 | Li Zhang | Method and system for marking scenes and images of scenes with optical tags |
DE102007009485A1 (en) * | 2007-02-22 | 2008-08-28 | Perner, Petra, Dr.-Ing. | Method and data processing system for modeling the image segmentation |
US20090082637A1 (en) * | 2007-09-21 | 2009-03-26 | Michael Galperin | Multi-modality fusion classifier with integrated non-imaging factors |
JP5159242B2 (en) | 2007-10-18 | 2013-03-06 | キヤノン株式会社 | Diagnosis support device, diagnosis support device control method, and program thereof |
US7853045B2 (en) * | 2007-10-31 | 2010-12-14 | Saudi Arabian Oil Company | Geostatistical analysis and classification of core data |
CA2744690C (en) * | 2007-12-13 | 2016-07-05 | University Of Saskatchewan | Image analysis |
US9990674B1 (en) | 2007-12-14 | 2018-06-05 | Consumerinfo.Com, Inc. | Card registry systems and methods |
US8127986B1 (en) | 2007-12-14 | 2012-03-06 | Consumerinfo.Com, Inc. | Card registry systems and methods |
CN101903883B (en) * | 2007-12-20 | 2017-02-15 | 皇家飞利浦电子股份有限公司 | Method and device for case-based decision support |
WO2009108791A1 (en) * | 2008-02-26 | 2009-09-03 | The Regents Of The University Of California | Diagnostic skin mapping by mrs, mri and other methods |
WO2009139161A1 (en) * | 2008-05-15 | 2009-11-19 | 株式会社ニコン | Image processing device, image processing method, processing device, processing method, and program |
US8090727B2 (en) * | 2008-05-20 | 2012-01-03 | Check Point Software Technologies Ltd. | Methods for automatically generating natural-language news items from log files and status traces |
US20090319388A1 (en) * | 2008-06-20 | 2009-12-24 | Jian Yuan | Image Capture for Purchases |
US8312033B1 (en) | 2008-06-26 | 2012-11-13 | Experian Marketing Solutions, Inc. | Systems and methods for providing an integrated identifier |
US8538941B2 (en) * | 2008-07-31 | 2013-09-17 | Adobe Systems Incorporated | Visual information search tool |
CN101348522B (en) * | 2008-08-01 | 2010-10-13 | 中国科学院海洋研究所 | Copper resistance binary signal conducting system regulation factor mutant, and construction and use thereof |
US9256904B1 (en) | 2008-08-14 | 2016-02-09 | Experian Information Solutions, Inc. | Multi-bureau credit file freeze and unfreeze |
US20100054555A1 (en) * | 2008-08-29 | 2010-03-04 | General Electric Company | Systems and methods for use of image recognition for hanging protocol determination |
WO2010030056A1 (en) * | 2008-09-10 | 2010-03-18 | Bionet Co., Ltd | Automatic contour detection method for ultrasonic diagnosis appartus |
JP5301232B2 (en) * | 2008-09-30 | 2013-09-25 | シスメックス株式会社 | Blood cell image display device, sample analysis system, blood cell image display method, and computer program |
US8060424B2 (en) | 2008-11-05 | 2011-11-15 | Consumerinfo.Com, Inc. | On-line method and system for monitoring and reporting unused available credit |
JP5374135B2 (en) * | 2008-12-16 | 2013-12-25 | オリンパス株式会社 | Image processing apparatus, method of operating image processing apparatus, and image processing program |
WO2010083021A1 (en) * | 2009-01-16 | 2010-07-22 | Thomson Licensing | Detection of field lines in sports videos |
EP2249286A1 (en) * | 2009-05-08 | 2010-11-10 | Honda Research Institute Europe GmbH | Robot with vision-based 3D shape recognition |
US8330807B2 (en) * | 2009-05-29 | 2012-12-11 | Convergent Medical Solutions, Inc. | Automated assessment of skin lesions using image library |
US9367523B2 (en) | 2009-09-25 | 2016-06-14 | Adobe Systems Incorporated | System and method for using design features to search for page layout designs |
US8687860B2 (en) * | 2009-11-24 | 2014-04-01 | Penrad Technologies, Inc. | Mammography statistical diagnostic profiler and prediction system |
US8799013B2 (en) * | 2009-11-24 | 2014-08-05 | Penrad Technologies, Inc. | Mammography information system |
US8411970B2 (en) | 2010-03-16 | 2013-04-02 | Pixia Corp. | Method and system for determining statistical data for image pixels having a higher bit depth per band |
US8463797B2 (en) * | 2010-07-20 | 2013-06-11 | Barracuda Networks Inc. | Method for measuring similarity of diverse binary objects comprising bit patterns |
US8737702B2 (en) | 2010-07-23 | 2014-05-27 | International Business Machines Corporation | Systems and methods for automated extraction of measurement information in medical videos |
US8792728B2 (en) * | 2010-09-27 | 2014-07-29 | Hewlett-Packard Development Company, L.P. | Near-duplicate image detection |
WO2012064169A2 (en) * | 2010-11-10 | 2012-05-18 | Rolando Rafael Hernandez Cisneros | Method for the detection and classification of microcalcification groups in digital mammographies |
US8737713B2 (en) | 2010-11-30 | 2014-05-27 | Siemens Medical Solutions Usa, Inc. | System for frame selection for optimal registration of a multi-frame dataset |
US9047319B2 (en) * | 2010-12-17 | 2015-06-02 | Microsoft Technology Licensing, Llc | Tag association with image regions |
US9665854B1 (en) | 2011-06-16 | 2017-05-30 | Consumerinfo.Com, Inc. | Authentication alerts |
US9483606B1 (en) | 2011-07-08 | 2016-11-01 | Consumerinfo.Com, Inc. | Lifescore |
US9106691B1 (en) | 2011-09-16 | 2015-08-11 | Consumerinfo.Com, Inc. | Systems and methods of identity protection and management |
WO2013052937A1 (en) | 2011-10-06 | 2013-04-11 | Nant Holdings Ip, Llc | Healthcare object recognition systems and methods |
US8738516B1 (en) | 2011-10-13 | 2014-05-27 | Consumerinfo.Com, Inc. | Debt services candidate locator |
US9179844B2 (en) | 2011-11-28 | 2015-11-10 | Aranz Healthcare Limited | Handheld skin measuring or monitoring device |
AU2011253982B9 (en) * | 2011-12-12 | 2015-07-16 | Canon Kabushiki Kaisha | Method, system and apparatus for determining a subject and a distractor in an image |
US9460390B1 (en) * | 2011-12-21 | 2016-10-04 | Emc Corporation | Analyzing device similarity |
CN104115149B (en) * | 2012-02-07 | 2017-11-17 | 皇家飞利浦有限公司 | Interactive optimization for the scan database of statistics test |
GB2499385A (en) * | 2012-02-09 | 2013-08-21 | Nokia Corp | Automated notification of images with changed appearance in common content |
JP5677348B2 (en) * | 2012-03-23 | 2015-02-25 | 富士フイルム株式会社 | CASE SEARCH DEVICE, CASE SEARCH METHOD, AND PROGRAM |
JP5882805B2 (en) | 2012-03-26 | 2016-03-09 | キヤノン株式会社 | Information processing apparatus, processing method thereof, and program |
US9292793B1 (en) * | 2012-03-31 | 2016-03-22 | Emc Corporation | Analyzing device similarity |
US9853959B1 (en) | 2012-05-07 | 2017-12-26 | Consumerinfo.Com, Inc. | Storage and maintenance of personal data |
US8923582B2 (en) * | 2012-05-31 | 2014-12-30 | General Electric Company | Systems and methods for computer aided detection using pixel intensity values |
CA2878455C (en) | 2012-07-06 | 2020-12-22 | Nant Holdings Ip, Llc | Healthcare analysis stream management |
US9654541B1 (en) | 2012-11-12 | 2017-05-16 | Consumerinfo.Com, Inc. | Aggregating user web browsing data |
US9916621B1 (en) | 2012-11-30 | 2018-03-13 | Consumerinfo.Com, Inc. | Presentation of credit score factors |
US10255598B1 (en) | 2012-12-06 | 2019-04-09 | Consumerinfo.Com, Inc. | Credit card account data extraction |
FR3000585A1 (en) * | 2012-12-31 | 2014-07-04 | Eads Europ Aeronautic Defence | METHOD FOR ANALYZING GEOGRAPHIC REGIONS AND DETECTING ZONES OF INTEREST |
US9224066B2 (en) * | 2013-03-01 | 2015-12-29 | Adobe Systems Incorporated | Object detection via validation with visual search |
IN2013CH01043A (en) * | 2013-03-12 | 2015-08-14 | Yahoo Inc | |
US9406085B1 (en) | 2013-03-14 | 2016-08-02 | Consumerinfo.Com, Inc. | System and methods for credit dispute processing, resolution, and reporting |
US9870589B1 (en) | 2013-03-14 | 2018-01-16 | Consumerinfo.Com, Inc. | Credit utilization tracking and reporting |
US10102570B1 (en) | 2013-03-14 | 2018-10-16 | Consumerinfo.Com, Inc. | Account vulnerability alerts |
US10685398B1 (en) | 2013-04-23 | 2020-06-16 | Consumerinfo.Com, Inc. | Presenting credit score information |
CN104142946A (en) * | 2013-05-08 | 2014-11-12 | 阿里巴巴集团控股有限公司 | Method and system for aggregating and searching service objects of same type |
CN104243787B (en) * | 2013-06-06 | 2017-09-05 | 华为技术有限公司 | Photographic method, photo management method and equipment |
US8947745B2 (en) | 2013-07-03 | 2015-02-03 | Symbol Technologies, Inc. | Apparatus and method for scanning and decoding information in an identified location in a document |
US9117104B2 (en) * | 2013-07-10 | 2015-08-25 | Cherif Algreatly | Object recognition for 3D models and 2D drawings |
US9443268B1 (en) | 2013-08-16 | 2016-09-13 | Consumerinfo.Com, Inc. | Bill payment and reporting |
US9830567B2 (en) | 2013-10-25 | 2017-11-28 | Location Labs, Inc. | Task management system and method |
US10325314B1 (en) | 2013-11-15 | 2019-06-18 | Consumerinfo.Com, Inc. | Payment reporting systems |
US9477737B1 (en) | 2013-11-20 | 2016-10-25 | Consumerinfo.Com, Inc. | Systems and user interfaces for dynamic access of multiple remote databases and synchronization of data based on user rules |
USD760256S1 (en) | 2014-03-25 | 2016-06-28 | Consumerinfo.Com, Inc. | Display screen or portion thereof with graphical user interface |
USD759690S1 (en) | 2014-03-25 | 2016-06-21 | Consumerinfo.Com, Inc. | Display screen or portion thereof with graphical user interface |
USD759689S1 (en) | 2014-03-25 | 2016-06-21 | Consumerinfo.Com, Inc. | Display screen or portion thereof with graphical user interface |
US9892457B1 (en) | 2014-04-16 | 2018-02-13 | Consumerinfo.Com, Inc. | Providing credit data in search results |
JP2015215482A (en) * | 2014-05-12 | 2015-12-03 | キヤノン株式会社 | Image display system and image display method |
US9704231B1 (en) | 2014-05-19 | 2017-07-11 | Google Inc. | Measuring and visualizing impact of image modifications |
JP6419560B2 (en) * | 2014-12-05 | 2018-11-07 | 株式会社東芝 | Search device, method and program |
US9349085B1 (en) * | 2015-04-16 | 2016-05-24 | Digicomp Inc. | Methods and system to decode hidden images |
WO2017020140A1 (en) * | 2015-08-03 | 2017-02-09 | Orand S.A. | System for searching for images by sketches using histograms of cell orientations and extraction of contours based on mid-level features |
US10080539B2 (en) * | 2015-11-25 | 2018-09-25 | General Electric Company | Systems and methods for X-ray image pasting |
ITUA20161570A1 (en) | 2016-03-11 | 2017-09-11 | Gruppo Cimbali Spa | Method for the automatic diagnosis of the quality of a dispensed beverage. |
US10013527B2 (en) | 2016-05-02 | 2018-07-03 | Aranz Healthcare Limited | Automatically assessing an anatomical surface feature and securely managing information related to the same |
US10176198B1 (en) * | 2016-05-09 | 2019-01-08 | A9.Com, Inc. | Techniques for identifying visually similar content |
US10796197B2 (en) * | 2016-08-16 | 2020-10-06 | Mykola Petrovych CHAPLAHIN | Automatic method and system for similar images and image fragments detection basing on image content |
AU2017204494B2 (en) * | 2016-09-01 | 2019-06-13 | Casio Computer Co., Ltd. | Diagnosis assisting device, image processing method in diagnosis assisting device, and non-transitory storage medium having stored therein program |
US11116407B2 (en) | 2016-11-17 | 2021-09-14 | Aranz Healthcare Limited | Anatomical surface assessment methods, devices and systems |
EP4183328A1 (en) | 2017-04-04 | 2023-05-24 | Aranz Healthcare Limited | Anatomical surface assessment methods, devices and systems |
CN109165309B (en) * | 2018-08-06 | 2020-10-16 | 北京邮电大学 | Negative example training sample acquisition method and device and model training method and device |
JPWO2020044736A1 (en) * | 2018-08-31 | 2021-08-10 | 富士フイルム株式会社 | Similarity determination device, method and program |
JP6914233B2 (en) * | 2018-08-31 | 2021-08-04 | 富士フイルム株式会社 | Similarity determination device, method and program |
US10880313B2 (en) | 2018-09-05 | 2020-12-29 | Consumerinfo.Com, Inc. | Database platform for realtime updating of user data from third party sources |
US11315179B1 (en) | 2018-11-16 | 2022-04-26 | Consumerinfo.Com, Inc. | Methods and apparatuses for customized card recommendations |
US11238656B1 (en) | 2019-02-22 | 2022-02-01 | Consumerinfo.Com, Inc. | System and method for an augmented reality experience via an artificial intelligence bot |
CN113614831A (en) * | 2019-03-22 | 2021-11-05 | 英芙勒玛提克斯公司 | System and method for deriving and optimizing classifiers from multiple data sets |
US11941065B1 (en) | 2019-09-13 | 2024-03-26 | Experian Information Solutions, Inc. | Single identifier platform for storing entity data |
US20210151169A1 (en) * | 2019-11-18 | 2021-05-20 | Fujifilm Medical Systems U.S.A., Inc. | Methods And Systems For Remote Analysis Of Medical Image Records |
US20230259547A1 (en) * | 2020-06-11 | 2023-08-17 | Nec Corporation | Image selection apparatus, image selection method, and non-transitory computer-readable medium |
Family Cites Families (149)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5019975A (en) | 1986-08-08 | 1991-05-28 | Fuji Photo Film Co., Ltd. | Method for constructing a data base in a medical image control system |
US4907156A (en) * | 1987-06-30 | 1990-03-06 | University Of Chicago | Method and system for enhancement and detection of abnormal anatomic regions in a digital image |
US5179651A (en) * | 1988-11-08 | 1993-01-12 | Massachusetts General Hospital | Apparatus for retrieval and processing of selected archived images for display at workstation terminals |
JPH02260073A (en) * | 1989-03-31 | 1990-10-22 | Toshiba Corp | Picture display device |
US5133020A (en) * | 1989-07-21 | 1992-07-21 | Arch Development Corporation | Automated method and system for the detection and classification of abnormal lesions and parenchymal distortions in digital medical images |
JP2589613B2 (en) | 1991-09-17 | 1997-03-12 | 株式会社日立製作所 | X-ray CT imaging method and X-ray CT apparatus |
US5289374A (en) * | 1992-02-28 | 1994-02-22 | Arch Development Corporation | Method and system for analysis of false positives produced by an automated scheme for the detection of lung nodules in digital chest radiographs |
US5603318A (en) | 1992-04-21 | 1997-02-18 | University Of Utah Research Foundation | Apparatus and method for photogrammetric surgical localization |
US5384862A (en) * | 1992-05-29 | 1995-01-24 | Cimpiter Corporation | Radiographic image evaluation apparatus and method |
US5537485A (en) * | 1992-07-21 | 1996-07-16 | Arch Development Corporation | Method for computer-aided detection of clustered microcalcifications from digital mammograms |
DE69328599T2 (en) | 1992-08-24 | 2000-08-24 | Casio Computer Co Ltd | Data search device |
JP2903904B2 (en) | 1992-10-09 | 1999-06-14 | 松下電器産業株式会社 | Image retrieval device |
US5568384A (en) | 1992-10-13 | 1996-10-22 | Mayo Foundation For Medical Education And Research | Biomedical imaging and analysis |
US5579471A (en) | 1992-11-09 | 1996-11-26 | International Business Machines Corporation | Image query system and method |
DE69432089T2 (en) | 1993-03-01 | 2004-02-12 | Kabushiki Kaisha Toshiba, Kawasaki | System for processing medical data to support diagnosis |
US6690817B1 (en) * | 1993-08-18 | 2004-02-10 | Applied Spectral Imaging Ltd. | Spectral bio-imaging data for cell classification using internal reference |
JP3234064B2 (en) | 1993-09-02 | 2001-12-04 | キヤノン株式会社 | Image retrieval method and apparatus |
US6434262B2 (en) * | 1993-09-29 | 2002-08-13 | Shih-Ping Wang | Computer-aided diagnosis system and method |
US6415048B1 (en) * | 1993-10-12 | 2002-07-02 | Schneider Medical Technologies, Inc. | Compositional analysis system |
WO1995014979A1 (en) | 1993-11-29 | 1995-06-01 | Arch Development Corporation | Automated method and system for improved computerized detection and classification of masses in mammograms |
US5452367A (en) | 1993-11-29 | 1995-09-19 | Arch Development Corporation | Automated method and system for the segmentation of medical images |
US5931780A (en) * | 1993-11-29 | 1999-08-03 | Arch Development Corporation | Method and system for the computerized radiographic analysis of bone |
EP0731956A4 (en) | 1993-11-30 | 1997-04-23 | Arch Dev Corp | Automated method and system for the alignment and correlation of images from two different modalities |
US5638458A (en) * | 1993-11-30 | 1997-06-10 | Arch Development Corporation | Automated method and system for the detection of gross abnormalities and asymmetries in chest images |
JP3067966B2 (en) | 1993-12-06 | 2000-07-24 | 松下電器産業株式会社 | Apparatus and method for retrieving image parts |
JP3026712B2 (en) * | 1993-12-09 | 2000-03-27 | キヤノン株式会社 | Image search method and apparatus |
JPH09508994A (en) * | 1994-01-28 | 1997-09-09 | シュナイダー メディカル テクノロジーズ インコーポレイテッド | Image forming apparatus and method |
US5531227A (en) * | 1994-01-28 | 1996-07-02 | Schneider Medical Technologies, Inc. | Imaging device and method |
JP4044971B2 (en) * | 1994-03-17 | 2008-02-06 | 株式会社日立製作所 | Search method using image information |
US5881124A (en) * | 1994-03-31 | 1999-03-09 | Arch Development Corporation | Automated method and system for the detection of lesions in medical computed tomographic scans |
US5982953A (en) * | 1994-09-02 | 1999-11-09 | Konica Corporation | Image displaying apparatus of a processed image from temporally sequential images |
US5802361A (en) * | 1994-09-30 | 1998-09-01 | Apple Computer, Inc. | Method and system for searching graphic images and videos |
US5659626A (en) * | 1994-10-20 | 1997-08-19 | Calspan Corporation | Fingerprint identification system |
US6246804B1 (en) * | 1994-11-15 | 2001-06-12 | Canon Kabushiki Kaisha | Image retrieval method and apparatus using a compound image formed from a plurality of detected regions |
US5657362A (en) * | 1995-02-24 | 1997-08-12 | Arch Development Corporation | Automated method and system for computerized detection of masses and parenchymal distortions in medical images |
CA2214101A1 (en) * | 1995-03-03 | 1996-09-12 | Ulrich Bick | Method and system for the detection of lesions in medical images |
JP2776294B2 (en) * | 1995-04-12 | 1998-07-16 | 日本電気株式会社 | Image feature extraction device and image processing device for skin pattern image |
US5790690A (en) * | 1995-04-25 | 1998-08-04 | Arch Development Corporation | Computer-aided method for automated image feature analysis and diagnosis of medical images |
US20010033680A1 (en) * | 1995-10-26 | 2001-10-25 | Bankman Isaac N. | Method and system for detecting small structures in images |
US6636635B2 (en) * | 1995-11-01 | 2003-10-21 | Canon Kabushiki Kaisha | Object extraction method, and image sensing apparatus using the method |
US5872859A (en) * | 1995-11-02 | 1999-02-16 | University Of Pittsburgh | Training/optimization of computer aided detection schemes based on measures of overall image quality |
US6165734A (en) | 1995-12-12 | 2000-12-26 | Applied Spectral Imaging Ltd. | In-situ method of analyzing cells |
TW514513B (en) * | 1996-02-06 | 2002-12-21 | Deus Technologies Inc | Method for the detection of lung nodule in radiological images using digital image processing and artificial neural network |
US5748173A (en) | 1996-02-29 | 1998-05-05 | University Of Pittsburgh | Hybrid display for simultaneous side-by-side review of radiographs |
US5893095A (en) | 1996-03-29 | 1999-04-06 | Virage, Inc. | Similarity engine for content-based retrieval of images |
US5911139A (en) * | 1996-03-29 | 1999-06-08 | Virage, Inc. | Visual image database search engine which allows for different schema |
CA2201057C (en) | 1996-03-29 | 2002-01-01 | Kenji Morimoto | A method of processing a sectional image of a sample bone including a cortical bone portion and a cancellous bone portion |
US6198838B1 (en) * | 1996-07-10 | 2001-03-06 | R2 Technology, Inc. | Method and system for detection of suspicious lesions in digital mammograms using a combination of spiculation and density signals |
US6263092B1 (en) * | 1996-07-10 | 2001-07-17 | R2 Technology, Inc. | Method and apparatus for fast detection of spiculated lesions in digital mammograms |
US5974201A (en) | 1996-10-01 | 1999-10-26 | Siemens Corporate Research, Inc. | Smart image system |
US5819288A (en) | 1996-10-16 | 1998-10-06 | Microsoft Corporation | Statistically based image group descriptor particularly suited for use in an image classification and retrieval system |
US5852823A (en) | 1996-10-16 | 1998-12-22 | Microsoft | Image classification and retrieval system using a query-by-example paradigm |
KR20000052802A (en) * | 1996-10-25 | 2000-08-25 | 모세 라르센 페터 | Proteome analysis for characterization of up- and down-regulated proteins in biological samples |
US5987094A (en) * | 1996-10-30 | 1999-11-16 | University Of South Florida | Computer-assisted method and apparatus for the detection of lung nodules |
US6012069A (en) * | 1997-01-28 | 2000-01-04 | Dainippon Screen Mfg. Co., Ltd. | Method and apparatus for retrieving a desired image from an image database using keywords |
US5930783A (en) | 1997-02-21 | 1999-07-27 | Nec Usa, Inc. | Semantic and cognition based image retrieval |
US5919135A (en) * | 1997-02-28 | 1999-07-06 | Lemelson; Jerome | System and method for treating cellular disorders in a living being |
US6032678A (en) * | 1997-03-14 | 2000-03-07 | Shraga Rottem | Adjunct to diagnostic imaging systems for analysis of images of an object or a body part or organ |
US6301378B1 (en) | 1997-06-03 | 2001-10-09 | R2 Technology, Inc. | Method and apparatus for automated detection of masses in digital mammograms |
US5906578A (en) * | 1997-06-18 | 1999-05-25 | Rajan; Govinda N. | Method and system for probe positioning in transesophageal echocardiography |
US6317617B1 (en) | 1997-07-25 | 2001-11-13 | Arch Development Corporation | Method, computer program product, and system for the automated analysis of lesions in magnetic resonance, mammogram and ultrasound images |
US6058322A (en) * | 1997-07-25 | 2000-05-02 | Arch Development Corporation | Methods for improving the accuracy in differential diagnosis on radiologic examinations |
US5984870A (en) | 1997-07-25 | 1999-11-16 | Arch Development Corporation | Method and system for the automated analysis of lesions in ultrasound images |
WO1999005640A1 (en) * | 1997-07-25 | 1999-02-04 | Arch Development Corporation | Method and system for the segmentation of lung regions in lateral chest radiographs |
US6970587B1 (en) | 1997-08-28 | 2005-11-29 | Icad, Inc. | Use of computer-aided detection system outputs in clinical practice |
US5999639A (en) * | 1997-09-04 | 1999-12-07 | Qualia Computing, Inc. | Method and system for automated detection of clustered microcalcifications from digital mammograms |
US5974412A (en) * | 1997-09-24 | 1999-10-26 | Sapient Health Network | Intelligent query system for automatically indexing information in a database and automatically categorizing users |
US6226548B1 (en) * | 1997-09-24 | 2001-05-01 | Surgical Navigation Technologies, Inc. | Percutaneous registration apparatus and method for use in computer-assisted surgical navigation |
US6630937B2 (en) | 1997-10-30 | 2003-10-07 | University Of South Florida | Workstation interface for use in digital mammography and associated methods |
US6181817B1 (en) * | 1997-11-17 | 2001-01-30 | Cornell Research Foundation, Inc. | Method and system for comparing data objects using joint histograms |
US6246790B1 (en) | 1997-12-29 | 2001-06-12 | Cornell Research Foundation, Inc. | Image indexing using color correlograms |
US6072904A (en) * | 1997-12-31 | 2000-06-06 | Philips Electronics North America Corp. | Fast image retrieval using multi-scale edge representation of images |
JP3688879B2 (en) * | 1998-01-30 | 2005-08-31 | 株式会社東芝 | Image recognition apparatus, image recognition method, and recording medium therefor |
US6181414B1 (en) * | 1998-02-06 | 2001-01-30 | Morphometrix Technologies Inc | Infrared spectroscopy for medical imaging |
US6282307B1 (en) | 1998-02-23 | 2001-08-28 | Arch Development Corporation | Method and system for the automated delineation of lung regions and costophrenic angles in chest radiographs |
US6434261B1 (en) * | 1998-02-23 | 2002-08-13 | Board Of Regents, The University Of Texas System | Method for automatic detection of targets within a digital image |
US6088473A (en) * | 1998-02-23 | 2000-07-11 | Arch Development Corporation | Method and computer readable medium for automated analysis of chest radiograph images using histograms of edge gradients for false positive reduction in lung nodule detection |
US6526168B1 (en) * | 1998-03-19 | 2003-02-25 | The Regents Of The University Of California | Visual neural classifier |
US6067373A (en) * | 1998-04-02 | 2000-05-23 | Arch Development Corporation | Method, system and computer readable medium for iterative image warping prior to temporal subtraction of chest radiographs in the detection of interval changes |
US6240423B1 (en) * | 1998-04-22 | 2001-05-29 | Nec Usa Inc. | Method and system for image querying using region based and boundary based image matching |
US6310967B1 (en) * | 1998-04-29 | 2001-10-30 | University Of South Florida | Normal and abnormal tissue identification system and method for medical images such as digital mammograms |
US6996549B2 (en) * | 1998-05-01 | 2006-02-07 | Health Discovery Corporation | Computer-aided image analysis |
US6282305B1 (en) * | 1998-06-05 | 2001-08-28 | Arch Development Corporation | Method and system for the computerized assessment of breast cancer risk |
US6138045A (en) | 1998-08-07 | 2000-10-24 | Arch Development Corporation | Method and system for the segmentation and classification of lesions |
US6442287B1 (en) * | 1998-08-28 | 2002-08-27 | Arch Development Corporation | Method and system for the computerized analysis of bone mass and structure |
US6112112A (en) * | 1998-09-18 | 2000-08-29 | Arch Development Corporation | Method and system for the assessment of tumor extent in magnetic resonance images |
US6146830A (en) | 1998-09-23 | 2000-11-14 | Rosetta Inpharmatics, Inc. | Method for determining the presence of a number of primary targets of a drug |
US6629065B1 (en) | 1998-09-30 | 2003-09-30 | Wisconsin Alumni Research Foundation | Methods and apparata for rapid computer-aided design of objects in virtual reality and other environments |
US6226636B1 (en) * | 1998-11-20 | 2001-05-01 | Philips Electronics North America Corp. | System for retrieving images using a database |
US6574378B1 (en) | 1999-01-22 | 2003-06-03 | Kent Ridge Digital Labs | Method and apparatus for indexing and retrieving images using visual keywords |
US6546137B1 (en) * | 1999-01-25 | 2003-04-08 | Siemens Corporate Research, Inc. | Flash system for fast and accurate pattern localization |
US7272593B1 (en) * | 1999-01-26 | 2007-09-18 | International Business Machines Corporation | Method and apparatus for similarity retrieval from iterative refinement |
US6424332B1 (en) * | 1999-01-29 | 2002-07-23 | Hunter Innovations, Inc. | Image comparison apparatus and method |
US6801645B1 (en) * | 1999-06-23 | 2004-10-05 | Icad, Inc. | Computer aided detection of masses and clustered microcalcifications with single and multiple input image context classification strategies |
CA2381653C (en) | 1999-08-09 | 2010-10-26 | Wake Forest University | A method and computer-implemented procedure for creating electronic, multimedia reports |
US6785410B2 (en) * | 1999-08-09 | 2004-08-31 | Wake Forest University Health Sciences | Image reporting method and system |
US6941323B1 (en) | 1999-08-09 | 2005-09-06 | Almen Laboratories, Inc. | System and method for image comparison and retrieval by enhancing, defining, and parameterizing objects in images |
US6379302B1 (en) | 1999-10-28 | 2002-04-30 | Surgical Navigation Technologies Inc. | Navigation information overlay onto ultrasound imagery |
US6804683B1 (en) * | 1999-11-25 | 2004-10-12 | Olympus Corporation | Similar image retrieving apparatus, three-dimensional image database apparatus and method for constructing three-dimensional image database |
JP2001157675A (en) * | 1999-12-02 | 2001-06-12 | Fuji Photo Film Co Ltd | Method and apparatus for displaying image |
US6553356B1 (en) * | 1999-12-23 | 2003-04-22 | University Of Pittsburgh - Of The Commonwealth System Of Higher Education | Multi-view computer-assisted diagnosis |
US6898303B2 (en) * | 2000-01-18 | 2005-05-24 | Arch Development Corporation | Method, system and computer readable medium for the two-dimensional and three-dimensional detection of lesions in computed tomography scans |
US20020009215A1 (en) * | 2000-01-18 | 2002-01-24 | Arch Development Corporation | Automated method and system for the segmentation of lung regions in computed tomography scans |
US6901156B2 (en) * | 2000-02-04 | 2005-05-31 | Arch Development Corporation | Method, system and computer readable medium for an intelligent search workstation for computer assisted interpretation of medical images |
US6549646B1 (en) * | 2000-02-15 | 2003-04-15 | Deus Technologies, Llc | Divide-and-conquer method and system for the detection of lung nodule in radiological images |
US6432053B1 (en) * | 2000-02-18 | 2002-08-13 | Advanced Diagnostics, Inc. | Process for non-invasively determining the dimensions of a lesion |
WO2001064124A1 (en) * | 2000-03-01 | 2001-09-07 | Surgical Navigation Technologies, Inc. | Multiple cannula image guided tool for image guided procedures |
US7035445B2 (en) * | 2000-03-06 | 2006-04-25 | Fuji Photo Film Co., Ltd. | Image position matching method, apparatus and storage medium |
US6535756B1 (en) * | 2000-04-07 | 2003-03-18 | Surgical Navigation Technologies, Inc. | Trajectory storage apparatus and method for surgical navigation system |
US7027642B2 (en) * | 2000-04-28 | 2006-04-11 | Orametrix, Inc. | Methods for registration of three-dimensional frames to create three-dimensional virtual models of objects |
US7085400B1 (en) | 2000-06-14 | 2006-08-01 | Surgical Navigation Technologies, Inc. | System and method for image based sensor calibration |
US6654728B1 (en) | 2000-07-25 | 2003-11-25 | Deus Technologies, Llc | Fuzzy logic based classification (FLBC) method for automated identification of nodules in radiological images |
WO2002023483A2 (en) | 2000-09-14 | 2002-03-21 | Leland Stanford Junior University | Technique for manipulating medical images |
AU2001291175A1 (en) * | 2000-09-21 | 2002-04-02 | Md Online Inc. | Medical image processing systems |
US6524246B1 (en) | 2000-10-13 | 2003-02-25 | Sonocine, Inc. | Ultrasonic cellular tissue screening tool |
JP4311598B2 (en) * | 2000-11-20 | 2009-08-12 | 富士フイルム株式会社 | Abnormal shadow detection method and apparatus |
US6470092B1 (en) | 2000-11-21 | 2002-10-22 | Arch Development Corporation | Process, system and computer readable medium for pulmonary nodule detection using multiple-templates matching |
US6577752B2 (en) | 2001-06-15 | 2003-06-10 | Arch Development Corporation | Automated method and system for the delineation of the chest wall in computed tomography scans for the assessment of pleural disease |
US6795521B2 (en) | 2001-08-17 | 2004-09-21 | Deus Technologies Llc | Computer-aided diagnosis system for thoracic computer tomography images |
US6728334B1 (en) * | 2001-10-24 | 2004-04-27 | Cornell Research Foundation, Inc. | Automatic detection of pulmonary nodules on volumetric computed tomography images using a local density maximum algorithm |
US7058210B2 (en) * | 2001-11-20 | 2006-06-06 | General Electric Company | Method and system for lung disease detection |
US7072498B1 (en) * | 2001-11-21 | 2006-07-04 | R2 Technology, Inc. | Method and apparatus for expanding the use of existing computer-aided detection code |
US7054473B1 (en) * | 2001-11-21 | 2006-05-30 | R2 Technology, Inc. | Method and apparatus for an improved computer aided diagnosis system |
US20030133601A1 (en) * | 2001-11-23 | 2003-07-17 | University Of Chicago | Automated method and system for the differentiation of bone disease on radiographic images |
US6855114B2 (en) * | 2001-11-23 | 2005-02-15 | Karen Drukker | Automated method and system for the detection of abnormalities in sonographic images |
WO2003067371A2 (en) | 2002-02-08 | 2003-08-14 | Giger Maryellen L | Method and system for risk-modulated diagnosis of disease |
US20030161513A1 (en) * | 2002-02-22 | 2003-08-28 | The University Of Chicago | Computerized schemes for detecting and/or diagnosing lesions on ultrasound images using analysis of lesion shadows |
US6878115B2 (en) * | 2002-03-28 | 2005-04-12 | Ultrasound Detection Systems, Llc | Three-dimensional ultrasound computed tomography imaging system |
SE524957C2 (en) * | 2002-04-12 | 2004-11-02 | Microdrug Ag | Method for dividing and distributing in air of dry powder drug |
US6819790B2 (en) | 2002-04-12 | 2004-11-16 | The University Of Chicago | Massive training artificial neural network (MTANN) for detecting abnormalities in medical images |
US6707878B2 (en) * | 2002-04-15 | 2004-03-16 | General Electric Company | Generalized filtered back-projection reconstruction in digital tomosynthesis |
US6724856B2 (en) * | 2002-04-15 | 2004-04-20 | General Electric Company | Reprojection and backprojection methods and algorithms for implementation thereof |
US6882700B2 (en) * | 2002-04-15 | 2005-04-19 | General Electric Company | Tomosynthesis X-ray mammogram system and method with automatic drive system |
US6748047B2 (en) * | 2002-05-15 | 2004-06-08 | General Electric Company | Scatter correction method for non-stationary X-ray acquisitions |
US7418123B2 (en) * | 2002-07-12 | 2008-08-26 | University Of Chicago | Automated method and system for computerized image analysis for prognosis |
US20040077952A1 (en) * | 2002-10-21 | 2004-04-22 | Rafter Patrick G. | System and method for improved diagnostic image displays |
US7244230B2 (en) * | 2002-11-08 | 2007-07-17 | Siemens Medical Solutions Usa, Inc. | Computer aided diagnostic assistance for medical imaging |
US6999549B2 (en) * | 2002-11-27 | 2006-02-14 | Ge Medical Systems Global Technology, Llc | Method and apparatus for quantifying tissue fat content |
US7298883B2 (en) | 2002-11-29 | 2007-11-20 | University Of Chicago | Automated method and system for advanced non-parametric classification of medical images and lesions |
US6937776B2 (en) | 2003-01-31 | 2005-08-30 | University Of Chicago | Method, system, and computer program product for computer-aided detection of nodules with three dimensional shape enhancement filters |
US7848558B2 (en) | 2003-02-14 | 2010-12-07 | The University Of Chicago | Method and system for fractal-based analysis of medical image texture |
US6909795B2 (en) * | 2003-06-16 | 2005-06-21 | R2 Technology, Inc. | Communicating computer-aided detection results in a standards-based medical imaging environment |
US7201772B2 (en) | 2003-07-08 | 2007-04-10 | Ventor Technologies, Ltd. | Fluid flow prosthetic device |
US7769215B2 (en) | 2003-11-28 | 2010-08-03 | The University Of Chicago | Method, system, and medium for prevalence-based computerized analysis of medical images and information |
US7444011B2 (en) * | 2004-02-10 | 2008-10-28 | University Of Chicago | Imaging system performing substantially exact reconstruction and using non-traditional trajectories |
WO2005079306A2 (en) * | 2004-02-13 | 2005-09-01 | University Of Chicago | Method, system, and computer software product for feature-based correlation of lesions from multiple images |
US7983732B2 (en) * | 2004-02-13 | 2011-07-19 | The University Of Chicago | Method, system, and computer software product for automated identification of temporal patterns with high initial enhancement in dynamic magnetic resonance breast imaging |
US7447670B1 (en) | 2005-09-08 | 2008-11-04 | Hrl Laboratories, Llc | Methods for monitoring conflicts in inference systems |
US8121360B2 (en) | 2006-07-31 | 2012-02-21 | Siemens Medical Solutions Usa, Inc. | Computer aided detection and decision support |
-
1999
- 1999-08-09 US US09/370,366 patent/US6941323B1/en not_active Expired - Fee Related
-
2000
- 2000-08-09 WO PCT/US2000/021735 patent/WO2001011489A2/en active Application Filing
- 2000-08-09 AU AU65338/00A patent/AU6533800A/en not_active Abandoned
-
2004
- 2004-12-07 US US11/007,062 patent/US7483919B2/en not_active Expired - Fee Related
-
2007
- 2007-08-20 US US11/841,888 patent/US8775451B2/en not_active Expired - Fee Related
Also Published As
Publication number | Publication date |
---|---|
US7483919B2 (en) | 2009-01-27 |
US20080037876A1 (en) | 2008-02-14 |
US6941323B1 (en) | 2005-09-06 |
WO2001011489A2 (en) | 2001-02-15 |
US20050149360A1 (en) | 2005-07-07 |
WO2001011489A3 (en) | 2004-04-22 |
US8775451B2 (en) | 2014-07-08 |
AU6533800A (en) | 2001-03-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6941323B1 (en) | System and method for image comparison and retrieval by enhancing, defining, and parameterizing objects in images | |
CN111178449B (en) | Liver cancer image classification method combining computer vision characteristics and imaging omics characteristics | |
US20090082637A1 (en) | Multi-modality fusion classifier with integrated non-imaging factors | |
Pan et al. | Accurate segmentation of nuclei in pathological images via sparse reconstruction and deep convolutional networks | |
Kumar et al. | Detection and classification of cancer from microscopic biopsy images using clinically significant and biologically interpretable features | |
US7492937B2 (en) | System and method for identifying objects of interest in image data | |
Singh et al. | Histopathological image analysis for breast cancer detection using cubic SVM | |
JP2004536367A (en) | Computer-based image analysis | |
US20210216745A1 (en) | Cell Detection Studio: a system for the development of Deep Learning Neural Networks Algorithms for cell detection and quantification from Whole Slide Images | |
Kurmi et al. | Content-based image retrieval algorithm for nuclei segmentation in histopathology images: CBIR algorithm for histopathology image segmentation | |
Ibrahim et al. | Deep learning based Brain Tumour Classification based on Recursive Sigmoid Neural Network based on Multi-Scale Neural Segmentation | |
Pezeshki et al. | Mass classification of mammograms using fractal dimensions and statistical features | |
JP2001117936A (en) | Three-dimensional similar image retrieving device | |
Ali et al. | Automatic cancerous tissue classification using discrete wavelet transformation and support vector machine | |
Wu et al. | Semiautomatic segmentation of glioma on mobile devices | |
de Brito Silva et al. | Classification of breast masses in mammograms using geometric and topological feature maps and shape distribution | |
Leite et al. | 3D texture-based classification applied on brain white matter lesions on MR images | |
Fooladi et al. | Segmenting the lesion area of brain tumor using convolutional neural networks and fuzzy k-means clustering | |
Valliappan et al. | A theoretical methodology and prototype implementation for detection segmentation classification of digital mammogram tumor by machine learning and problem solving approach | |
Dobrescu et al. | Image processing applications based on texture and fractal analysis | |
Liu et al. | Deep learning technology in pathological image analysis of breast tissue | |
Hibbard et al. | Automated recognition and mapping of immunolabelled neurons in the developing brain | |
Catanzariti et al. | Computer aided detection of clustered microcalcifications in digitized mammograms using Gabor functions | |
Afifi et al. | An efficient system for Melanoma diagnosis in dermoscopic images | |
Iqbal et al. | AMIAC: adaptive medical image analyzes and classification, a robust self-learning framework |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A2 Designated state(s): AE AG AL AM AT AT AU AZ BA BB BG BR BY BZ CA CH CN CR CU CZ CZ DE DE DK DK DM DZ EE EE ES FI FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SK SL TJ TM TR TT TZ UA UG UZ VN YU ZA ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A2 Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
AK | Designated states |
Kind code of ref document: C2 Designated state(s): AE AG AL AM AT AT AU AZ BA BB BG BR BY BZ CA CH CN CR CU CZ CZ DE DE DK DK DM DZ EE EE ES FI FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SK SL TJ TM TR TT TZ UA UG UZ VN YU ZA ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: C2 Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG |
|
COP | Corrected version of pamphlet |
Free format text: PAGES 1/9-9/9, DRAWINGS, REPLACED BY NEW PAGES 1/9-9/9; DUE TO LATE TRANSMITTAL BY THE RECEIVING OFFICE |
|
REG | Reference to national code |
Ref country code: DE Ref legal event code: 8642 |
|
122 | Ep: pct application non-entry in european phase | ||
NENP | Non-entry into the national phase |
Ref country code: JP |