WO2006057740A2 - Detecting and classifying lesions in ultrasound images - Google Patents

Detecting and classifying lesions in ultrasound images Download PDF

Info

Publication number
WO2006057740A2
WO2006057740A2 PCT/US2005/037843 US2005037843W WO2006057740A2 WO 2006057740 A2 WO2006057740 A2 WO 2006057740A2 US 2005037843 W US2005037843 W US 2005037843W WO 2006057740 A2 WO2006057740 A2 WO 2006057740A2
Authority
WO
WIPO (PCT)
Prior art keywords
regions
lesion
candidate lesion
region
ultrasound image
Prior art date
Application number
PCT/US2005/037843
Other languages
French (fr)
Other versions
WO2006057740A3 (en
Inventor
Jiebo Luo
Zhimin Huo
Wei Hao
Original Assignee
Eastman Kodak Company
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Eastman Kodak Company filed Critical Eastman Kodak Company
Priority to JP2007543061A priority Critical patent/JP2008520345A/en
Priority to EP05811996A priority patent/EP1815429A2/en
Publication of WO2006057740A2 publication Critical patent/WO2006057740A2/en
Publication of WO2006057740A3 publication Critical patent/WO2006057740A3/en

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/08Detecting organic movements or changes, e.g. tumours, cysts, swellings
    • A61B8/0825Detecting organic movements or changes, e.g. tumours, cysts, swellings for diagnosis of the breast, e.g. mammography
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/08Detecting organic movements or changes, e.g. tumours, cysts, swellings
    • A61B8/0833Detecting organic movements or changes, e.g. tumours, cysts, swellings involving detecting or locating foreign bodies or organic structures
    • A61B8/085Detecting organic movements or changes, e.g. tumours, cysts, swellings involving detecting or locating foreign bodies or organic structures for locating body or organic structures, e.g. tumours, calculi, blood vessels, nodules
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10132Ultrasound image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30068Mammography; Breast
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S128/00Surgery
    • Y10S128/92Computer assisted medical diagnostics
    • Y10S128/922Computer assisted medical diagnostics including image analysis

Definitions

  • the invention relates to digital image processing of ultrasound images, and more particularly to detecting and classifying lesions within such images.
  • Breast cancer risk assessment provides an opportunity to devise surveillance plans that may include enhanced screening for individuals at increased risk of breast cancer.
  • Computerized analysis of mammographic parenchymal patterns may provide an objective and quantitative characterization and classification of these patterns, which may be associated with breast cancer risk.
  • Computerized assessment of breast cancer risk that are based on the analysis of mammograms alone or combined with epidemiologic risk factors (for example, age) can serve as an alternative to current existing clinical methods, which are costly and/or information-dependent, in predicting breast cancer risk.
  • the breast is composed primarily of two components: fibroglandular tissue and fatty tissue.
  • An average breast is comprised of 50% fibroglandular tissue and 50% fat.
  • Fibroglandular tissue is a mixture of fibrous connective tissue and the glandular epithelial cells that line the ducts of the breast (the parenchyma).
  • Major breast diseases develop from the terminal ductal lobular units of the breast, and arise predominantly from the epithelial cells that line the ducts. However, the fibrous or connective tissue can also be involved. It is thought by some individuals that malignant breast disease develops through a process that starts with epithelial hyperplasia, i.e., an increase in the number of epithelial cells. Epithelial hyperplasia can progress to atypical hyperplasia in which the epithelial cells not only increase in number, but also change in a way that is not normal for these cells. The process, at this stage, is believed to be reversible.
  • carcinoma-in-situ can be made, in which there is no invasion of malignant cells outside of the duct.
  • the process of malignant transformation is considered irreversible at this stage.
  • the cancer cells break out of the ductal walls and invade the surrounding stromal tissue, and at this point the disease is called infiltrating or invasive carcinoma.
  • breast carcinomas can be seen on a mammogram as a mass, a cluster of tiny calcifications, or a combination of both.
  • Other mammographic abnormalities are of lesser specificity and prevalence than masses and/or calcifications, and include skin or nipple changes, abnormalities in the axilla, asymmetric density, and architectural distortion.
  • US Patent No. 6,282,305 (Huo et al) is directed to a method and system for the computerized assessment of breast cancer risk, wherein a digital image of a breast is obtained and at least one feature is extracted from a region of interest in the digital image.
  • the extracted features are compared with a predetermined model associating patterns of the extracted features with a risk estimate derived from corresponding feature patterns associated with a predetermined model based on gene carrier information or clinical information, or both gene carrier information and clinical information, and a risk classification index is output as a result of the comparison.
  • Preferred features to be extracted from the digital image include 1) one or more features based on absolute values of gray levels of pixels in said region of interest, 2) one or more features based on gray-level histogram analysis of pixels in said region of interest; 3) one or more features based on Fourier analysis of pixel values in said region of interest; and 4) one or more features based on a spatial relationship among gray levels of pixels within the region of interest.
  • US Patent No. 5,984,870 (Giger et al.) is directed to a method for the analysis of a lesion existing in anatomical tissue, comprising the steps of (a) obtaining first digital image data derived from an ultrasound image of the anatomical tissue in which the lesion exists; (b) determining a location of the lesion in relation to the first digital data; (c) selecting for feature extraction analysis at least one of 1) a region of interest on the margin of the lesion, and 2) a region of interest which includes the lesion and a region of interest which is in the surrounding vicinity of the lesion, and 3) a region of interest which includes the lesion and a region of interest which is on the margin of the lesion; (d) extracting from each selected region of interest selected in said selecting step at least one first feature that characterize a lesion within said first image data; and (e) characterizing said lesion based on the extracted at least one first feature.
  • a difficulty associated with a computerized system for detecting and diagnosing breast lesions is segmentation of the lesion regions from the surrounding tissues.
  • segmentation is obtained by manual outlining the lesions using a graphic user interface, for example, US Patent No. 5,984,870 (Giger et al.).
  • This manual procedure is labor-intensive, can disrupt full automation, and can be prone to human error, inconsistency and subjectivity.
  • the resulting inaccuracy in the outline of the lesion has adverse effect on the sub ⁇ sequent computerized diagnosis because features computed from inaccurate lesion outline (e.g., those taught by US Patent No. 5,984,870 (Giger et al.)), tend to be inaccurate for lesion classification.
  • Pixel-based, edge-based, region-based, and model-based segmentation techniques are known in medical image processing. Each approach has its own limitations. For example, pixel-based segmentation techniques tend to have difficulties when there is a significant amount of noise in the image; edge- based techniques tend to experience problems when the boundary of the object is not well defined and when the image contrast is poor; while model-based techniques tend to fail when there is a significant amount of variation in the shape and appearance of the object of interest. Region-growing techniques require a good seed point (typically provided by manual interaction) and are subject to critical errors when adjoining objects closely match an object of interest in their appearance.
  • the present invention provides a lesion segmentation and classification method wherein segmentation is automatic and relatively insensitive to variations in image noise and target appearance (color and shape). Further, the methods can provide a fully automatic system for segmenting and classifying lesions by cascading automatic lesion segmentation and automatic lesion classification.
  • An object of the present invention is to provide an lesion segmentation method in which segmentation is automatic and relatively insensitive to variations in image noise and target appearance (e.g., color and shape).
  • Another object of the present invention is to provide an automatic system for segmenting and classifying lesions by cascading automatic lesion segmentation and automatic lesion classification.
  • the present invention provides a method, computer program, and system, in which a lesion region is automatically located within an ultrasound (mammographic) image, and such a lesion region is classified according to a plurality of features computed upon the lesion region itself to decide whether the lesion is benign or malignant.
  • an improved lesion diagnosis method and apparatus in which segmentation is automatic and is relatively insensitive to in image noise and target appearance (e.g., color and shape).
  • a method for determining a candidate lesion region within an ultrasound image includes the steps of: accessing a digital ultrasound image of anatomical tissue; segmenting spatially contiguous pixels in the ultrasound image into a plurality of regions in accordance with substantially similar intensity values and spatial smoothness constraints; and selecting, from the plurality of regions, one or more candidate lesion regions having an intensity value lower than a pre ⁇ determined intensity value.
  • the one or more candidate lesion region is classified into at least one of the following classes: benign, malignant, or unknown.
  • FIG. 1 is a block diagram illustrating the steps of an embodiment of the method in accordance with the present invention.
  • FIG. 2 is a block diagram illustrating steps of segmenting lesion regions from normal tissue regions.
  • FIGS. 3A-3D are examples of ultrasound images comprising both lesion (benign and malignant) and normal tissue regions.
  • FIGS. 4A-4D correspond with FIGS. 3A-3D, and are examples of segmented lesion regions.
  • FIGS. 5A-5D correspond with FIGS. 3 A-3D and FIGS. 4A-4D, and show examples of classified lesion regions.
  • FIGS. 6 A and 6B shows additional examples of classified lesion regions.
  • the present invention also relates to systems including specific pieces of apparatus for performing the operations described herein.
  • Apparatus such as a programmable computer may be specially constructed for the required purposes, or may comprise a general purpose computer selectively activated or reconfigured by a computer program stored in the computer.
  • a computer program may be stored in a computer readable storage medium, such as, but is not limited to, any type of disk including floppy disks, optical disks, CD-ROMs, and magnetic- optical disks, read-only memories (ROMs), random access memories (RAMs) such as Dynamic RAM (DRAM), EPROMs, EEPROMs, magnetic or optical cards, or any type of media suitable for storing electronic instructions, and each of the above storage components is coupled to a computer system bus.
  • a machine-readable medium includes any mechanism for storing or transmitting information in a form readable by a machine (e.g., a computer).
  • a machine-readable medium includes read only memory ("ROM”); random access memory (“RAM”); magnetic disk storage media; optical storage media; flash memory devices; electrical, optical, acoustical or other form of propagated signals (e.g., carrier waves, infrared signals, digital signals, etc.); etc.
  • references to "in a particular embodiment” and the like refer to features that are present in at least one embodiment of the invention.
  • references to "an embodiment” or “particular embodiments” or the like do not necessarily refer to the same embodiment or embodiments; however, such embodiments are not mutually exclusive, unless so indicated or as are readily apparent to one of skill in the art.
  • the invention is thus inclusive of combinations of the embodiments described herein.
  • tissue is used herein to refer to a single abnormal mass within otherwise normal anatomical tissue.
  • mammographic image is used herein to refer to an image that shows a portion of a patient's breast, which includes normal breast tissue and in some cases lesions.
  • an input ultrasound image 10 is accessed/acquired/provided for analysis.
  • Region segmentation 20 is performed on the input ultrasound image to produce a number of regions, each having a distinctive intensity characteristic. Among segmented regions, some may be selected as candidate regions 30 for lesions.
  • a plurality of features are extracted 40 for each candidate region. Each candidate region is then classified 50, for example, as a benign lesion, or a malignant lesion, or in some cases, not a lesion.
  • pre-process 15 the input image prior to region segmentation 20.
  • FIG. 2 there is shown an block diagram of a suitable region segmentation method (step 20) according to a preferred embodiment of the present invention.
  • the method is modified from the adaptive Bayesian color segmentation algorithm described in Luo et al., "Towards physics- based segmentation of photographic color images," Proceedings of the IEEE International Conference on Image Processing, 1997.
  • Region segmentation generates a tractable number of physically coherent regions, in terms of similar intensity patterns, of arbitrary shape.
  • Similar intensity patterns can refer to similar intensity values, or similar intensity textures, and/or a combination of both.
  • the regions having similar intensity patterns fall into the same intensity class, but otherwise are spatially separated and have different shapes.
  • this segmentation method is preferred, it will be appreciated that a person of ordinary skill in the art can use a different segmentation method to obtain regions of arbitrary shape without departing from the scope of the present invention. Segmentation of arbitrarily shaped regions facilitates accurate measurement of the size, location, shape, as well as internal characteristics of potential objects.
  • An initial segmentation of the image into regions is obtained S50.
  • An intensity histogram of the image is computed and then partitioned into a plurality of clusters that correspond to distinctive, prominent intensities in the image.
  • Each pixel of the image is classified to the closest cluster in the intensity space according to a preferred intensity distance metric with respect to the mean values of the intensity clusters (Luo et al., "Towards physics-based segmentation of photographic color images," Proceedings of the IEEE International Conference on Image Processing, 1997).
  • This classification process results in an initial segmentation of the image.
  • a neighborhood window is placed at each pixel in order to determine what neighborhood pixels are used to compute the local intensity histogram for this pixel.
  • the window size is initially set at the size of the entire image S52, so that the local intensity histogram is the same as the one for the entire image and does not need to be recomputed.
  • an iterative procedure is performed between two alternating processes: re-computing S54 the local mean values of each intensity class based on the current segmentation, and re-classifying the pixels according to the updated local mean values of intensity classes S56.
  • This iterative procedure is performed until a convergence is reached S60.
  • the strength of the spatial constraints can be adjusted in a gradual manner S58 (for example, the value of ⁇ , which indicates the strength of the spatial constraints, is increased linearly with each iteration).
  • the window used to estimate the local mean values for intensity classes is reduced by half in size S62.
  • the iterative procedure is repeated for the reduced window size to allow more accurate estimation of the local mean values for intensity classes.
  • This mechanism introduces spatial adaptivity into the segmentation process.
  • Applicants set/pre-defined the number of distinctive intensity classes to be 3, roughly corresponding to lesion, normal tissue, and unknown structure.
  • An advantage of the preferred segmentation method described in Figure. 2 is that small lesions are not lost because of its size, which is often a challenge for other non-preferred segmentation method, such as the well-known classic k-means algorithm. Without proper region segmentation, can be difficult to perform object recognition and proper assessment of such properties as size and shape.
  • the result of segmentation contains labels of regions, i.e., each pixel is represented by its label value.
  • the image of labeled regions is referred to as a label image, or segmentation map.
  • pre-process 15 the input image prior to region segmentation 20. More particularly, it may be desirable to enhance the contrast between lesion and normal tissues before segmentation 20 is applied.
  • a gamma adjustment of 2.0 is used.
  • Alternative contrast enhancement can be performed for the same purpose.
  • I(x,y) denote an intensity value if a pixel coordinates (x,y)
  • segmented regions may be selected as candidate regions for lesions in step 30.
  • lesions typically appear as relatively darker regions in an ultrasound image because they tend to form boundaries against the normal tissue and have higher water content. Therefore, darker regions are selected as lesion candidates in the present invention.
  • regions with an average intensity lower than a pre-determined value are selected as candidate regions.
  • Additional criteria can be used to further eliminate/reduce unlikely candidate regions. For example, it is assumed that an operator would usually place the lesion in the center of the view. In other words, a lesion, if present, would be completely contained in the image, therefore regions that touch image borders are unlikely to be the lesion and thus can be eliminated from further analysis.
  • Candidate selection facilitates the subsequent lesion classification by focusing on a small number of regions.
  • a set of features are extracted to characterize each region in step 40.
  • the list of features can comprise one or more categories. In a particular application, three categories are employed: shape features, interior features, and boundary features. Other categories can be included, for example, shadow features. Other categories may be known to those skilled in the art.
  • raw measurements of features maybe converted such that their values are normalized to be within [0, 1.0], by functions with appropriate nonlinearity characteristics. Normalization is well known for persons skilled in the art. Those who are skilled in the art can recognize alternative types of features, without departing from the scope and spirit of the present invention.
  • lesions that have compact, elliptical shape and smooth contour tend to be benign.
  • the basic features are height, width, and perimeter, and size.
  • An aspect ratio of a region is the major axis/minor axis, or height over width, of the region. This feature detector is used to discount long narrow shapes from being considered a lesion.
  • malignant lesions are harder to compress and may appear taller than wider along the direction of ultrasound scattering, i.e., the direction of the transducer. Benign lesions or non-lesions tend to appear as wider than taller. Such characteristics can be indicated by the aspect ratio as well.
  • the compactness is the square of the perimeter of the region divided by the area of region.
  • a sigmoid belief function maps the circularity to a belief value.
  • the interior features include mean and standard deviation of the intensity values within the interior of a candidate region.
  • the boundary features include mean and standard deviation of the intensity gradient values along the boundary of a candidate region.
  • the shadow feature is used to measure the shadow cast by a candidate lesion region.
  • the features are used to classify whether a candidate region is a benign lesion, or a malignant lesion, or not a lesion at all, in step 50.
  • the third class of "not a lesion" is provided should the candidate selection step 30 occasionally include non-lesion regions to be selected because of an ambiguity. Examples of such cases, as classified by medical professionals, need to be collected as so-called training examples.
  • the features and the associated classification labels of these training examples are the basis for deriving a classifier, which is then used to classify a novel candidate region not seen by the classifier.
  • Those skilled in the art can choose from many existing classifiers, including a decision tree, a neural network, a Bayesian classifier, or the like.
  • a decision tree is used. If a confidence value is desired, alternative classifiers such as a neural network or a Bayesian classifier should be used.
  • Figures 3A-3D shows four example ultrasound images of the breast, hi Figures 4A-4D, corresponding to Figures 3 A-3D, each image is segmented into three intensity classes, with the intensity class of the darkest regions indicated by the highest brightness in the segmented images.
  • lesion candidates are selected from these regions, hi Figures 5A-5D, all the candidate regions are classified into benign (e.g., part 100), malignant (e.g., part 110), or not a lesion (e.g., part 120). More particularly, Figure 5 A shows no lesion, Figure 5B shows a benign lesion, Figure 5 C shows a malignant lesion, and Figure 5D shows a malignant lesion.
  • a colored bar 200 (shown proximate Figures 5A-5D) can be displayed next to a region upon which the operator places the cursor, wherein the shades (color saturation) of the bar gives a visual presentation of the lesions classification and the associated confidence value.
  • a tick 201 (or other symbol) could point to the range of deep red of the color bar when the lesion is classified as malignant with a high confidence value, or the range of light blue of the color bar when lesion is classified as benign but with a low confidence value, or the like.
  • Figures 6A and 6B provide additional examples, showing a cursor operable by an operator and wherein a candidate region is outlined.
  • the present invention is described in a preferred embodiment that is related to lesions in the breast, it will be understood that variations and modifications can be effected within the spirit and scope of the invention for lesions in other tissues, such as the liver.
  • the present invention may be practiced online, real-time, or offline.
  • it could provide the operator with instant warning or feedback
  • hi the offline situation it could analyze a database of pre-recorded images.
  • it may be desirable to display the ultrasound image and an outline of the segmented lesion region following automatic segmentation, and accept adjustment of the region outline by an operator using a graphical user interface.
  • a computer program product may include one or more storage media, for example; magnetic storage media such as magnetic disk (such as a floppy disk) or magnetic tape; optical storage media such as optical disk, optical tape, or machine readable bar code; solid-state electronic storage devices such as random access memory (RAM), or read-only memory (ROM); or any other physical device or media employed to store a computer program having instructions for controlling one or more computers to practice the method according to the present invention.
  • magnetic storage media such as magnetic disk (such as a floppy disk) or magnetic tape
  • optical storage media such as optical disk, optical tape, or machine readable bar code
  • solid-state electronic storage devices such as random access memory (RAM), or read-only memory (ROM); or any other physical device or media employed to store a computer program having instructions for controlling one or more computers to practice the method according to the present invention.
  • the system of the invention includes a programmable computer having a microprocessor, computer memory, and a computer program stored in said computer memory for performing the steps of the method.
  • the computer has a memory interface operatively connected to the microprocessor. This can be a port, such as a USB port, over a drive that accepts removable memory, or some other device that allows access to camera memory.
  • the system includes a digital camera that has memory that is compatible with the memory interface. A photographic film camera and scanner can be used in place of the digital camera, if desired.
  • a graphical user interface (GUI) and user input unit, such as a mouse and keyboard can be provided as part of the computer.
  • GUI graphical user interface

Abstract

A method for determining a candidate lesion region within an ultrasound image. The method includes the steps of: accessing a digital ultrasound image of anatomical tissue; segmenting spatially contiguous pixels in the ultrasound image into a plurality of regions in accordance with substantially similar intensity values and spatial smoothness constraints; and selecting, from the plurality of regions, one or more candidate lesion regions having an intensity value lower than a pre-determined intensity value. In one arrangement, the one or more candidate lesion region is classified into at least one of the following classes: benign, malignant, or unknown.

Description

DETECTING AND CLASSIFYING LESIONS IN ULTRASOUND IMAGES
FIELD OF THE INVENTION
The invention relates to digital image processing of ultrasound images, and more particularly to detecting and classifying lesions within such images.
BACKGROUND OF THE INVENTION
Breast cancer risk assessment provides an opportunity to devise surveillance plans that may include enhanced screening for individuals at increased risk of breast cancer. Computerized analysis of mammographic parenchymal patterns may provide an objective and quantitative characterization and classification of these patterns, which may be associated with breast cancer risk. Computerized assessment of breast cancer risk that are based on the analysis of mammograms alone or combined with epidemiologic risk factors (for example, age) can serve as an alternative to current existing clinical methods, which are costly and/or information-dependent, in predicting breast cancer risk.
The breast is composed primarily of two components: fibroglandular tissue and fatty tissue. An average breast is comprised of 50% fibroglandular tissue and 50% fat. Fibroglandular tissue is a mixture of fibrous connective tissue and the glandular epithelial cells that line the ducts of the breast (the parenchyma).
Major breast diseases develop from the terminal ductal lobular units of the breast, and arise predominantly from the epithelial cells that line the ducts. However, the fibrous or connective tissue can also be involved. It is thought by some individuals that malignant breast disease develops through a process that starts with epithelial hyperplasia, i.e., an increase in the number of epithelial cells. Epithelial hyperplasia can progress to atypical hyperplasia in which the epithelial cells not only increase in number, but also change in a way that is not normal for these cells. The process, at this stage, is believed to be reversible. Once a certain criterion level of atypia is reached, the diagnosis of carcinoma-in-situ can be made, in which there is no invasion of malignant cells outside of the duct. The process of malignant transformation is considered irreversible at this stage. In the last phase of development, the cancer cells break out of the ductal walls and invade the surrounding stromal tissue, and at this point the disease is called infiltrating or invasive carcinoma.
Most (80%-85%) breast carcinomas can be seen on a mammogram as a mass, a cluster of tiny calcifications, or a combination of both. Other mammographic abnormalities are of lesser specificity and prevalence than masses and/or calcifications, and include skin or nipple changes, abnormalities in the axilla, asymmetric density, and architectural distortion.
Early detection of breast cancer can improve survival rates. Some statistics indicate that the overall five-year survival rate for women diagnosed with breast cancer is 84%, but when found at a small, localized stage, the 5-year survival rate is 97%. At least one study has shown that the use of screening mammography can reduce lesion size and stage at detection, improving the prognosis for survival. Currently, mammography is an established imaging technique for early detection of breast cancer. At least one organization has recommended annual screening mammography for all women over the age of 40.
US Patent No. 6,282,305 (Huo et al) is directed to a method and system for the computerized assessment of breast cancer risk, wherein a digital image of a breast is obtained and at least one feature is extracted from a region of interest in the digital image. The extracted features are compared with a predetermined model associating patterns of the extracted features with a risk estimate derived from corresponding feature patterns associated with a predetermined model based on gene carrier information or clinical information, or both gene carrier information and clinical information, and a risk classification index is output as a result of the comparison. Preferred features to be extracted from the digital image include 1) one or more features based on absolute values of gray levels of pixels in said region of interest, 2) one or more features based on gray-level histogram analysis of pixels in said region of interest; 3) one or more features based on Fourier analysis of pixel values in said region of interest; and 4) one or more features based on a spatial relationship among gray levels of pixels within the region of interest.
US Patent No. 5,984,870 (Giger et al.) is directed to a method for the analysis of a lesion existing in anatomical tissue, comprising the steps of (a) obtaining first digital image data derived from an ultrasound image of the anatomical tissue in which the lesion exists; (b) determining a location of the lesion in relation to the first digital data; (c) selecting for feature extraction analysis at least one of 1) a region of interest on the margin of the lesion, and 2) a region of interest which includes the lesion and a region of interest which is in the surrounding vicinity of the lesion, and 3) a region of interest which includes the lesion and a region of interest which is on the margin of the lesion; (d) extracting from each selected region of interest selected in said selecting step at least one first feature that characterize a lesion within said first image data; and (e) characterizing said lesion based on the extracted at least one first feature.
US Patent Applications No. 2003/0161513 and 2003/0125621 describe similar systems, using analysis of lesion shadows in an ultrasound image and a radial gradient index (RGI) feature in a sonographic image, respectively.
A difficulty associated with a computerized system for detecting and diagnosing breast lesions is segmentation of the lesion regions from the surrounding tissues. Some systems assume that segmentation is obtained by manual outlining the lesions using a graphic user interface, for example, US Patent No. 5,984,870 (Giger et al.). This manual procedure is labor-intensive, can disrupt full automation, and can be prone to human error, inconsistency and subjectivity. The resulting inaccuracy in the outline of the lesion has adverse effect on the sub¬ sequent computerized diagnosis because features computed from inaccurate lesion outline (e.g., those taught by US Patent No. 5,984,870 (Giger et al.)), tend to be inaccurate for lesion classification.
Accordingly, it is desirable to devise an automatic segmentation module to streamline the computerized mammography analysis system.
Pixel-based, edge-based, region-based, and model-based segmentation techniques are known in medical image processing. Each approach has its own limitations. For example, pixel-based segmentation techniques tend to have difficulties when there is a significant amount of noise in the image; edge- based techniques tend to experience problems when the boundary of the object is not well defined and when the image contrast is poor; while model-based techniques tend to fail when there is a significant amount of variation in the shape and appearance of the object of interest. Region-growing techniques require a good seed point (typically provided by manual interaction) and are subject to critical errors when adjoining objects closely match an object of interest in their appearance.
As such, there exists a need for a method which overcomes the limitations of existing methods.
The present invention provides a lesion segmentation and classification method wherein segmentation is automatic and relatively insensitive to variations in image noise and target appearance (color and shape). Further, the methods can provide a fully automatic system for segmenting and classifying lesions by cascading automatic lesion segmentation and automatic lesion classification.
SUMMARY OF THE INVENTION
An object of the present invention is to provide an lesion segmentation method in which segmentation is automatic and relatively insensitive to variations in image noise and target appearance (e.g., color and shape).
Another object of the present invention is to provide an automatic system for segmenting and classifying lesions by cascading automatic lesion segmentation and automatic lesion classification.
These objects are given only by way of illustrative example, and such objects may be exemplary of one or more embodiments of the invention. Other desirable objectives and advantages inherently achieved by the disclosed invention may occur or become apparent to those skilled in the art. The invention is defined by the appended claims. The present invention provides a method, computer program, and system, in which a lesion region is automatically located within an ultrasound (mammographic) image, and such a lesion region is classified according to a plurality of features computed upon the lesion region itself to decide whether the lesion is benign or malignant.
It is an advantageous effect of the invention that an improved lesion diagnosis method and apparatus is provided, in which segmentation is automatic and is relatively insensitive to in image noise and target appearance (e.g., color and shape).
It is a further advantageous effect of the invention to provide a folly automated lesion segmentation and classification method.
According to one aspect of the present invention there is provided a method for determining a candidate lesion region within an ultrasound image. The method includes the steps of: accessing a digital ultrasound image of anatomical tissue; segmenting spatially contiguous pixels in the ultrasound image into a plurality of regions in accordance with substantially similar intensity values and spatial smoothness constraints; and selecting, from the plurality of regions, one or more candidate lesion regions having an intensity value lower than a pre¬ determined intensity value. In one arrangement, the one or more candidate lesion region is classified into at least one of the following classes: benign, malignant, or unknown.
BRIEF DESCRIPTION OF THE DRAWINGS
The above-mentioned and other features and objects of this invention and the manner of attaining them will become more apparent and the invention itself will be better understood by reference to the following description of an embodiment of the invention taken in conjunction with the accompanying figures wherein:
FIG. 1 is a block diagram illustrating the steps of an embodiment of the method in accordance with the present invention. FIG. 2 is a block diagram illustrating steps of segmenting lesion regions from normal tissue regions.
FIGS. 3A-3D are examples of ultrasound images comprising both lesion (benign and malignant) and normal tissue regions.
FIGS. 4A-4D correspond with FIGS. 3A-3D, and are examples of segmented lesion regions.
FIGS. 5A-5D correspond with FIGS. 3 A-3D and FIGS. 4A-4D, and show examples of classified lesion regions.
FIGS. 6 A and 6B shows additional examples of classified lesion regions.
DETAILED DESCRIPTION OF THE INVENTION
The following is a detailed description of the preferred embodiments of the invention, reference being made to the drawings in which the same reference numerals identify the same elements of structure in each of the several figures.
Some portions of the detailed descriptions which follow are presented in terms of algorithms and symbolic representations of operations on data bits within a computer memory. These algorithmic descriptions and representations are used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. An algorithm is, and generally, conceived to be a self-consistent sequence of operations leading to a desired result. The operations are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated. For convenience, these signals as sometimes referred to herein in relation to the underlying information using referents such as bits, values, elements, symbols, characters, terms, numbers, or the like.
It should be recognized, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. Unless specifically stated otherwise as apparent from the following discussion, it is appreciated that throughout the description, discussions utilizing terms such as "processing" or "computing" or "calculating" or "determining" or "displaying" or the like, refer to the action and processes of a computer system, or similar data processing device, that manipulates and transforms data represented as physical (e.g. electronic) quantities within the computer system's registers and memories into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage, transmission or display devices.
The present invention also relates to systems including specific pieces of apparatus for performing the operations described herein. Apparatus such as a programmable computer may be specially constructed for the required purposes, or may comprise a general purpose computer selectively activated or reconfigured by a computer program stored in the computer. Such a computer program may be stored in a computer readable storage medium, such as, but is not limited to, any type of disk including floppy disks, optical disks, CD-ROMs, and magnetic- optical disks, read-only memories (ROMs), random access memories (RAMs) such as Dynamic RAM (DRAM), EPROMs, EEPROMs, magnetic or optical cards, or any type of media suitable for storing electronic instructions, and each of the above storage components is coupled to a computer system bus.
The algorithms and displays presented herein are not inherently related to any particular computer or other apparatus. Various general purpose systems may be used with programs in accordance with the teachings herein, or it may prove convenient to construct more specialized apparatus to perform the methods. The structure for a variety of these systems will appear from the description below. In addition, the present invention is not described with reference to any particular programming language. It will be appreciated that a variety of programming languages may be used to implement the teachings of the invention as described herein.
A machine-readable medium includes any mechanism for storing or transmitting information in a form readable by a machine (e.g., a computer). For example, a machine-readable medium includes read only memory ("ROM"); random access memory ("RAM"); magnetic disk storage media; optical storage media; flash memory devices; electrical, optical, acoustical or other form of propagated signals (e.g., carrier waves, infrared signals, digital signals, etc.); etc.
References to "in a particular embodiment" and the like refer to features that are present in at least one embodiment of the invention. Separate references to "an embodiment" or "particular embodiments" or the like do not necessarily refer to the same embodiment or embodiments; however, such embodiments are not mutually exclusive, unless so indicated or as are readily apparent to one of skill in the art. The invention is thus inclusive of combinations of the embodiments described herein.
The term "lesion" is used herein to refer to a single abnormal mass within otherwise normal anatomical tissue.
The term "mammographic image" is used herein to refer to an image that shows a portion of a patient's breast, which includes normal breast tissue and in some cases lesions.
Elements not specifically shown or described herein may be selected from those known in the art. Certain aspects of the embodiment to be described may be provided in software. Given the method as shown and described according to the invention in the following materials, software not specifically shown, described or suggested herein that is useful for implementation of the invention is conventional and within the ordinary skill in such arts. The method of the invention is useful with, but not limited to, use with color image processing procedures well known to those of skill in the art.
Referring to Figure 1, the present invention is described with reference to the shown block diagram. First, an input ultrasound image 10 is accessed/acquired/provided for analysis. Region segmentation 20 is performed on the input ultrasound image to produce a number of regions, each having a distinctive intensity characteristic. Among segmented regions, some may be selected as candidate regions 30 for lesions. Next, a plurality of features are extracted 40 for each candidate region. Each candidate region is then classified 50, for example, as a benign lesion, or a malignant lesion, or in some cases, not a lesion.
As will be more particularly described below, it may be desirable to pre-process 15 the input image prior to region segmentation 20.
Referring now to Figure 2, there is shown an block diagram of a suitable region segmentation method (step 20) according to a preferred embodiment of the present invention. The method is modified from the adaptive Bayesian color segmentation algorithm described in Luo et al., "Towards physics- based segmentation of photographic color images," Proceedings of the IEEE International Conference on Image Processing, 1997.
Region segmentation generates a tractable number of physically coherent regions, in terms of similar intensity patterns, of arbitrary shape.
Similar intensity patterns can refer to similar intensity values, or similar intensity textures, and/or a combination of both. The regions having similar intensity patterns fall into the same intensity class, but otherwise are spatially separated and have different shapes. Although this segmentation method is preferred, it will be appreciated that a person of ordinary skill in the art can use a different segmentation method to obtain regions of arbitrary shape without departing from the scope of the present invention. Segmentation of arbitrarily shaped regions facilitates accurate measurement of the size, location, shape, as well as internal characteristics of potential objects.
An initial segmentation of the image into regions is obtained S50. An intensity histogram of the image is computed and then partitioned into a plurality of clusters that correspond to distinctive, prominent intensities in the image. Each pixel of the image is classified to the closest cluster in the intensity space according to a preferred intensity distance metric with respect to the mean values of the intensity clusters (Luo et al., "Towards physics-based segmentation of photographic color images," Proceedings of the IEEE International Conference on Image Processing, 1997). This classification process results in an initial segmentation of the image. A neighborhood window is placed at each pixel in order to determine what neighborhood pixels are used to compute the local intensity histogram for this pixel. The window size is initially set at the size of the entire image S52, so that the local intensity histogram is the same as the one for the entire image and does not need to be recomputed.
Next, an iterative procedure is performed between two alternating processes: re-computing S54 the local mean values of each intensity class based on the current segmentation, and re-classifying the pixels according to the updated local mean values of intensity classes S56.
This iterative procedure is performed until a convergence is reached S60. During this iterative procedure, the strength of the spatial constraints can be adjusted in a gradual manner S58 (for example, the value of β, which indicates the strength of the spatial constraints, is increased linearly with each iteration).
After a convergence is reached for a particular window size, the window used to estimate the local mean values for intensity classes is reduced by half in size S62.
The iterative procedure is repeated for the reduced window size to allow more accurate estimation of the local mean values for intensity classes. This mechanism introduces spatial adaptivity into the segmentation process.
Finally, segmentation of the image is obtained when the iterative procedure reaches convergence for the minimum window size S64.
In a preferred embodiment of the present invention, Applicants set/pre-defined the number of distinctive intensity classes to be 3, roughly corresponding to lesion, normal tissue, and unknown structure.
An advantage of the preferred segmentation method described in Figure. 2 is that small lesions are not lost because of its size, which is often a challenge for other non-preferred segmentation method, such as the well-known classic k-means algorithm. Without proper region segmentation, can be difficult to perform object recognition and proper assessment of such properties as size and shape. The result of segmentation contains labels of regions, i.e., each pixel is represented by its label value. The image of labeled regions is referred to as a label image, or segmentation map.
As indicated above, it may be desirable to pre-process 15 the input image prior to region segmentation 20. More particularly, it may be desirable to enhance the contrast between lesion and normal tissues before segmentation 20 is applied.
In a preferred embodiment of the present invention, a gamma adjustment of 2.0 is used. Alternative contrast enhancement can be performed for the same purpose. Let I(x,y) denote an intensity value if a pixel coordinates (x,y), gamma adjusted value is given by the equation: I'(x,y) = I(x,y)Λ(l/gamma).
Following region segmentation 20, the segmented regions may be selected as candidate regions for lesions in step 30.
In general, lesions typically appear as relatively darker regions in an ultrasound image because they tend to form boundaries against the normal tissue and have higher water content. Therefore, darker regions are selected as lesion candidates in the present invention.
In a preferred embodiment of the present invention, regions with an average intensity lower than a pre-determined value, (for example, 20% of the possible maximum intensity) are selected as candidate regions.
Additional criteria can be used to further eliminate/reduce unlikely candidate regions. For example, it is assumed that an operator would usually place the lesion in the center of the view. In other words, a lesion, if present, would be completely contained in the image, therefore regions that touch image borders are unlikely to be the lesion and thus can be eliminated from further analysis.
Candidate selection facilitates the subsequent lesion classification by focusing on a small number of regions.
For each of the selected candidate regions, a set of features are extracted to characterize each region in step 40. The list of features can comprise one or more categories. In a particular application, three categories are employed: shape features, interior features, and boundary features. Other categories can be included, for example, shadow features. Other categories may be known to those skilled in the art.
In addition, raw measurements of features maybe converted such that their values are normalized to be within [0, 1.0], by functions with appropriate nonlinearity characteristics. Normalization is well known for persons skilled in the art. Those who are skilled in the art can recognize alternative types of features, without departing from the scope and spirit of the present invention.
In general, lesions that have compact, elliptical shape and smooth contour tend to be benign.
The basic features are height, width, and perimeter, and size. , An aspect ratio of a region is the major axis/minor axis, or height over width, of the region. This feature detector is used to discount long narrow shapes from being considered a lesion. In addition, malignant lesions are harder to compress and may appear taller than wider along the direction of ultrasound scattering, i.e., the direction of the transducer. Benign lesions or non-lesions tend to appear as wider than taller. Such characteristics can be indicated by the aspect ratio as well.
The compactness is the square of the perimeter of the region divided by the area of region. A sigmoid belief function maps the circularity to a belief value.
The interior features include mean and standard deviation of the intensity values within the interior of a candidate region.
The boundary features include mean and standard deviation of the intensity gradient values along the boundary of a candidate region.
The shadow feature is used to measure the shadow cast by a candidate lesion region.
Once the particular features are computed for all the candidate regions, the features are used to classify whether a candidate region is a benign lesion, or a malignant lesion, or not a lesion at all, in step 50. The third class of "not a lesion" is provided should the candidate selection step 30 occasionally include non-lesion regions to be selected because of an ambiguity. Examples of such cases, as classified by medical professionals, need to be collected as so-called training examples. The features and the associated classification labels of these training examples are the basis for deriving a classifier, which is then used to classify a novel candidate region not seen by the classifier. Those skilled in the art can choose from many existing classifiers, including a decision tree, a neural network, a Bayesian classifier, or the like. Duda, Hart, and Stork, "Pattern Classifier", describes examples of such classifiers. In a preferred embodiment of the present invention, a decision tree is used. If a confidence value is desired, alternative classifiers such as a neural network or a Bayesian classifier should be used.
Figures 3A-3D shows four example ultrasound images of the breast, hi Figures 4A-4D, corresponding to Figures 3 A-3D, each image is segmented into three intensity classes, with the intensity class of the darkest regions indicated by the highest brightness in the segmented images. In general, lesion candidates are selected from these regions, hi Figures 5A-5D, all the candidate regions are classified into benign (e.g., part 100), malignant (e.g., part 110), or not a lesion (e.g., part 120). More particularly, Figure 5 A shows no lesion, Figure 5B shows a benign lesion, Figure 5 C shows a malignant lesion, and Figure 5D shows a malignant lesion.
Alternatively, if a confidence value is desirable, a colored bar 200 (shown proximate Figures 5A-5D) can be displayed next to a region upon which the operator places the cursor, wherein the shades (color saturation) of the bar gives a visual presentation of the lesions classification and the associated confidence value. For example, a tick 201 (or other symbol) could point to the range of deep red of the color bar when the lesion is classified as malignant with a high confidence value, or the range of light blue of the color bar when lesion is classified as benign but with a low confidence value, or the like.
Figures 6A and 6B provide additional examples, showing a cursor operable by an operator and wherein a candidate region is outlined. Although the present invention is described in a preferred embodiment that is related to lesions in the breast, it will be understood that variations and modifications can be effected within the spirit and scope of the invention for lesions in other tissues, such as the liver.
The present invention may be practiced online, real-time, or offline. In the real-time situation, it could provide the operator with instant warning or feedback, hi the offline situation, it could analyze a database of pre-recorded images. hi addition, it may be desirable to display the ultrasound image and an outline of the segmented lesion region following automatic segmentation, and accept adjustment of the region outline by an operator using a graphical user interface.
The present invention may be implemented for example in a computer program product. A computer program product may include one or more storage media, for example; magnetic storage media such as magnetic disk (such as a floppy disk) or magnetic tape; optical storage media such as optical disk, optical tape, or machine readable bar code; solid-state electronic storage devices such as random access memory (RAM), or read-only memory (ROM); or any other physical device or media employed to store a computer program having instructions for controlling one or more computers to practice the method according to the present invention.
The system of the invention includes a programmable computer having a microprocessor, computer memory, and a computer program stored in said computer memory for performing the steps of the method. The computer has a memory interface operatively connected to the microprocessor. This can be a port, such as a USB port, over a drive that accepts removable memory, or some other device that allows access to camera memory. The system includes a digital camera that has memory that is compatible with the memory interface. A photographic film camera and scanner can be used in place of the digital camera, if desired. A graphical user interface (GUI) and user input unit, such as a mouse and keyboard can be provided as part of the computer.

Claims

CLAIMS:
1. A method for detecting a lesion in a digital ultrasound image of anatomical tissue, the method comprising the steps of: accessing the digital ultrasound image of anatomical tissue; segmenting spatially contiguous pixels in the digital image into a plurality of regions in accordance with substantially similar intensity patterns; selecting, from the plurality of regions, one or more candidate lesion regions having an intensity value lower than a pre-determined intensity value; and classifying the one or more candidate lesion regions into at least one of the following classes: benign, malignant, or unknown.
2. The method of claim 1 , wherein the step of classifying the one or more candidate lesion regions includes the steps of: classifying the one or more candidate lesion regions into the benign or malignant class in accordance with at least one of the following features: a shape feature, an interior feature, or a boundary feature.
3. The method of claim 1 , further comprising the step of providing a confidence value for the lesion classification.
4. The method of claim 1, further comprising the step of, prior to the step of segmenting, pre-processing the digital ultrasound image.
5. The method of claim 4, wherein the digital ultrasound image comprises a lesion, and the digital ultrasound image is pre-processed to enhance the contrast between the lesion and the normal anatomical tissue.
6. The method of claim 1 , wherein the selected candidate lesion regions have an average intensity value lower than a pre-determined intensity value.
7. A method of determining a candidate lesion region within an ultrasound image, the method comprising the steps of: accessing a digital ultrasound image of anatomical tissue; segmenting spatially contiguous pixels in the ultrasound image into a plurality of regions in accordance with substantially similar intensity values and spatial smoothness constraints; and selecting, from the plurality of regions, one or more candidate lesion regions having an intensity value lower than a pre-determined intensity value.
8. The method of claim 7, further comprising the step of, prior to segmenting, pre-processing the ultrasound image such that a contrast between a lesion and normal anatomical tissue would be enhanced.
9. The method of claim 8, where in the step of pre-processing comprises a gamma adjustment.
10. The method of claim 7, further comprising the step of classifying the one or more candidate lesion region into at least one of the following classes: benign, malignant, or unknown.
11. The method of claim 10, wherein the step of classifying the one or more candidate lesion regions includes the steps of: classifying the one or more candidate lesion region into the benign or malignant class in accordance with at least one of the following features: a shape feature, an interior feature, or a boundary feature.
12. The method of claim 7, further comprising the step of, prior to selecting one or more candidate lesion regions, reducing the number of the plurality of regions by removing small holes within interior of a region and smoothing an outline of the region.
13. The method of claim 7, wherein the step of segmenting is accomplished using a Markov random field model in a modified k-means algorithm.
14. The method of claim 10, wherein the step of classifying step further comprises the steps of: extracting an outline of the selected candidate lesion region; determining at least one feature from the group: a shape feature, an interior feature, or a boundary feature; and classifying the selected candidate lesion region as either benign, malignant, or unknown or malignant in response to the computed at least one feature.
15. The method of claim 14, wherein at least two features are determined for the selected candidate lesion region.
16. The method of claim 14, wherein the shape feature is an aspect ratio of the selected candidate lesion region.
17. The method of claim 14, wherein the boundary feature includes a ratio of region perimeter and square root of the selected candidate lesion region.
18. The method of claim 14, wherein the interior feature is a variance of pixel values within the selected candidate lesion region.
19. The method of claim 14, wherein the boundary feature is an average gradient across boundary of the selected candidate lesion region.
20. The method of claim 7, further comprising the step of: displaying the ultrasound image on a display with an outline of the selected candidate lesion region.
21. A computer storage product having at least one computer storage medium having instructions stored therein causing one or more computers to perform the method of claim 1.
22. A system for detecting a lesion in a digital ultrasound image of anatomical tissue, the method comprising the steps of: means for accessing the digital ultrasound image of anatomical tissue; means for segmenting spatially contiguous pixels in the digital image into a plurality of regions in accordance with substantially similar intensity patterns; means for selecting, from the plurality of regions, one or more candidate lesion regions having an intensity value lower than a pre-determined intensity value; and means for classifying the one or more candidate lesion regions into at least one of the following classes: benign, malignant, or unknown.
PCT/US2005/037843 2004-11-22 2005-10-20 Detecting and classifying lesions in ultrasound images WO2006057740A2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP2007543061A JP2008520345A (en) 2004-11-22 2005-10-20 Method and system for detecting and classifying lesions in ultrasound images
EP05811996A EP1815429A2 (en) 2004-11-22 2005-10-20 Detecting and classifying lesions in ultrasound images

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US10/994,794 US7736313B2 (en) 2004-11-22 2004-11-22 Detecting and classifying lesions in ultrasound images
US10/994,794 2004-11-22

Publications (2)

Publication Number Publication Date
WO2006057740A2 true WO2006057740A2 (en) 2006-06-01
WO2006057740A3 WO2006057740A3 (en) 2006-08-10

Family

ID=36216983

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2005/037843 WO2006057740A2 (en) 2004-11-22 2005-10-20 Detecting and classifying lesions in ultrasound images

Country Status (5)

Country Link
US (1) US7736313B2 (en)
EP (1) EP1815429A2 (en)
JP (1) JP2008520345A (en)
CN (1) CN101061513A (en)
WO (1) WO2006057740A2 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009153715A (en) * 2007-12-26 2009-07-16 Toshiba Corp Ultrasonic diagnostic system, ultrasonic image processor and ultrasonic image processing program
EP2378978A1 (en) * 2008-12-22 2011-10-26 The Medipattern Corporation Method and system of automated detection of lesions in medical images
EP2754395A1 (en) * 2013-01-10 2014-07-16 Samsung Electronics Co., Ltd Lesion diagnosis apparatus and method
WO2016015994A1 (en) * 2014-07-29 2016-02-04 Koninklijke Philips N.V. Ultrasound imaging apparatus
US11096667B2 (en) 2016-11-17 2021-08-24 Samsung Medison Co., Ltd. Ultrasound imaging apparatus and method of controlling the same

Families Citing this family (66)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7778450B2 (en) * 2005-01-20 2010-08-17 Scimed Life Systems, Inc. Pattern recognition systems and methods
US8929621B2 (en) * 2005-12-20 2015-01-06 Elekta, Ltd. Methods and systems for segmentation and surface matching
US9451928B2 (en) * 2006-09-13 2016-09-27 Elekta Ltd. Incorporating internal anatomy in clinical radiotherapy setups
US7804989B2 (en) * 2006-10-30 2010-09-28 Eigen, Inc. Object recognition system for medical imaging
CA2675619C (en) * 2007-01-19 2016-08-16 Sunnybrook Health Sciences Centre Scanning mechanisms for imaging probe
US7929762B2 (en) * 2007-03-12 2011-04-19 Jeffrey Kimball Tidd Determining edgeless areas in a digital image
US8249317B2 (en) * 2007-07-20 2012-08-21 Eleckta Ltd. Methods and systems for compensating for changes in anatomy of radiotherapy patients
WO2009012576A1 (en) * 2007-07-20 2009-01-29 Resonant Medical Inc. Methods and systems for guiding the acquisition of ultrasound images
US8135198B2 (en) * 2007-08-08 2012-03-13 Resonant Medical, Inc. Systems and methods for constructing images
US9549713B2 (en) 2008-04-24 2017-01-24 Boston Scientific Scimed, Inc. Methods, systems, and devices for tissue characterization and quantification using intravascular ultrasound signals
WO2009132188A1 (en) * 2008-04-24 2009-10-29 Boston Scientific Scimed, Inc. Methods, systems, and devices for tissue characterization by spectral similarity of intravascular ultrasound signals
US8189738B2 (en) * 2008-06-02 2012-05-29 Elekta Ltd. Methods and systems for guiding clinical radiotherapy setups
KR101100498B1 (en) * 2008-08-05 2011-12-29 삼성메디슨 주식회사 Ultrasound system and method for forming color map
US8331641B2 (en) * 2008-11-03 2012-12-11 Siemens Medical Solutions Usa, Inc. System and method for automatically classifying regions-of-interest
JP5117353B2 (en) * 2008-11-07 2013-01-16 オリンパス株式会社 Image processing apparatus, image processing program, and image processing method
US9144397B2 (en) * 2008-12-04 2015-09-29 Real Imaging Ltd. Method apparatus and system for determining a thermal signature
US10542962B2 (en) * 2009-07-10 2020-01-28 Elekta, LTD Adaptive radiotherapy treatment using ultrasound
US9248316B2 (en) 2010-01-12 2016-02-02 Elekta Ltd. Feature tracking using ultrasound
US20110172526A1 (en) 2010-01-12 2011-07-14 Martin Lachaine Feature Tracking Using Ultrasound
WO2012082994A2 (en) * 2010-12-15 2012-06-21 H. Lee Moffitt Cancer Center And Research Institute, Inc. Method for assessing breast density
CN102855483B (en) * 2011-06-30 2017-09-12 北京三星通信技术研究有限公司 Handle the method and apparatus and breast cancer diagnosis apparatus of ultrasonoscopy
JP6006307B2 (en) 2011-07-07 2016-10-12 ザ ボード オブ トラスティーズ オブ ザ レランド スタンフォード ジュニア ユニバーシティー Comprehensive cardiovascular analysis by volumetric phase contrast MRI
KR20130012297A (en) * 2011-07-25 2013-02-04 삼성전자주식회사 Apparatus for detecting lesion, method for detecting lesion and lesion diagnosis apparatus
US10667790B2 (en) 2012-03-26 2020-06-02 Teratech Corporation Tablet ultrasound system
US9877699B2 (en) 2012-03-26 2018-01-30 Teratech Corporation Tablet ultrasound system
KR102043133B1 (en) 2012-11-16 2019-11-12 삼성전자주식회사 Computer-aided diagnosis supporting apparatus and method
US9378551B2 (en) 2013-01-03 2016-06-28 Siemens Aktiengesellschaft Method and system for lesion candidate detection
JP2016508769A (en) * 2013-01-28 2016-03-24 コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. Medical image processing
KR102042202B1 (en) * 2013-02-25 2019-11-08 삼성전자주식회사 Lesion segmentation apparatus and method in medical image
EP2790152B1 (en) * 2013-04-12 2015-12-02 Alcatel Lucent Method and device for automatic detection and tracking of one or multiple objects of interest in a video
US8995739B2 (en) * 2013-08-21 2015-03-31 Seiko Epson Corporation Ultrasound image object boundary localization by intensity histogram classification using relationships among boundaries
US9014452B2 (en) * 2013-08-21 2015-04-21 Seiko Epson Corporation Orientation-aware average intensity histogram to indicate object boundary depth in ultrasound images
US10117597B2 (en) 2014-01-17 2018-11-06 Arterys Inc. Apparatus, methods and articles for four dimensional (4D) flow magnetic resonance imaging using coherency identification for magnetic resonance imaging flow data
US10331852B2 (en) 2014-01-17 2019-06-25 Arterys Inc. Medical imaging and efficient sharing of medical imaging information
KR20150108701A (en) * 2014-03-18 2015-09-30 삼성전자주식회사 System and method for visualizing anatomic elements in a medical image
KR102372351B1 (en) * 2014-08-19 2022-03-08 삼성메디슨 주식회사 Ultrasound diagnosis apparatus and ultrasound diagnosis method
KR101656582B1 (en) * 2014-11-25 2016-09-12 재단법인 아산사회복지재단 Method and system for detection of thyroid nodules
JP6475363B2 (en) 2015-05-05 2019-02-27 ボストン サイエンティフィック サイムド,インコーポレイテッドBoston Scientific Scimed,Inc. System and method comprising a swellable material disposed on a transducer of an ultrasound imaging system
EP3298581A1 (en) * 2015-05-18 2018-03-28 Koninklijke Philips N.V. Self-aware image segmentation methods and systems
CN108603922A (en) 2015-11-29 2018-09-28 阿特瑞斯公司 Automatic cardiac volume is divided
US20170164924A1 (en) * 2015-12-15 2017-06-15 Konica Minolta, Inc. Ultrasound image diagnostic apparatus
TWI574671B (en) * 2016-06-27 2017-03-21 太豪生醫股份有限公司 Analysis method for breast image and electronic apparatus thereof
US11484247B2 (en) 2016-07-01 2022-11-01 Bostel Technologies, Llc Phonodermoscopy, a medical device system and method for skin diagnosis
US11298072B2 (en) 2016-07-01 2022-04-12 Bostel Technologies, Llc Dermoscopy diagnosis of cancerous lesions utilizing dual deep learning algorithms via visual and audio (sonification) outputs
AU2017290785B2 (en) * 2016-07-01 2021-11-25 The Board Of Regents Of The University Of Texas System Methods, apparatuses, and systems for creating 3-dimensional representations exhibiting geometric and surface characteristics of brain lesions
CN106326856A (en) * 2016-08-18 2017-01-11 厚凯(天津)医疗科技有限公司 Surgery image processing method and surgery image processing device
CN109923617B (en) * 2016-10-07 2023-09-22 皇家飞利浦有限公司 Background sensitive magnifier
GB201617444D0 (en) * 2016-10-14 2016-11-30 Fujitsu Ltd Processing digital image to remove unwanted portion
US10902598B2 (en) 2017-01-27 2021-01-26 Arterys Inc. Automated segmentation utilizing fully convolutional networks
JP6955303B2 (en) * 2017-04-12 2021-10-27 富士フイルム株式会社 Medical image processing equipment and methods and programs
US20200085382A1 (en) * 2017-05-30 2020-03-19 Arterys Inc. Automated lesion detection, segmentation, and longitudinal identification
EP3673814B1 (en) * 2017-08-23 2022-11-16 FUJIFILM Corporation Acoustic wave diagnostic apparatus and method for controlling acoustic wave diagnostic apparatus
EP3714467A4 (en) 2017-11-22 2021-09-15 Arterys Inc. Content based image retrieval for lesion analysis
AU2019265531A1 (en) * 2018-05-07 2020-10-08 Hologic, Inc. Breast ultrasound workflow application
CN112367896A (en) * 2018-07-09 2021-02-12 富士胶片株式会社 Medical image processing apparatus, medical image processing system, medical image processing method, and program
WO2020026349A1 (en) * 2018-07-31 2020-02-06 オリンパス株式会社 Diagnostic imaging assistance system and diagnostic imaging assistance device
CN109308488B (en) * 2018-08-30 2022-05-03 深圳大学 Mammary gland ultrasonic image processing device, method, computer equipment and storage medium
CN110490850B (en) * 2019-02-14 2021-01-08 腾讯科技(深圳)有限公司 Lump region detection method and device and medical image processing equipment
CN110033456B (en) * 2019-03-07 2021-07-09 腾讯科技(深圳)有限公司 Medical image processing method, device, equipment and system
KR20200109093A (en) * 2019-03-12 2020-09-22 삼성메디슨 주식회사 Apparatus and method for displaying ultrasound image and computer program product
US11308611B2 (en) 2019-10-09 2022-04-19 Siemens Healthcare Gmbh Reducing false positive detections of malignant lesions using multi-parametric magnetic resonance imaging
US11250563B2 (en) * 2019-10-31 2022-02-15 Tencent America LLC Hierarchical processing technique for lesion detection, classification, and segmentation on microscopy images
US11817204B2 (en) * 2019-12-09 2023-11-14 Case Western Reserve University Specialized computer-aided diagnosis and disease characterization with a multi-focal ensemble of convolutional neural networks
CN110992351B (en) * 2019-12-12 2022-08-16 南京邮电大学 sMRI image classification method and device based on multi-input convolution neural network
CN111862045B (en) * 2020-07-21 2021-09-07 上海杏脉信息科技有限公司 Method and device for generating blood vessel model
KR102406080B1 (en) * 2020-09-01 2022-06-07 성균관대학교산학협력단 Apparatus and method for lesion analysis based on marginal feature

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5984870A (en) * 1997-07-25 1999-11-16 Arch Development Corporation Method and system for the automated analysis of lesions in ultrasound images
US6317617B1 (en) * 1997-07-25 2001-11-13 Arch Development Corporation Method, computer program product, and system for the automated analysis of lesions in magnetic resonance, mammogram and ultrasound images
EP1398721A2 (en) * 2002-09-13 2004-03-17 GE Medical Systems Global Technology Company LLC Computer assisted analysis of tomographic mammography data

Family Cites Families (41)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US125621A (en) * 1872-04-09 Improvement in furnaces for smelting ores
US161513A (en) * 1873-11-13 1875-03-30 Improvement in watches
US5016173A (en) * 1989-04-13 1991-05-14 Vanguard Imaging Ltd. Apparatus and method for monitoring visually accessible surfaces of the body
CA2177472A1 (en) * 1993-11-29 1995-06-01 Maryellen L. Giger Automated method and system for improved computerized detection and classification of masses in mammograms
US5671294A (en) * 1994-09-15 1997-09-23 The United States Of America As Represented By The Secretary Of The Navy System and method for incorporating segmentation boundaries into the calculation of fractal dimension features for texture discrimination
AU705713B2 (en) * 1995-03-03 1999-05-27 Arch Development Corporation Method and system for the detection of lesions in medical images
US5825909A (en) * 1996-02-29 1998-10-20 Eastman Kodak Company Automated method and system for image segmentation in digital radiographic images
WO1998037811A1 (en) * 1997-02-28 1998-09-03 Electro-Optical Sciences, Inc. Systems and methods for the multispectral imaging and characterization of skin tissue
AU752829B2 (en) * 1998-01-26 2002-10-03 Brigham And Women's Hospital Fluorescence imaging endoscope
US6282305B1 (en) * 1998-06-05 2001-08-28 Arch Development Corporation Method and system for the computerized assessment of breast cancer risk
US6138045A (en) * 1998-08-07 2000-10-24 Arch Development Corporation Method and system for the segmentation and classification of lesions
US6901156B2 (en) * 2000-02-04 2005-05-31 Arch Development Corporation Method, system and computer readable medium for an intelligent search workstation for computer assisted interpretation of medical images
US7212668B1 (en) * 2000-08-18 2007-05-01 Eastman Kodak Company Digital image processing system and method for emphasizing a main subject of an image
US6697502B2 (en) * 2000-12-14 2004-02-24 Eastman Kodak Company Image processing method for detecting human figures in a digital image
US6956975B2 (en) * 2001-04-02 2005-10-18 Eastman Kodak Company Method for improving breast cancer diagnosis using mountain-view and contrast-enhancement presentation of mammography
US7062085B2 (en) * 2001-09-13 2006-06-13 Eastman Kodak Company Method for detecting subject matter regions in images
US6855114B2 (en) * 2001-11-23 2005-02-15 Karen Drukker Automated method and system for the detection of abnormalities in sonographic images
US7123762B2 (en) * 2002-02-08 2006-10-17 University Of Chicago Method and system for risk-modulated diagnosis of disease
US20030161513A1 (en) 2002-02-22 2003-08-28 The University Of Chicago Computerized schemes for detecting and/or diagnosing lesions on ultrasound images using analysis of lesion shadows
US7035461B2 (en) * 2002-08-22 2006-04-25 Eastman Kodak Company Method for detecting objects in digital images
US6574304B1 (en) * 2002-09-13 2003-06-03 Ge Medical Systems Global Technology Company, Llc Computer aided acquisition of medical images
US7346209B2 (en) * 2002-09-30 2008-03-18 The Board Of Trustees Of The Leland Stanford Junior University Three-dimensional pattern recognition method to detect shapes in medical images
US7139437B2 (en) * 2002-11-12 2006-11-21 Eastman Kodak Company Method and system for removing artifacts in compressed images
US7221787B2 (en) * 2002-12-10 2007-05-22 Eastman Kodak Company Method for automated analysis of digital chest radiographs
US7466848B2 (en) * 2002-12-13 2008-12-16 Rutgers, The State University Of New Jersey Method and apparatus for automatically detecting breast lesions and tumors in images
US7333644B2 (en) * 2003-03-11 2008-02-19 Siemens Medical Solutions Usa, Inc. Systems and methods for providing automatic 3D lesion segmentation and measurements
US7529394B2 (en) * 2003-06-27 2009-05-05 Siemens Medical Solutions Usa, Inc. CAD (computer-aided decision) support for medical imaging using machine learning to adapt CAD process with knowledge collected during routine use of CAD system
US7171058B2 (en) * 2003-07-31 2007-01-30 Eastman Kodak Company Method and computer program product for producing an image of a desired aspect ratio
US7555165B2 (en) * 2003-11-13 2009-06-30 Eastman Kodak Company Method for semantic scene classification using camera metadata and content-based cues
US7463757B2 (en) * 2003-12-09 2008-12-09 Carestream Health, Inc. Tooth locating within dental images
US6853318B1 (en) * 2003-12-30 2005-02-08 Eastman Kodak Company Digital image compression utilizing shrinkage of subband coefficients
US7983732B2 (en) * 2004-02-13 2011-07-19 The University Of Chicago Method, system, and computer software product for automated identification of temporal patterns with high initial enhancement in dynamic magnetic resonance breast imaging
US7430321B2 (en) * 2004-09-09 2008-09-30 Siemens Medical Solutions Usa, Inc. System and method for volumetric tumor segmentation using joint space-intensity likelihood ratio test
US7519207B2 (en) * 2004-11-19 2009-04-14 Carestream Health, Inc. Detection and correction method for radiograph orientation
US7317821B2 (en) * 2004-11-22 2008-01-08 Carestream Health, Inc. Automatic abnormal tissue detection in MRI images
US7382907B2 (en) * 2004-11-22 2008-06-03 Carestream Health, Inc. Segmenting occluded anatomical structures in medical images
US20060110035A1 (en) * 2004-11-23 2006-05-25 Hui Luo Method for classifying radiographs
US7599542B2 (en) * 2005-04-08 2009-10-06 John Philip Brockway System and method for detection and display of diseases and abnormalities using confidence imaging
US8517945B2 (en) * 2005-04-28 2013-08-27 Carestream Health, Inc. Segmentation of lesions in ultrasound images
US7756317B2 (en) * 2005-04-28 2010-07-13 Carestream Health, Inc. Methods and systems for automated detection and analysis of lesion on magnetic resonance images
US7672497B2 (en) * 2005-12-29 2010-03-02 Carestream Health, Inc. Computer aided disease detection system for multiple organ systems

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5984870A (en) * 1997-07-25 1999-11-16 Arch Development Corporation Method and system for the automated analysis of lesions in ultrasound images
US6317617B1 (en) * 1997-07-25 2001-11-13 Arch Development Corporation Method, computer program product, and system for the automated analysis of lesions in magnetic resonance, mammogram and ultrasound images
EP1398721A2 (en) * 2002-09-13 2004-03-17 GE Medical Systems Global Technology Company LLC Computer assisted analysis of tomographic mammography data

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
ARMATO S G ET AL: "COMPUTERIZED DETECTION OF LUNG NODULES IN COMPUTED TOMOGRAPHY SCANS" COMPUTER-AIDED DIAGNOSIS IN MEDICAL IMAGING, ELSEVIER, AMSTERDAM,, NL, 20 September 1998 (1998-09-20), pages 119-123, XP009016979 *
ZHIMIN HUO ET AL: "Computer-aided diagnosis: analysis of mammographic parenchymal patterns and classification of masses on digitized mammograms" ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY, 1998. PROCEEDINGS OF THE 20TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE HONG KONG, CHINA 29 OCT.-1 NOV. 1998, PISCATAWAY, NJ, USA,IEEE, US, vol. 2, 29 October 1998 (1998-10-29), pages 1017-1020, XP010320579 ISBN: 0-7803-5164-9 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009153715A (en) * 2007-12-26 2009-07-16 Toshiba Corp Ultrasonic diagnostic system, ultrasonic image processor and ultrasonic image processing program
EP2378978A1 (en) * 2008-12-22 2011-10-26 The Medipattern Corporation Method and system of automated detection of lesions in medical images
EP2378978A4 (en) * 2008-12-22 2013-05-22 Medipattern Corp Method and system of automated detection of lesions in medical images
EP2712554A1 (en) * 2008-12-22 2014-04-02 Salient Imaging, Inc. Method and system of automated detection of lesions in medical images
EP2754395A1 (en) * 2013-01-10 2014-07-16 Samsung Electronics Co., Ltd Lesion diagnosis apparatus and method
WO2016015994A1 (en) * 2014-07-29 2016-02-04 Koninklijke Philips N.V. Ultrasound imaging apparatus
US11096667B2 (en) 2016-11-17 2021-08-24 Samsung Medison Co., Ltd. Ultrasound imaging apparatus and method of controlling the same

Also Published As

Publication number Publication date
CN101061513A (en) 2007-10-24
WO2006057740A3 (en) 2006-08-10
JP2008520345A (en) 2008-06-19
EP1815429A2 (en) 2007-08-08
US7736313B2 (en) 2010-06-15
US20060120608A1 (en) 2006-06-08

Similar Documents

Publication Publication Date Title
US7736313B2 (en) Detecting and classifying lesions in ultrasound images
US10470734B2 (en) Characterizing lung nodule risk with quantitative nodule and perinodular radiomics
Guo et al. Radiomics analysis on ultrasound for prediction of biologic behavior in breast invasive ductal carcinoma
Shi et al. Characterization of mammographic masses based on level set segmentation with new image features and patient information
US7418123B2 (en) Automated method and system for computerized image analysis for prognosis
US6901156B2 (en) Method, system and computer readable medium for an intelligent search workstation for computer assisted interpretation of medical images
US7760941B2 (en) Method and apparatus of segmenting an object in a data set and of determination of the volume of segmented object
Norton et al. Three‐phase general border detection method for dermoscopy images using non‐uniform illumination correction
SG194407A1 (en) A method and system for determining a stage of fibrosis in a liver
JP2010504129A (en) Advanced computer-aided diagnosis of pulmonary nodules
JP2004532067A (en) An automated and computerized mechanism for discriminating between benign and malignant solitary pulmonary nodules on chest images
US20100183210A1 (en) Computer-assisted analysis of colonic polyps by morphology in medical images
Cavalcanti et al. Macroscopic pigmented skin lesion segmentation and its influence on lesion classification and diagnosis
US20200058125A1 (en) Comparative cancer survival models to assist physicians to choose optimal treatment
KR20180022607A (en) Determination of result data on the basis of medical measurement data from various measurements
Riaz et al. Impact of visual features on the segmentation of gastroenterology images using normalized cuts
US9430844B2 (en) Automated mammographic density estimation and display method using prior probability information, system for the same, and media storing computer program for the same
Boubakar Khalifa Albargathe et al. Blood vessel segmentation and extraction using H-minima method based on image processing techniques
TW201726064A (en) Medical image processing apparatus and breast image processing method thereof
CN116230237B (en) Lung cancer influence evaluation method and system based on ROI focus features
CN113924046A (en) Computer-based method for classifying organ masses as cysts
Touil et al. A new conditional region growing approach for microcalcification delineation in mammograms
Lu et al. A Review of the Role of Ultrasound Radiomics and Its Application and Limitations in the Investigation of Thyroid Disease
Zhang et al. Development and external validation of a simple-to-Use dynamic nomogram for predicting breast malignancy based on ultrasound morphometric features: a retrospective multicenter study
WO2022153100A1 (en) A method for detecting breast cancer using artificial neural network

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KP KR KZ LC LK LR LS LT LU LV LY MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU LV MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 2005811996

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2007543061

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 200580039892.5

Country of ref document: CN

NENP Non-entry into the national phase

Ref country code: DE

WWP Wipo information: published in national office

Ref document number: 2005811996

Country of ref document: EP