US20150374316A1 - Image display to display internal structure with change of depth - Google Patents
Image display to display internal structure with change of depth Download PDFInfo
- Publication number
- US20150374316A1 US20150374316A1 US14/766,856 US201414766856A US2015374316A1 US 20150374316 A1 US20150374316 A1 US 20150374316A1 US 201414766856 A US201414766856 A US 201414766856A US 2015374316 A1 US2015374316 A1 US 2015374316A1
- Authority
- US
- United States
- Prior art keywords
- image
- image display
- depth
- user
- display
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000008859 change Effects 0.000 title description 6
- 238000013500 data storage Methods 0.000 claims abstract description 15
- 230000004044 response Effects 0.000 claims abstract description 7
- 238000000034 method Methods 0.000 claims description 17
- 210000002455 dental arch Anatomy 0.000 claims description 8
- 210000000988 bone and bone Anatomy 0.000 claims description 6
- 210000004746 tooth root Anatomy 0.000 description 9
- 238000010586 diagram Methods 0.000 description 7
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000013170 computed tomography imaging Methods 0.000 description 2
- 238000012217 deletion Methods 0.000 description 2
- 230000037430 deletion Effects 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 238000003745 diagnosis Methods 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 238000012549 training Methods 0.000 description 2
- 230000003213 activating effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000009429 electrical wiring Methods 0.000 description 1
- 230000005670 electromagnetic radiation Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 238000003325 tomography Methods 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus for radiation diagnosis, e.g. combined with radiation therapy equipment
- A61B6/14—Applications or adaptations for dentistry
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus for radiation diagnosis, e.g. combined with radiation therapy equipment
- A61B6/02—Devices for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
- A61B6/03—Computerised tomographs
- A61B6/032—Transmission computed tomography [CT]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus for radiation diagnosis, e.g. combined with radiation therapy equipment
- A61B6/46—Apparatus for radiation diagnosis, e.g. combined with radiation therapy equipment with special arrangements for interfacing with the operator or the patient
- A61B6/461—Displaying means of special interest
- A61B6/463—Displaying means of special interest characterised by displaying multiple images or images and diagnostic data on one display
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus for radiation diagnosis, e.g. combined with radiation therapy equipment
- A61B6/46—Apparatus for radiation diagnosis, e.g. combined with radiation therapy equipment with special arrangements for interfacing with the operator or the patient
- A61B6/461—Displaying means of special interest
- A61B6/466—Displaying means of special interest adapted to display 3D data
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus for radiation diagnosis, e.g. combined with radiation therapy equipment
- A61B6/46—Apparatus for radiation diagnosis, e.g. combined with radiation therapy equipment with special arrangements for interfacing with the operator or the patient
- A61B6/467—Apparatus for radiation diagnosis, e.g. combined with radiation therapy equipment with special arrangements for interfacing with the operator or the patient characterised by special input means
-
- A61B6/51—
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus for radiation diagnosis, e.g. combined with radiation therapy equipment
- A61B6/52—Devices using data or image processing specially adapted for radiation diagnosis
- A61B6/5211—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
- A61B6/5223—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data generating planar views from image data, e.g. extracting a coronal view from a 3D image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/20—Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/14—Display of multiple viewports
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10081—Computed x-ray tomography [CT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10116—X-ray image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30036—Dental; Teeth
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2210/00—Indexing scheme for image generation or computer graphics
- G06T2210/41—Medical
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2219/00—Indexing scheme for manipulating 3D models or images for computer graphics
- G06T2219/028—Multiple view windows (top-side-front-sagittal-orthogonal)
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2380/00—Specific applications
- G09G2380/08—Biomedical applications
Definitions
- the present disclosure relates to an image display, and more particularly to an image display to display internal structures within an image with variation of depth of views.
- An X-ray computerized tomography (CT) imaging apparatus emits X-rays to an object and detects the X-rays transmitted through the object using an X-ray detector. Based on the detected X-rays, the X-ray CT imaging apparatus generates image data. The X-ray CT imaging apparatus produces and displays a three-dimensional (3D) image and internal structure of the object based on the generated image data.
- CT computerized tomography
- FIG. 1 to FIG. 5 illustrate dental X-ray images in three dimensions (i.e., 3D image) produced and displayed by a typical image display apparatus.
- the typical image display apparatus may produce and display a volume image of a dental arch of a patient.
- a volume image i.e., 3D dental image
- a user such as a dentist, examines a tooth including a dental crown extending from a gum and a dental root covered with the gum.
- the user manipulates the volume image as follows. The user rotates the volume image so as to see a bottom of the dental arch, as shown in FIG. 2 .
- the user designates a region to be deleted in order to see a dental root of interest, as shown in FIG. 3 .
- the user deletes the designated region, as shown in FIG. 4 .
- the user rotates the volume image to see the dental roots of interest revealed after the deletion, as shown in FIG. 5 .
- the user generally performs at least three steps of: i) rotating the volume image, ii) designating the region to be deleted, and iii) rotating the volume image after deletion of the designated region, as described.
- the user should be accustomed to using the typical display apparatus. Even if the user is proficient in using the typical apparatus, it is not easy for the user to designate the region to be deleted accurately to a depth to be examined. From time to time, the user returns to and restores the entire volume image in order to change the region of interest with another region of interest. In this case, all the three steps should be performed again one after another, and which is time consuming.
- Manipulation of an image, such as a dental image, between several views to focus on a particular area of one of the teeth in interest requires a user to perform many steps before obtaining the desired image. Such excessive steps in manipulation requires time, skill and training on the part of the user, and increases the possibility of error.
- a volume image and a surface at a designated depth thereof may be shown by selecting a depth value based on a user input. From this point, via user input, a depth value can be selected, allowing the user to change the depth of view into the image.
- an image display apparatus may include a data storage, a processor, and an image display.
- the data storage may be configured to store image data.
- the processor may be configured to generate image data for a three dimensional (3D) image including internal surface images at one or more depths.
- the image display may be configured to display the 3D image with internal surfaces at each depth.
- the 3D image may be an X-ray image.
- the X-ray image may be an image of a dental arch.
- the internal surface of the 3D image may include images of an internal structure of teeth including tooth root and an alveolar bone.
- the image display may display a model of a head and a field of view indication thereon.
- an image display apparatus may include a data storage, a user input interface, a processor, and an image display.
- the data storage may be configured to store image data.
- the user input interface may be configured to receive a request from a user.
- the processor may be configured to generate image data for a 3D image with internal surface images at one or more depths in response to the user request.
- the image display may be configured to display the 3D image with internal surfaces at each depth.
- the image display may display a depth indicator including a scroll bar and a plurality of buttons. Based on user input received via the user input interface, a user may control the depth of the 3D image.
- the 3D image may be an X-ray image.
- the X-ray image may be of a dental arch. At least one internal surface of the 3D image may show an internal structure of teeth including tooth root and an alveolar bone.
- the image display may display a model of a head and a field of view indication thereon.
- a method may be provided for display images.
- the method may include displaying a three dimensional (3D) image, receiving a user request, and displaying a 3D image with internal surface images at one or more depths in response to the user request.
- a user does not require substantial skill or training to obtain appropriate images in association with one another for diagnosis and treatment, while also reducing the time required to obtain the desired image.
- FIG. 1 to FIG. 5 are schematic diagrams showing a procedure to examine dental roots with 3D images displayed on a typical X-ray image display device
- FIG. 6 is a schematic diagram showing an image display apparatus in accordance with at least one embodiment
- FIG. 7 is a schematic diagram showing explaining the image data of a CT image
- FIG. 8 is a schematic diagram showing a volume image displayed on an image display
- FIG. 9 to FIG. 5C are schematic diagrams showing surfaces as a change of a depth in accordance with at least one embodiment
- FIG. 12 to FIG. 14 are schematic diagrams showing surfaces as a change of a depth in accordance with at least one embodiment.
- FIG. 15 and FIG. 16 are schematic diagrams showing a model of a head with a field of view.
- an image display apparatus includes: a data storage configured to store image data; a processor configured to generate image data for a three dimensional (3D) image including internal surface images at one or more depths; and an image display configured to display the 3D image with internal surfaces at each depth.
- a data storage configured to store image data
- a processor configured to generate image data for a three dimensional (3D) image including internal surface images at one or more depths
- an image display configured to display the 3D image with internal surfaces at each depth.
- an image display apparatus includes: a data storage configured to store image data; a user input interface configured to receive a request from a user; a processor configured to generate image data for a 3D image with internal surface images at one or more depths in response to the user request; and an image display configured to display the 3D image with internal surfaces at each depth.
- a display device may produce and display a predetermined internal surface (e.g., a predetermined plane view) of an object according to a user input designating a depth thereof.
- the display device may detect a user input to designate a depth of a three dimensional (3D) image. Based on the detected user input, the display device may produce and display an internal surface of the 3D image, which is corresponding to a depth indicated by the detected user input, without requiring further user inputs or user interaction.
- 3D three dimensional
- FIG. 6 illustrates a display device in accordance with at least one exemplary embodiment.
- display device 100 displays a volume image (i.e., a three-dimensional (3D) image).
- Display device 100 may include data storage 110 , image display 120 , a user input interface 130 , and a processor 140 .
- the display device 100 may be configured and implemented via a typical computer, such as a laptop computer.
- Data storage 110 may store 3D X-ray image data.
- the stored 3D X-ray image data may be generated based on image signals obtained from an X-ray computerized tomographic (CT) imaging apparatus.
- CT computerized tomographic
- FIG. 7 illustrates the generation of 3D X-ray image data.
- image signals may be obtained by emitting an X-ray to an object (OB) and detecting the X-ray transmitted through the object (OB) using an X-ray detector.
- the obtained image signals may be processed to generate the 3D X-ray image data.
- the 3D X-ray image data may be stored in data storage 110 .
- the 3D X-ray image data may include information on voxel values, CT numbers (e.g., Hounsfield scale), and information on surfaces in various directions.
- User input interface 130 represents one or more devices that allow a user to interact with the display device 100 such as, but not limited to, touch screen input, mouse input, and keyboard input.
- some embodiments employ a mouse input to select portions of the images on the screen for further processing.
- user input interface 130 is configured to receive various types of inputs, such as a request, from a user.
- the inputs may include inputs for changing a direction (e.g., a display direction) of volume X-ray image 3 D 1 , for changing a depth of an internal surface, and for displaying internal surfaces one by one.
- User input interface 130 may include a mouse, a key board, a touch pad, and a touch panel disposed on a screen of image display 120 .
- User input interface 130 may be combined with image display 120 . That is, the user may move a cursor or an indicator displayed on the screen of image display 120 in order to control the display direction.
- user input interface 130 may receive inputs for scrolling or for activating (e.g., touching or clicking) a menu bar or tool bars displayed on the screen in order to receive the information on the depth of surface to be examined.
- Processor 140 may produce a CT image of the object (OB) based on the 3D X-ray image data stored in data storage 110 and a user input detected by user input interface 130 .
- processor 140 may produce 3D image data based on the user input made through user input interface 130 .
- Image display 120 may receive 3D image data from processor 140 and display a 3D image on the screen.
- Image display 120 may provide first region R 1 for displaying 3D images and second region R 2 for displaying menu bars and toolbars.
- image display 120 may display volume X-ray image 3 D 1 within first region R 1 on the screen, based on the 3D mage data received from processor 140 .
- image display 120 display volume X-ray image 3 D 1 rotatable about an axis (not shown) according to a display direction controlled or appointed by a user.
- Processor 140 may generate 3D image data with the 3D X-ray image data stored in data storage 110 based on user inputs or requests received through user input interface 140 .
- Such user inputs may include information for controlling or manipulating the 3D image displayed on the screen of image display 120 .
- processor 140 may generate 3D image data to display corresponding volume X-ray image 3D rotating according to the rotation direction as shown in FIG. 9 .
- processor 140 may generate the 3D mage data to display a surface (e.g., a sectional plane view) at the designated depth, as shown in FIG. 10 .
- processor 140 may generate the 3D image data for displaying a 3D image with an internal surface at each depth, one by one depth value (e.g., by a predetermined interval), as shown in FIG. 10 and FIG. 11 .
- a user may input the depth information by scrolling menu bar BC displayed in second display region R 2 , as shown in FIG. 9 to FIG. 11 .
- processor 140 may determine a depth value based on the amount of movement of the scroll.
- the length of the menu bar may be proportional to the number of voxels from a contour surface of volume image 3 D 1 to a predetermined surface (e.g., the deepest surface) in a direction to be examined.
- the processor 140 determines an associated depth (e.g., associated voxel value) and generates the 3D image data so as to show the surface (e.g., a contour surface, a sectional view image, or a sagittal plane view) at the determined depth.
- display device 100 may allow a user to conveniently, clearly, and efficiently examine the internal dental structure including a tooth root, a cementum, an alveolar bone, and so forth by changing a depth through scrolling menu bar BC.
- the user may designate the depth by selecting one of depth indication buttons 1 to 9 shown in FIG. 12 to FIG. 14 .
- the depth indication buttons correspond to the respective depths ranging from i) the contour surface of volume image 3 D 1 to ii) the deepest surface of a dental arch thereof.
- the number of the depth indication buttons may depend on a depth difference between adjacent buttons. In other words, the number of the buttons may be determined in consideration of ease of use. For a detailed analysis, the number may be increased to a maximum number of the voxels in the direction to be examined.
- a user may be enabled to designate a region of interest in an object.
- display device 100 may produce and display internal surfaces of the designated region according to a depth controlled or selected based on a user input made through user input interface 140 .
- a user might select a particular tooth for careful examination.
- Display device 100 may produce and display internal surfaces (e.g., sectional image views) of the designated tooth.
- display device 100 might display a field of view (FOV) on a model of a head, as shown in FIG. 15 and FIG. 16 , where the FOV is changed from a relatively large area ( FIG. 15 ) to a relatively small area ( FIG. 16 ).
- FOV field of view
- a volume image and a surface at a designated depth thereof may be shown by selecting a depth value based on a user input. From this point, via user input, a depth value can be selected, allowing the user to change the depth of view into the image.
- the word exemplary is used herein to mean serving as an example, instance, or illustration. Any aspect or design described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects or designs. Rather, use of the word exemplary is intended to present concepts in a concrete fashion.
- the term “or” is intended to mean an inclusive “or” rather than an exclusive “or”. That is, unless specified otherwise, or clear from context, “employs A or B” is intended to mean any of the natural inclusive permutations. That is, if X employs A; X employs B; or X employs both A and B, then “X employs A or B” is satisfied under any of the foregoing instances.
- the articles “a” and “an” as used in this application and the appended claims should generally be construed to mean “one or more” unless specified otherwise or clear from context to be directed to a singular form.
- a component may be, but is not limited to being, a process running on a processor, a processor, an object, an executable, a thread of execution, a program, and/or a computer.
- an application running on a controller and the controller can be a component.
- One or more components may reside within a process and/or thread of execution and a component may be localized on one computer and/or distributed between two or more computers.
- the present invention can be embodied in the form of methods and apparatuses for practicing those methods.
- the present invention can also be embodied in the form of program code embodied in tangible media, non-transitory media, such as magnetic recording media, optical recording media, solid state memory, floppy diskettes, CD-ROMs, hard drives, or any other machine-readable storage medium, wherein, when the program code is loaded into and executed by a machine, such as a computer, the machine becomes an apparatus for practicing the invention.
- the present invention can also be embodied in the form of program code, for example, whether stored in a storage medium, loaded into and/or executed by a machine, or transmitted over some transmission medium or carrier, such as over electrical wiring or cabling, through fiber optics, or via electromagnetic radiation, wherein, when the program code is loaded into and executed by a machine, such as a computer, the machine becomes an apparatus for practicing the invention.
- program code When implemented on a general-purpose processor, the program code segments combine with the processor to provide a unique device that operates analogously to specific logic circuits.
- the present invention can also be embodied in the form of a bitstream or other sequence of signal values electrically or optically transmitted through a medium, stored magnetic-field variations in a magnetic recording medium, etc., generated using a method and/or an apparatus of the present invention.
- the term “compatible” means that the element communicates with other elements in a manner wholly or partially specified by the standard, and would be recognized by other elements as sufficiently capable of communicating with the other elements in the manner specified by the standard.
- the compatible element does not need to operate internally in a manner specified by the standard.
- the present invention is applicable to display internal structures within an image with variation of depth of views.
Abstract
The disclosure is related to a display device that may produce and display a predetermined internal surface (e.g., a predetermined plane view) of an object according to a user input designating a depth thereof. Such display device may include a data storage, a user input interface, a processor, and an image display. The data storage may be configured to store image data. The user input interface may be configured to receive a request from a user. The processor may be configured to generate image data for a 3D image with internal surface images at one or more depths in response to the user request. The image display may be configured to display the 3D image with internal surfaces at each depth.
Description
- The present disclosure relates to an image display, and more particularly to an image display to display internal structures within an image with variation of depth of views.
- An X-ray computerized tomography (CT) imaging apparatus emits X-rays to an object and detects the X-rays transmitted through the object using an X-ray detector. Based on the detected X-rays, the X-ray CT imaging apparatus generates image data. The X-ray CT imaging apparatus produces and displays a three-dimensional (3D) image and internal structure of the object based on the generated image data.
-
FIG. 1 toFIG. 5 illustrate dental X-ray images in three dimensions (i.e., 3D image) produced and displayed by a typical image display apparatus. As shown inFIG. 1 , the typical image display apparatus may produce and display a volume image of a dental arch of a patient. Based on such a volume image (i.e., 3D dental image), a user, such as a dentist, examines a tooth including a dental crown extending from a gum and a dental root covered with the gum. In order to examine the dental root with the volume image, the user manipulates the volume image as follows. The user rotates the volume image so as to see a bottom of the dental arch, as shown inFIG. 2 . The user designates a region to be deleted in order to see a dental root of interest, as shown inFIG. 3 . The user deletes the designated region, as shown inFIG. 4 . The user rotates the volume image to see the dental roots of interest revealed after the deletion, as shown inFIG. 5 . - To see the dental roots of interest with the typical display apparatus, the user generally performs at least three steps of: i) rotating the volume image, ii) designating the region to be deleted, and iii) rotating the volume image after deletion of the designated region, as described. In addition, the user should be accustomed to using the typical display apparatus. Even if the user is proficient in using the typical apparatus, it is not easy for the user to designate the region to be deleted accurately to a depth to be examined. From time to time, the user returns to and restores the entire volume image in order to change the region of interest with another region of interest. In this case, all the three steps should be performed again one after another, and which is time consuming.
- Manipulation of an image, such as a dental image, between several views to focus on a particular area of one of the teeth in interest requires a user to perform many steps before obtaining the desired image. Such excessive steps in manipulation requires time, skill and training on the part of the user, and increases the possibility of error.
- A volume image and a surface at a designated depth thereof may be shown by selecting a depth value based on a user input. From this point, via user input, a depth value can be selected, allowing the user to change the depth of view into the image.
- In accordance with at least one embodiment, an image display apparatus may include a data storage, a processor, and an image display. The data storage may be configured to store image data. The processor may be configured to generate image data for a three dimensional (3D) image including internal surface images at one or more depths. The image display may be configured to display the 3D image with internal surfaces at each depth.
- The 3D image may be an X-ray image. The X-ray image may be an image of a dental arch. The internal surface of the 3D image may include images of an internal structure of teeth including tooth root and an alveolar bone. The image display may display a model of a head and a field of view indication thereon.
- In accordance with at least one embodiment, an image display apparatus may include a data storage, a user input interface, a processor, and an image display. The data storage may be configured to store image data. The user input interface may be configured to receive a request from a user. The processor may be configured to generate image data for a 3D image with internal surface images at one or more depths in response to the user request. The image display may be configured to display the 3D image with internal surfaces at each depth.
- The image display may display a depth indicator including a scroll bar and a plurality of buttons. Based on user input received via the user input interface, a user may control the depth of the 3D image. The 3D image may be an X-ray image. The X-ray image may be of a dental arch. At least one internal surface of the 3D image may show an internal structure of teeth including tooth root and an alveolar bone. The image display may display a model of a head and a field of view indication thereon.
- In accordance with at least one embodiment, a method may be provided for display images. The method may include displaying a three dimensional (3D) image, receiving a user request, and displaying a 3D image with internal surface images at one or more depths in response to the user request.
- A user does not require substantial skill or training to obtain appropriate images in association with one another for diagnosis and treatment, while also reducing the time required to obtain the desired image.
- The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate one or more embodiments described herein and, together with the description, explain these embodiments. In the drawings:
-
FIG. 1 toFIG. 5 are schematic diagrams showing a procedure to examine dental roots with 3D images displayed on a typical X-ray image display device; -
FIG. 6 is a schematic diagram showing an image display apparatus in accordance with at least one embodiment; -
FIG. 7 is a schematic diagram showing explaining the image data of a CT image; -
FIG. 8 is a schematic diagram showing a volume image displayed on an image display; -
FIG. 9 toFIG. 5C are schematic diagrams showing surfaces as a change of a depth in accordance with at least one embodiment; -
FIG. 12 toFIG. 14 are schematic diagrams showing surfaces as a change of a depth in accordance with at least one embodiment; and -
FIG. 15 andFIG. 16 are schematic diagrams showing a model of a head with a field of view. - In accordance with an embodiment of the present invention, an image display apparatus includes: a data storage configured to store image data; a processor configured to generate image data for a three dimensional (3D) image including internal surface images at one or more depths; and an image display configured to display the 3D image with internal surfaces at each depth.
- In accordance with another embodiment of the present invention, an image display apparatus includes: a data storage configured to store image data; a user input interface configured to receive a request from a user; a processor configured to generate image data for a 3D image with internal surface images at one or more depths in response to the user request; and an image display configured to display the 3D image with internal surfaces at each depth.
- The following detailed description refers to the accompanying drawings. The same reference numbers in different drawings may identify the same or similar elements. Also, the following detailed description does not limit the invention.
- In accordance with at least one embodiment, a display device may produce and display a predetermined internal surface (e.g., a predetermined plane view) of an object according to a user input designating a depth thereof. Particularly, the display device may detect a user input to designate a depth of a three dimensional (3D) image. Based on the detected user input, the display device may produce and display an internal surface of the 3D image, which is corresponding to a depth indicated by the detected user input, without requiring further user inputs or user interaction. Hereinafter, overall configuration and operation of a display device will be described with reference to
FIG. 6 . For convenience and ease of understanding, the display device will be described as displaying a 3D image and internal surface views of dental structure. However, the present invention is not limited thereto. The display device may be applied to producing and displaying 3D image and internal surface views of any objects. -
FIG. 6 illustrates a display device in accordance with at least one exemplary embodiment. Referring toFIG. 6 ,display device 100 displays a volume image (i.e., a three-dimensional (3D) image).Display device 100 may includedata storage 110,image display 120, auser input interface 130, and aprocessor 140. Thedisplay device 100 may be configured and implemented via a typical computer, such as a laptop computer. -
Data storage 110 may store 3D X-ray image data. The stored 3D X-ray image data may be generated based on image signals obtained from an X-ray computerized tomographic (CT) imaging apparatus. As an example,FIG. 7 illustrates the generation of 3D X-ray image data. As shown inFIG. 7 , image signals may be obtained by emitting an X-ray to an object (OB) and detecting the X-ray transmitted through the object (OB) using an X-ray detector. The obtained image signals may be processed to generate the 3D X-ray image data. After the generation of 3D X-ray image data, the 3D X-ray image data may be stored indata storage 110. The 3D X-ray image data may include information on voxel values, CT numbers (e.g., Hounsfield scale), and information on surfaces in various directions.User input interface 130 represents one or more devices that allow a user to interact with thedisplay device 100 such as, but not limited to, touch screen input, mouse input, and keyboard input. Advantageously, some embodiments employ a mouse input to select portions of the images on the screen for further processing. - In accordance with at least one embodiment,
user input interface 130 is configured to receive various types of inputs, such as a request, from a user. For example, the inputs (e.g., request) may include inputs for changing a direction (e.g., a display direction) of volume X-ray image 3D1, for changing a depth of an internal surface, and for displaying internal surfaces one by one.User input interface 130 may include a mouse, a key board, a touch pad, and a touch panel disposed on a screen ofimage display 120.User input interface 130 may be combined withimage display 120. That is, the user may move a cursor or an indicator displayed on the screen ofimage display 120 in order to control the display direction. Furthermore,user input interface 130 may receive inputs for scrolling or for activating (e.g., touching or clicking) a menu bar or tool bars displayed on the screen in order to receive the information on the depth of surface to be examined. -
Processor 140 may produce a CT image of the object (OB) based on the 3D X-ray image data stored indata storage 110 and a user input detected byuser input interface 130. For example,processor 140 may produce 3D image data based on the user input made throughuser input interface 130. -
Image display 120 may receive 3D image data fromprocessor 140 and display a 3D image on the screen.Image display 120 may provide first region R1 for displaying 3D images and second region R2 for displaying menu bars and toolbars. As shown inFIG. 8 ,image display 120 may display volume X-ray image 3D1 within first region R1 on the screen, based on the 3D mage data received fromprocessor 140. - Furthermore,
image display 120 display volume X-ray image 3D1 rotatable about an axis (not shown) according to a display direction controlled or appointed by a user. -
Processor 140 may generate 3D image data with the 3D X-ray image data stored indata storage 110 based on user inputs or requests received throughuser input interface 140. Such user inputs may include information for controlling or manipulating the 3D image displayed on the screen ofimage display 120. For example, in response to a user input for controlling a direction for rotation (e.g., rotation direction),processor 140 may generate 3D image data to display correspondingvolume X-ray image 3D rotating according to the rotation direction as shown inFIG. 9 . When a user designates a depth through user input interface 140 (e.g., entering information on a depth of interest),processor 140 may generate the 3D mage data to display a surface (e.g., a sectional plane view) at the designated depth, as shown inFIG. 10 . When a user requests displaying an internal surface,processor 140 may generate the 3D image data for displaying a 3D image with an internal surface at each depth, one by one depth value (e.g., by a predetermined interval), as shown inFIG. 10 andFIG. 11 . - For example, a user may input the depth information by scrolling menu bar BC displayed in second display region R2, as shown in
FIG. 9 toFIG. 11 . In this case,processor 140 may determine a depth value based on the amount of movement of the scroll. In accordance with at least one exemplary embodiment, the length of the menu bar may be proportional to the number of voxels from a contour surface of volume image 3D1 to a predetermined surface (e.g., the deepest surface) in a direction to be examined. According to the scroll amount, theprocessor 140 determines an associated depth (e.g., associated voxel value) and generates the 3D image data so as to show the surface (e.g., a contour surface, a sectional view image, or a sagittal plane view) at the determined depth. Accordingly,display device 100 may allow a user to conveniently, clearly, and efficiently examine the internal dental structure including a tooth root, a cementum, an alveolar bone, and so forth by changing a depth through scrolling menu bar BC. In accordance with another embodiment, the user may designate the depth by selecting one ofdepth indication buttons 1 to 9 shown inFIG. 12 toFIG. 14 . The depth indication buttons correspond to the respective depths ranging from i) the contour surface of volume image 3D1 to ii) the deepest surface of a dental arch thereof. The number of the depth indication buttons may depend on a depth difference between adjacent buttons. In other words, the number of the buttons may be determined in consideration of ease of use. For a detailed analysis, the number may be increased to a maximum number of the voxels in the direction to be examined. - In accordance with still another embodiment of this invention, a user may be enabled to designate a region of interest in an object. By designating the region,
display device 100 may produce and display internal surfaces of the designated region according to a depth controlled or selected based on a user input made throughuser input interface 140. For example, a user might select a particular tooth for careful examination.Display device 100 may produce and display internal surfaces (e.g., sectional image views) of the designated tooth. In accordance with yet another embodiment,display device 100 might display a field of view (FOV) on a model of a head, as shown inFIG. 15 andFIG. 16 , where the FOV is changed from a relatively large area (FIG. 15 ) to a relatively small area (FIG. 16 ). Namely, the model of head shown inFIG. 9 toFIG. 11 andFIG. 12 toFIG. 14 may be displayed with the FOV so that the user is shown the position and the size of the FOV. - According to this invention, a volume image and a surface at a designated depth thereof may be shown by selecting a depth value based on a user input. From this point, via user input, a depth value can be selected, allowing the user to change the depth of view into the image.
- Thus, the user does not need to improve his skill to obtain appropriate images in association with one another for diagnosis and time for obtain appropriate can be reduced effectively.
- Reference herein to “one embodiment” or “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment can be included in at least one embodiment of the invention. The appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments necessarily mutually exclusive of other embodiments. The same applies to the term “implentation.”
- As used in this application, the word exemplary is used herein to mean serving as an example, instance, or illustration. Any aspect or design described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects or designs. Rather, use of the word exemplary is intended to present concepts in a concrete fashion.
- Additionally, the term “or” is intended to mean an inclusive “or” rather than an exclusive “or”. That is, unless specified otherwise, or clear from context, “employs A or B” is intended to mean any of the natural inclusive permutations. That is, if X employs A; X employs B; or X employs both A and B, then “X employs A or B” is satisfied under any of the foregoing instances. In addition, the articles “a” and “an” as used in this application and the appended claims should generally be construed to mean “one or more” unless specified otherwise or clear from context to be directed to a singular form.
- Moreover, the terms “system,” “component,” “module,” “interface,” “model” or the like are generally intended to refer to a computer-related entity, either hardware, a combination of hardware and software, software, or software in execution. For example, a component may be, but is not limited to being, a process running on a processor, a processor, an object, an executable, a thread of execution, a program, and/or a computer. By way of illustration, both an application running on a controller and the controller can be a component. One or more components may reside within a process and/or thread of execution and a component may be localized on one computer and/or distributed between two or more computers.
- The present invention can be embodied in the form of methods and apparatuses for practicing those methods. The present invention can also be embodied in the form of program code embodied in tangible media, non-transitory media, such as magnetic recording media, optical recording media, solid state memory, floppy diskettes, CD-ROMs, hard drives, or any other machine-readable storage medium, wherein, when the program code is loaded into and executed by a machine, such as a computer, the machine becomes an apparatus for practicing the invention. The present invention can also be embodied in the form of program code, for example, whether stored in a storage medium, loaded into and/or executed by a machine, or transmitted over some transmission medium or carrier, such as over electrical wiring or cabling, through fiber optics, or via electromagnetic radiation, wherein, when the program code is loaded into and executed by a machine, such as a computer, the machine becomes an apparatus for practicing the invention. When implemented on a general-purpose processor, the program code segments combine with the processor to provide a unique device that operates analogously to specific logic circuits. The present invention can also be embodied in the form of a bitstream or other sequence of signal values electrically or optically transmitted through a medium, stored magnetic-field variations in a magnetic recording medium, etc., generated using a method and/or an apparatus of the present invention.
- It should be understood that the steps of the exemplary methods set forth herein are not necessarily required to be performed in the order described, and the order of the steps of such methods should be understood to be merely exemplary. Likewise, additional steps may be included in such methods, and certain steps may be omitted or combined, in methods consistent with various embodiments of the present invention.
- As used herein in reference to an element and a standard, the term “compatible” means that the element communicates with other elements in a manner wholly or partially specified by the standard, and would be recognized by other elements as sufficiently capable of communicating with the other elements in the manner specified by the standard. The compatible element does not need to operate internally in a manner specified by the standard.
- Although embodiments of the present invention have been described herein, it should be understood that the foregoing embodiments and advantages are merely examples and are not to be construed as limiting the present invention or the scope of the claims. Numerous other modifications and embodiments can be devised by those skilled in the art that will fall within the spirit and scope of the principles of this disclosure, and the present teaching can also be readily applied to other types of apparatuses. More particularly, various variations and modifications are possible in the component parts and/or arrangements of the subject combination arrangement within the scope of the disclosure, the drawings and the appended claims. In addition to variations and modifications in the component parts and/or arrangements, alternative uses will also be apparent to those skilled in the art.
- The present invention is applicable to display internal structures within an image with variation of depth of views.
Claims (18)
1. An image display apparatus comprising:
a data storage configured to store image data;
a processor configured to generate image data for a three dimensional (3D) image including internal surface images at one or more depths; and
an image display configured to display the 3D image with internal surfaces at each depth.
2. An image display apparatus according to claim 1 , wherein the 3D image is an X-ray image.
3. An image display apparatus according to claim 2 , wherein the X-ray image is an image of a dental arch.
4. An image display apparatus according to claim 2 , wherein the internal surface of the 3D image includes images of an internal structure of teeth including tooth root and an alveolar bone.
5. An image display apparatus according to claim 2 , wherein the image display displays a model of a head and a field of view indication thereon.
6. An image display apparatus comprising:
a data storage configured to store image data;
a user input interface configured to receive a request from a user;
a processor configured to generate image data for a 3D image with internal surface images at one or more depths in response to the user request; and
an image display configured to display the 3D image with internal surfaces at each depth.
7. An image display apparatus according to claim 6 , wherein the image display displays a depth indicator including a scroll bar and a plurality of buttons.
8. An image display apparatus according to claim 6 , wherein the image display displays a scroll bar, wherein, based on user input received via the user input interface, a user controls the depth of the 3D image.
9. An image display apparatus according to claim 8 , wherein the 3D image is an X-ray image.
10. An image display apparatus according to claim 8 , wherein the X-ray image is of a dental arch.
11. An image display apparatus according to claim 10 , wherein at least one internal surface of the 3D image shows an internal structure of teeth including tooth root and an alveolar bone.
12. An image display apparatus according to claim 11 , wherein the image display displays a model of a head and a field of view indication thereon.
13. A method of displaying images comprising:
displaying a three dimensional (3D) image;
receiving a user request; and
displaying a 3D image with internal surface images at one or more depths in response to the user request.
14. A method of displaying image according to claim 13 , wherein the 3D image is an X-ray image.
15. A method of displaying image according to claim 13 , wherein the X-ray image is of a dental arch.
16. A method of displaying image according to claim 13 , wherein at least one internal surface of the 3D image shows an internal structure of teeth including teeth root and an alveolar bone.
17. A method of displaying image according to claim 13 , wherein the display displays a model of a head and a field of view indication thereon.
18. (canceled)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR20130014496 | 2013-02-08 | ||
KR10-2013-0014496 | 2013-02-08 | ||
PCT/KR2014/001081 WO2014123396A1 (en) | 2013-02-08 | 2014-02-08 | Image display to display internal structure with change of depth |
Publications (2)
Publication Number | Publication Date |
---|---|
US20150374316A1 true US20150374316A1 (en) | 2015-12-31 |
US10210667B2 US10210667B2 (en) | 2019-02-19 |
Family
ID=51299939
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/766,856 Active US10210667B2 (en) | 2013-02-08 | 2014-02-08 | Displaying 3D image with a plurality of surface images at depths of interest |
US14/766,747 Active US9665990B2 (en) | 2013-02-08 | 2014-02-08 | Image display to display 3D image and sectional images |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/766,747 Active US9665990B2 (en) | 2013-02-08 | 2014-02-08 | Image display to display 3D image and sectional images |
Country Status (3)
Country | Link |
---|---|
US (2) | US10210667B2 (en) |
KR (2) | KR101731593B1 (en) |
WO (2) | WO2014123395A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112288886A (en) * | 2020-09-15 | 2021-01-29 | 陈学鹏 | Tooth position arrangement method of accurate digital tooth model |
Families Citing this family (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104600141B (en) * | 2015-02-06 | 2018-04-03 | 协鑫集成科技股份有限公司 | Solar cell module |
IT201600083061A1 (en) * | 2016-08-05 | 2018-02-05 | Aldo Amato | METHOD OF DETERMINING AND DESIGNING THE INDIVIDUAL IDEAL FORM OF TWO UPPER FRONT TEETH |
KR101877895B1 (en) * | 2016-10-06 | 2018-07-12 | 주식회사 메가젠임플란트 | Image Generation System for implant Diagnosis and the same |
KR101865701B1 (en) * | 2016-10-06 | 2018-06-11 | 주식회사 메가젠임플란트 | Mobile iinked implant diagnosis system |
WO2019045144A1 (en) * | 2017-08-31 | 2019-03-07 | (주)레벨소프트 | Medical image processing apparatus and medical image processing method which are for medical navigation device |
CN109963100B (en) * | 2017-12-14 | 2021-11-16 | 浙江宇视科技有限公司 | Cache additional recording method and device based on multicast sharing |
US20210137653A1 (en) * | 2019-11-12 | 2021-05-13 | Align Technology, Inc. | Digital 3d models of dental arches with accurate arch width |
KR102352985B1 (en) * | 2020-07-07 | 2022-01-20 | 한국과학기술원 | Method and apparatus system for volume data visualization interface through virtual widget based on touch screen |
KR102633824B1 (en) * | 2021-07-21 | 2024-02-06 | 오스템임플란트 주식회사 | Dental image providing apparatus and method thereof |
WO2023175001A1 (en) * | 2022-03-15 | 2023-09-21 | Avatar Medical | Method for displaying a 3d model of a patient |
US20230298163A1 (en) * | 2022-03-15 | 2023-09-21 | Avatar Medical | Method for displaying a 3d model of a patient |
EP4258216A1 (en) * | 2022-04-06 | 2023-10-11 | Avatar Medical | Method for displaying a 3d model of a patient |
CN116779093B (en) * | 2023-08-22 | 2023-11-28 | 青岛美迪康数字工程有限公司 | Method and device for generating medical image structured report and computer equipment |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6633789B1 (en) * | 2000-02-17 | 2003-10-14 | Align Technology, Inc. | Effiicient data representation of teeth model |
US20030199769A1 (en) * | 2002-04-08 | 2003-10-23 | Adrian Podoleanu | Apparatus for high resolution imaging of moving organs |
US20050149877A1 (en) * | 1999-11-15 | 2005-07-07 | Xenogen Corporation | Graphical user interface for 3-D in-vivo imaging |
US20100255445A1 (en) * | 2007-10-03 | 2010-10-07 | Bernard Gantes | Assisted dental implant treatment |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5734384A (en) | 1991-11-29 | 1998-03-31 | Picker International, Inc. | Cross-referenced sectioning and reprojection of diagnostic image volumes |
US5371778A (en) | 1991-11-29 | 1994-12-06 | Picker International, Inc. | Concurrent display and adjustment of 3D projection, coronal slice, sagittal slice, and transverse slice images |
IL120867A0 (en) * | 1997-05-20 | 1997-09-30 | Cadent Ltd | Computer user interface for orthodontic use |
KR200303415Y1 (en) | 2002-11-01 | 2003-02-07 | 박창수 | computed tomography |
US8295432B2 (en) | 2005-05-02 | 2012-10-23 | Oy Ajat Ltd | Radiation imaging device with irregular rectangular shape and extraoral dental imaging system therefrom |
US7742560B2 (en) | 2005-05-02 | 2010-06-22 | Oy Ajat Ltd. | Radiation imaging device with irregular rectangular shape and extraoral dental imaging system therefrom |
US7336763B2 (en) | 2005-05-02 | 2008-02-26 | Oy Ajat Ltd | Dental extra-oral x-ray imaging system and method |
US7676022B2 (en) | 2005-05-02 | 2010-03-09 | Oy Ajat Ltd. | Extra-oral digital panoramic dental x-ray imaging system |
CN101288102B (en) | 2005-08-01 | 2013-03-20 | 拜奥普蒂根公司 | Methods and systems for analysis of three dimensional data sets obtained from samples |
KR100947826B1 (en) | 2006-05-24 | 2010-03-18 | 주식회사 메디슨 | Apparatus and method for displaying an ultrasound image |
US8401257B2 (en) | 2007-01-19 | 2013-03-19 | Bioptigen, Inc. | Methods, systems and computer program products for processing images generated using Fourier domain optical coherence tomography (FDOCT) |
US8439672B2 (en) | 2008-01-29 | 2013-05-14 | Align Technology, Inc. | Method and system for optimizing dental aligner geometry |
DE102008008733A1 (en) | 2008-02-12 | 2009-08-13 | Sirona Dental Systems Gmbh | Method for creating a tomogram |
JP5390377B2 (en) | 2008-03-21 | 2014-01-15 | 淳 高橋 | 3D digital magnifier surgery support system |
WO2009140582A2 (en) * | 2008-05-16 | 2009-11-19 | Geodigm Corporation | Method and apparatus for combining 3d dental scans with other 3d data sets |
EP2123223B1 (en) | 2008-05-19 | 2017-07-05 | Cefla S.C. | Method and Apparatus for Simplified Patient Positioning in Dental Tomographic X-Ray Imaging |
EP2335596A1 (en) | 2009-12-15 | 2011-06-22 | Medison Co., Ltd. | Ultrasound system and method of selecting slice image from three-dimensional ultrasound image |
KR101183767B1 (en) | 2009-12-15 | 2012-09-17 | 삼성메디슨 주식회사 | Ultrasound system and method of selecting two-dimensional slice image from three-dimensional ultrasound image |
KR101126891B1 (en) | 2010-01-12 | 2012-03-20 | 삼성메디슨 주식회사 | Ultrasound system and method for providing slice image |
KR101117930B1 (en) * | 2010-05-13 | 2012-02-29 | 삼성메디슨 주식회사 | Ultrasound system and method for providing additional information with slice image |
-
2014
- 2014-02-08 KR KR1020157024564A patent/KR101731593B1/en active IP Right Grant
- 2014-02-08 US US14/766,856 patent/US10210667B2/en active Active
- 2014-02-08 WO PCT/KR2014/001080 patent/WO2014123395A1/en active Application Filing
- 2014-02-08 WO PCT/KR2014/001081 patent/WO2014123396A1/en active Application Filing
- 2014-02-08 KR KR1020157024562A patent/KR101731589B1/en active IP Right Grant
- 2014-02-08 US US14/766,747 patent/US9665990B2/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050149877A1 (en) * | 1999-11-15 | 2005-07-07 | Xenogen Corporation | Graphical user interface for 3-D in-vivo imaging |
US6633789B1 (en) * | 2000-02-17 | 2003-10-14 | Align Technology, Inc. | Effiicient data representation of teeth model |
US20030199769A1 (en) * | 2002-04-08 | 2003-10-23 | Adrian Podoleanu | Apparatus for high resolution imaging of moving organs |
US20100255445A1 (en) * | 2007-10-03 | 2010-10-07 | Bernard Gantes | Assisted dental implant treatment |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112288886A (en) * | 2020-09-15 | 2021-01-29 | 陈学鹏 | Tooth position arrangement method of accurate digital tooth model |
Also Published As
Publication number | Publication date |
---|---|
KR101731593B1 (en) | 2017-05-11 |
KR101731589B1 (en) | 2017-05-11 |
WO2014123396A1 (en) | 2014-08-14 |
WO2014123395A1 (en) | 2014-08-14 |
KR20160002703A (en) | 2016-01-08 |
KR20150122678A (en) | 2015-11-02 |
US9665990B2 (en) | 2017-05-30 |
US20150379780A1 (en) | 2015-12-31 |
US10210667B2 (en) | 2019-02-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10210667B2 (en) | Displaying 3D image with a plurality of surface images at depths of interest | |
Gallo et al. | 3D interaction with volumetric medical data: experiencing the Wiimote | |
US9053565B2 (en) | Interactive selection of a region of interest in an image | |
US10380787B2 (en) | Method and system for indicating light direction for a volume-rendered image | |
US10796498B2 (en) | Image processing apparatus, image processing method, and non-transitory computer-readable medium | |
JP2005342511A (en) | Method and apparatus for automatic protocol selection | |
JP6434959B2 (en) | Enabling users to study image data | |
JP2013192939A (en) | Method for interactive examination of root fracture | |
RU2706231C2 (en) | Visualization of three-dimensional image of anatomical structure | |
CN103534733A (en) | Medical image system and method | |
JP5838215B2 (en) | Image and annotation display | |
KR20160087772A (en) | Method for an exchange of data between a medical imaging apparatus and a user and a medical imaging apparatus therefor | |
JP2014512607A (en) | Image processing system and method | |
US20140055448A1 (en) | 3D Image Navigation Method | |
JP6276529B2 (en) | Interpretation support apparatus, method and program | |
US10699424B2 (en) | Image processing apparatus, image processing method, and non-transitory computer readable medium with generation of deformed images | |
EP3423968B1 (en) | Medical image navigation system | |
JPH11283018A (en) | Medical image processor | |
JP2020028583A (en) | Medical report creation device, medical report creation method, and medical report creation program | |
Cheng et al. | GPU based multi-histogram volume navigation for virtual bronchoscopy | |
CN104185448A (en) | Medical image display device, medical image display method and medical image display program | |
JP2004173727A (en) | Diagnostic imaging supporting equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: EWOOSOFT CO., LTD, KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:JIN, YOUNGKYU;REEL/FRAME:036290/0144 Effective date: 20140514 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |