CA2459732A1 - Concealed object recognition - Google Patents

Concealed object recognition Download PDF

Info

Publication number
CA2459732A1
CA2459732A1 CA002459732A CA2459732A CA2459732A1 CA 2459732 A1 CA2459732 A1 CA 2459732A1 CA 002459732 A CA002459732 A CA 002459732A CA 2459732 A CA2459732 A CA 2459732A CA 2459732 A1 CA2459732 A1 CA 2459732A1
Authority
CA
Canada
Prior art keywords
data
object data
values
computer code
program product
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CA002459732A
Other languages
French (fr)
Other versions
CA2459732C (en
Inventor
Gene Arthur Grindstaff
Susan Heath Calvin Fletcher
Therman Ward Mckay, Iii
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Intergraph Corp
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Publication of CA2459732A1 publication Critical patent/CA2459732A1/en
Application granted granted Critical
Publication of CA2459732C publication Critical patent/CA2459732C/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/68Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/68Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
    • H04N23/682Vibration or motion blur correction
    • H04N23/683Vibration or motion blur correction performed by a processor, e.g. controlling the readout of an image memory
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/90Arrangement of cameras or camera modules, e.g. multiple cameras in TV studios or sports stadiums
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/181Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/64Circuits for processing colour signals
    • H04N9/646Circuits for processing colour signals for image enhancement, e.g. vertical detail restoration, cross-colour elimination, contour correction, chrominance trapping filters

Abstract

A method for viewing a first object that is obstructed by a second object. In such a method, the first object has a contrasting color to the second object and the second object is constructed from a material that allows visible light to pass therethrough. The amount of visible light that passes through the second object is not enough such that the first object is visible to the human eye. The method involves taking a digital image of the first and second object using a visible light sensor, such as a CCD camera sensor. The digital image data that is received into a computer system contains both first object data and second object data. It should be understood that the first object data and the second object data include color information. The amount of contrast between the first and the second object should be approximately 10% of the total scale such that on a 256 color scale the difference is approximately 25 levels. The computer system substantially filters out the second object data and then the values associated with the first object data are increased until the data is visibly displayable.

Description

Concealed Object Recognition Technical Field The present invention relates to analysis of digital images. More specifically, the present invention relates to viewing of obstructed objects in digital images.
Back r~ ound Systems have been designed for viewing objects which are obstructed by other objects. For instance, x-ray detectors and infra-red detectors allow a user to subject an object to either infra-red light, x-rays, or another signal other than visible light in order to expose obstructed objects. Such systems require that the object is exposed to the signal in order for the object to be observed. It would be desirable to have a method and system for viewing obstructed objects within a digital image without having to first expose the objects to a created signal.
Summary of the Invention The invention in one embodiment is a method for viewing a first object that is obstructed by a second object. In such a method the first object has a contrasting color to the second object and the second object is constructed from a material that allows visible light to pass therethrough. The amount of visible light that passes through the second object is not enough such that the first object is visible to the human eye. The method involves taking a digital image of the first and second object using a visible light sensor, such as a CCD
camera sensor. The digital image data that is received into a computer system contains both first object data and second object data. It should be understood that the first object data and the second object data include color information. The amount of contrast between the first and the second object should be approximately 10% of the total scale such that on a 256 color scale the difference is approximately 25 levels. The computer system substantially filters out the second object data and then the values associated with the first object data are increased until the data is visibly displayable.
The computer system may create a histogram of the digital image data to allow a user of the system to select second object data based upon the histogram. In such an embodiment, the histogram is displayed to the user and a user selection mechanism is provided for selecting data from the histogram.
The step of substantially filtering out the second object data and increasing the first object data may be achieved by increasing gamma correction to a value that is outside of the range of current graphics programs. For example, the gamma correction value may be set to a value above 2.5 and more generally in the range above 6. Filtering of the second object data can also be achieved by increasing the brightness in combination with increasing the gamma correction value.
The first object data is generally increased such that the original first object data is mapped to values where the contrast between data levels is visible to the human eye when the first object data is displayed on a display device. The mapping is also such that the original values that are low level light values which cannot be perceived when displayed on a display device are mapped to values that can be perceived by a human eye.
The method may include the step of converting the color information to greyscale before filtering the second object data and increasing/remapping the color information of the first object data.
The computer system may identify whether the first object is lighter than the second object or darker than the second object. If the second object is substantially lighter than the first object, the system may invert the first and second object data such that the second object appears darker when represented on a display device prior to filtering. The second object data may be filtered by decreasing all of the values of the second object data until the values are that for black which generally is value 0 on a scale from 0-255.
The methodology as described may also be created in a computer program product for use with a computer wherein the computer program
2
3 PCT/US02/28351 product contains computer code thereon. The method in one embodiment is completely automated. The methodology may also be borne out in a system that includes a camera for taking a digital image using visible light wherein the digital image that is created by the digital camera contains data representative of the first object and of the second object. The system further includes a processor for receiving the digital image and processing the digital image to filter out the second object data and increasing the values of the first object data until visibly displayable. The system may be automated or receive user activation signals for the identification of data as belonging to the second object and also gamma correction values for example.
Brief Description of the Drawings The features of the invention will be more readily understood by reference to the following detailed description, taken with reference to the accompanying drawings, in which:
Fig.1 shows one environment for implementing a first embodiment for concealed object recognition;
Fig 2 is a flow diagram illustrating recognition of a first object concealed by a second object from a video image;
Fig. 3 shows a more detailed flowchart of the steps that are taken for concealed object recognition;
Fig. 4 is a sample histogram;
Fig. 5 is a flow chart showing one method for implementing concealed object recognition; and Fig. 6 is a flow chart showing an automated version of the method for concealed object recognition.
Detailed Description of Specific Embodiments As used in the following text, the term "digital video" implies a digital representation of a sequence of images which may be temporally displayed on a display device. Typically, digital video contains a plurality of frames wherein each frame represents a separate image. The frames may be further subdivided such that the frames are made up of a series of pixels. As used in the following description the term "pixel°' shall mean a single point of an image.
The greater the number of pixels that are contained in an image, the greater the resolution of the video. Resolutions are conventionally referenced by length and width measurements of the number of pixels, for example, in a resolution of 800x600, there are 800 pixels along the length of an image by 600 pixels along the width of the image.
The term "brightness" shall mean the attribute of a visual sensation according to which an area appears to emit more or less light. Brightness is a relative expression of the intensity of the energy output of a visible light source. The term "gamma" and "gamma correction value" shall mean the exponent of the luminance power law response to voltage. The term "gamma correction" shall mean the insertion of a nonlinear output-input characteristic for the purpose of changing the system transfer characteristic. The term "normal video signal" shall mean a video signal that is produced using visible light as opposed to infrared or subjecting the subject matter of the image to some type of signal other than visible light waves. Tile term "porous material"
is any material that allows a portion of direct light to reflect through the material.
In an embodiment of the present invention, a method of detecting concealed objects using analysis of video image frames is presented. Under certain conditions, objects under a porous material can be detected in a video frame. Examples of porous materials include clothing (cotton, silk etc.), fabric and glass. The concealed object can be displayed if the concealed object and the porous material are of contrasting colors. A contrasting color is a color which when transferred to a greyscale varies by approximately 10% in grey scale level. For example, if there are 256 greyscale levels the porous material and the concealed object should vary by 25 levels or more.
Fig. 1 shows a digital image camera 100 capturing image data that is the
4 result of reflected light. The majority of the direct light 105 is reflected off of a porous object 110. A small portion of the direct light 115 passes through the porous material and is reflected off of the concealed object 120. Some of the reflected light from the concealed object passes through the porous material and is captured by a video image sensor 130. The video image sensor 130 turns the reflected light 105,115 from both the porous material and the concealed material into an electric signal 140 which is processed by a digital image processor 150. The digital image processor 150 then sends the image data 155 to memory in a digital format such as RGB for each pixel/receptor in the sensor.
When the image data 155 is then displayed on a display device the concealed object is not visible, due to the fact that the pixel values which represent the porous material are much greater than that for the concealed object. As a result, the light for the porous material that is emitted from the display device engulfs the emitted light which results from the pixels values of the concealed object.
Therefore, the human eye cannot perceive the concealed object. In general, the human eye is not capable of observing intensity levels below 80 on a grey-scale between 0-255. As such, values below 80 will appear to be black. In Fig. 1 the concealed object is shown to be a solid object, although the concealed object could also be porous:
Fig.2 shows a flow chart of the steps that are taken for altering digital image data in order to display a concealed object. The method is performed on a computer system using video or image editing software. First information which is digital data that is representative of a first digital image/frame is received 200. The data is then converted from color digital image data to greyscale data. The digital greyscale data is then filtered 210. Filtering occurs in the following manner. If the porous object is not black, all of the pixel values are adjusted such that the pixel values associated with the porous object default to black. This may be done by applying a filter that cuts off values above a threshold greyscale value. Then the remaining values which are generally representative of the concealed object are increased 220. The remaining values
5 which have not been set to 0 (black) are increased such that the values are in a range which is visibly perceptible. For example, assuming a greyscale between 0-255 all of the values would be increased above 100 for example. The values are thus remapped, such that the values are spread out causing the levels of contrast to be perceivable and the overall brightness/intensity level for the pixels representative of the concealed object are increased such that the pixels are visible to the human eye. It should be understood by one of ordinary skill in the art that the human eye does not readily differentiate between pixels that have closely proximate greyscale values. Therefore, one object of the filter is to spread out the low level pixel values such that the human eye can perceive the contrast.
Fig. 3 shows a more detailed flowchart of the steps that are taken for concealed object recognition. As previously stated once the digital image information is received into a computer system, the color information is converted to greyscale information by techniques known to those of ordinary skill in the art. Although the same methodology could also be applied to the color information.
First, a histogram of the digital image data forming a first image in a sequence of images is produced 300. A sample histogram is shown in Fig. 4.
The number of occurrences of a value is represented on the vertical axis and the greyscale values are represented on the horizontal axis. This histogram thus indicates how many times intensity occurs for a given image. The histogram provides a weighting to allow a user to select the low intensity levels. In general the high intensity values will predominate and the less visible values (low intensity values) will be much smaller both in intensity and generally in number of occurrences. Thus, the visible values can be readily identified and removed through filtering or other techniques known to one of ordinary skill in the art.
In one embodiment, the computer system allows a user to view the histogram and determine the low level intensity light portions from the
6 graphical representation of the histogram 310. Low light intensity levels are those which result from visible light but which would not be perceived by the human eye if displayed on a display device. The user then selects a filter and filters out all other light (high intensity light) 320. Normally the low level intensity light can be visually determined and selected by a user of the computer system. Once the high intensity light is filtered out, the low level light is re-mapped to higher level values so that a display device will display the re-mapped values 330. These values are also spread out so that the contrast is more dramatic. For example, if the cut off between low and high intensity is 40 on a scale from 0-255 and the values are being mapped between 135 and 255, contrast is increased by three such that a value of 1 and 2 would not be 135 and 139. The two mapped low intensity levels are then displayed to a user of the system 340. If the object represented by the low intensity levels is visible, the user responds that the image is unacceptable 350. The user indicates this through a user input device such as a keyboard or mouse. The method then provides for the user to either re-determine the threshold between low and high intensity levels or provide for a new remapping of the low intensity values such that the values are either brighter, more contrast is provided between values or both. The system may also let the user select a new value for separating between a low-level intensity value and high-level intensity value 350. If the user indicates that the concealed object is visible, the process ends.
The process may also be achieved by increasing gamma correction and/or brightness as shown in Fig. 5, first the brightness and/or the gamma correction value is applied to the digital image data 410. By increasing the brightness level, the high level intensity light eventually saturates toward white, thus filtering the high level intensity light. All values that are saturated can then be removed leaving the low level values. The gamma correction can then be used to increase the remaining values. Adjusting gamma increases the intensity values without over amplifying the noise that occurs by just increasing brightness proportionally across all values. Increasing the gamma
7 increases the intensity in a non-linear fashion such that the new intensity is:
current intensity gQ,nn,a new intensity = ( ) Because of the non-linear nature of gamma correction, the noise is also increased in a non-linear fashion.
In this embodiment of the invention, the gamma correction value applied must be greater in value than is normally used with a cathode ray tube ("CRT"') display Typically the gamma correction value should be between six to ten for this embodiment wherein a standard cathode display has a gamma correction value of approximately 2.2-2.5 and prior art video editing software have maximum gamma correction values of 5.
Once a filter has been applied to the video image data and the low level intensity video image data has been amplified, the low level intensity video image data representing the current frame is displayed 420 to the user. The data that is displayed to the user may also include some of the high level intensity video data. The user can then interactively change 430 the gamma and brightness values and the resulting values are again displayed to the user.
Through this process, the image represented by a low level light is displayed.
When the image is satisfactory, the user causes the computer system to write the altered' video data of the frame 450 to the video store. Succeeding video frames are processed in a similar manner until all frames have been processed 480. Concealed object recognition is then complete 490.
In a first example of the method for concealed object recognition, an image of a person wearing a black shirt with a white gun underneath is analyzed. A digital image is captured of the person under normal lighting conditions. If the digital image is displayed on a display device, the human eye can only see a black shirt which has a value of zero for the pixels displayed on the screen. The light that is reflected through by the white gun is muted in the process and thus may be of such an intensity level that is not perceivable by the human eye. Although the light is not perceivable, the light information is captured by the sensor of the digital image device. If gamma and brightness are
8 increased the black values will not increase. However, the slight amount of light coming from the white gun will be amplified since it is not a zero value.
Some noise will also be amplified and thus displayed, but if enough light, preferably 500 Lux or more, is available to the camera, the system will clearly show the white gun. In another example, an image of a person wearing a white shirt with a black gun underneath is analyzed. In such an image, the video is inverted thus filtering the high intensity data, such that the shirt now appears black and therefore provides a 0 intensity value. The low level intensity data is then mapped to a range in which the values are in the visible range, for example, between 100-255 on a scale from 0-255. This may be achieved by first converting the data to greyscale data or using the color-based data that is the result of the digital imaging sensor. Additionally, the values are spaced such that the distance between values is increased, thus making contrast perceptions more likely to the human eye.
In a third example, an image of a person wearing a dark grey shirt and having a light grey gun underneath the shirt is analyzed. The intensity level corresponding to the dark grey suit is decreased such that it is set to black.
The light grey gun will then be at some value which is above that of black. The values corresponding to the light grey gun as well as noise can then be increased using either brightness or gamma correction, thus the low-light values of the light grey gun are filtered and mapped to values which are perceivable by the human eye, as previously explained.
This process can be automated as shown in Fig. 5. In one variation, the mean for all the digital data of the image is computed 510. The mean determines where the high intensity values are. Next a standard deviation from the mean is determined 520. All of the values within one standard deviation below the mean are then identified as high intensity values as well as all values above the mean 530. The high intensity values are then filtered and thus removed from the digital data set that makes up the image 540. The low level values are then mapped to values that are visible 550. The spacing between
9 values is also increased such that the contrast between values is more perceptible to the human eye. For example, the automated system may determine the lowest and highest value within the low level values and subtract the two values to determine the range. This range is then mapped to a visible light range. For example, if the scale is between 0-255 and the low level values are between 1-41 the range is 40. Values above 80 are normally perceivable to the human eye so the range of 40 would be mapped to a range of 175 (80-255). Thus, all values from the low level light would extend between the values of 80 and 255. The low level data values are then displayed on a display device. 560.
The system may be further automated to change the color data to greyscale data for processing. It should be understood, that although the data may be converted to grayscale for determining high intensity and low intensity values, the color information is still maintained and may be used for the remapping of the low intensity values, such that, the low intensity values may be displayed in color after being remapped to visible values.
The camera angle should be within plus or minus 40 degrees of normal to the object for best results. If the outside clothing is very light or white, then the video may be inverted so that light objects become dark and contrasting objects underneath the porous materials are detected. An ambient light condition of 500 Lux or greater should illuminate the object under study.
It should be noted that the flow diagrams are used herein to demonstrate various aspects of the invention, and should not be construed to limit the present invention to any particular logic flow or logic implementation.
The described logic may be partitioned into different logic blocks (e.g., programs, modules, functions, or subroutines) without changing the overall results or otherwise departing from the true scope of the invention.
Oftentimes, logic elements may be added, modified, omitted, performed in a different order, or implemented using different logic constructs (e.g., logic gates, looping primitives, conditional logic, and other logic constructs) without changing the overall results or otherwise departing from the true scope of the invention.
The present invention may be embodied in many different forms, including, but in no way limited to, computer program logic for use with a processor (e.g., a microprocessor, microcontroller, digital signal processor, or general purpose computer), programmable logic for use with a programmable logic device (e.g., a Field Programmable Gate Array (FPGA) or other PLD), discrete components, integrated circuitry (e.g., an Application Specific Integrated Circuit (ASIC)), or any other means including any combination thereof.
Computer program logic implementing all or part of the functionality previously described herein may be embodied in various forms, including, but in no way limited to, a source code form, a computer executable form, and various intermediate forms (e.g., forms generated by an assembler, compiler, linker, or locator.) Source code may include a series of computer program instructions implemented in any of various programming languages (e.g., an object code, an assembly language, or a high-level language such as Fortran, C, C++, JAVA, or HTML) for use with various operating systems or operating environments. The source code may define and use various data structures and communication messages. The source code may be in a computer executable form (e.g., via an interpreter), or the source code may be converted (e.g., via a translator, assembler, or compiler) into a computer executable form.
The computer program may be fixed in any form (e.g., source code form, computer executable form, or an intermediate form) either permanently or transitorily in a tangible storage medium, such as a semiconductor memory device (e.g., a RAM, ROM, PROM, EEPROM, or Flash-Programmable RAM), a magnetic memory device (e.g., a diskette or fixed disk), an optical memory device (e.g., a CD-ROM), a PC card (e.g., PCMCIA card), or other memory device. The computer program may be fixed in any form in a signal that is transmittable to a computer using any of various communication technologies, including, but in no way limited to, analog technologies, digital technologies, optical technologies, wireless technologies, networking technologies, and internetworking technologies. The computer program may be distributed in any form as a removable storage medium with accompanying printed or electronic documentation (e.g., shrink wrapped software or a magnetic tape), preloaded with a computer system (e.g., on system ROM or fixed disk), or distributed from a server or electronic bulletin board over the communication system (e.g., the Internet or World Wide Web.) Hardware logic (including programmable logic for use with a programmable logic device) implementing all or part of the functionality previously described herein may be designed using traditional manual methods, or may be designed, captured, simulated, or documented electronically using various tools, such as Computer Aided Design (CAD), a hardware description language (e.g., VHDL or AHDL), or a PLD programming language (e.g., PALASM, ABEL, or CUPL.) The present invention may be embodied in other specific forms without departing from the true scope of the invention. The described embodiments are to be considered in all respects only as illustrative and not restrictive.
01247/00A66 216724.1

Claims (55)

1. A method for viewing a first object having a contrasting color to a second object, wherein the second object obstructs visibility of the first object, wherein visible light passes through the second object, the method comprising:
receiving digital image data representative of the first object obstructed by the second object into a computer system; wherein the digital image data contains at least first object data and second object data produced by visible light and each datum has an associated value;
substantially filtering out the second object data; and increasing the value of the first object data until visibly displayable.
2. The method according to claim 1, further comprising creating a histogram of the digital image data; and selecting the second object data based upon the histogram.
3. The method according to claim 1, wherein substantially filtering out the second object data and increasing the first object data is achieved at least by increasing gamma correction.
4. The method according to claim 3 wherein the gamma correction is set above 2.5.
5. The method according to claim 3 wherein the gamma correction is set at or above 6.
6. The method according to claim 1, wherein increasing the first object data includes mapping the first object data to values such that the contrast between data levels is such that a human eye can detect the contrast when the first object data is displayed on a display device.
7. The method according to claim 6, wherein mapping is performed such that the mapping is to data values which when displayed by a display device are observable to the human eye.
8. The method according to claim 1, wherein increasing the first object data is achieved at least by increasing brightness.
9. The method according to claim 1, wherein the first object and the second object have contrasting colors.
10. The method according to claim 1, wherein the first object and the second object have contrasting colors which average more than 25 levels apart on a 256 level scale.
11. The method according to claim 1, wherein the image data is represented on a scale of values and wherein the first object and the second object have contrasting colors such that the colors of the first object data and the second object data vary approximately by ten percent of the scale.
12. The method according to claim 1, wherein the process is automated.
13. The method according to claim 2 wherein the process is automated.
14. The method according to claim 1, wherein the first object data and the second object data contain color information.
15. The method according to claim 14, further comprising:
converting the color information from the first object data and the second object data to greyscale.
16. A method for displaying first object data representative of a first object in a digital image on a display device wherein the first object is obstructed by a second object, visible light passing through the second object, the first object and the second object having contrasting colors, the method comprising:
receiving digital data representative of the digital image into a computer system; and filtering the first and second object data such that the first object data is visibly displayed on the display device and the second object data is clipped.
17. The method according to claim 16, further comprising:
taking a digital image of the first object and the second object using a visible light sensor that produces the first object data and the second object data.
18. The method according to claim 16, wherein filtering comprises:
mapping the first object data to a scale such that contrasts between data values when displayed are visible to a human eye.
19. The method according to claim 16, further comprising:
displaying the first object data on a display device.
20. The method according to claim 19, further comprising:
providing manual adjustment of the filtering after displaying the first object data on the display device.~
21. The method according to claim 16 wherein filtering includes:
creating a visually displayable histogram of the digital data representative of the digital image;
providing user selection of a group of the digital data representative of the digital image;
removing the selected group from the digital data representative of the digital image; and adjusting data values of the remaining digital data by applying gamma correction to the digital data by increasing a gamma correction value until the first object is displayable when the digital data is displayed on a display device.
22. The method according to claim 16, wherein the first object data and the second object data contain color information.
23. The method according to claim 22, further comprising:
converting the color information of the first object data and the second object data to greyscale.
24. The method according to claim 23, wherein if the second object is darker than the first object, adjusting at least the second object data until substantially all of the second object data produces zero output when represented on a display.
25. The method according to claim 23, wherein if the second object is substantially lighter than the first object, inverting the first and second object data such that the second object appears darker when represented on a display device.
26. The method according to claim 25, adjusting at least the second object data until substantially all of the second object data produces zero output when represented on a display.
27. The method according to claim 25 further comprising:
mapping the first object data to data values such that the data values when presented on a display are visible and scaling between data values is increased over scaling between the first object data.
28. The method according to claim 26 further comprising:
mapping the first object data to data values such that the data values when presented on a display are visible and scaling between data values is increased over scaling between the first object data.
29. A computer program product having computer readable computer code for viewing first object data representative of a first object having a contrasting color to a second object, wherein the second object obstructs visibility of the first object, wherein the visible light passes through the second object, the computer code comprising:
computer code for receiving digital image data representative of the first object obstructed by the second object into a computer system; wherein the digital image data contains at least first object data and second object data produced by visible light;
computer code for substantially filtering out the second object data; and computer code for increasing the first object data until visibly displayable.
30. The computer program product according to claim 29, further comprising:
computer code for creating a histogram of the digital image data; and computer code for selecting the second object data based upon the histogram.
31. The computer program product according to claim 29, wherein the computer code for substantially filtering out the second object data and increasing the first object data increases gamma correction.
32. The computer program product according to claim 31 wherein the gamma correction is set above 2.5.
33. The computer program product according to claim 31 wherein the gamma correction is set at or above 6.
34. The computer program product according to claim 1, wherein the computer code for increasing the first object data includes computer code for mapping the first object data to values such that the contrast between data levels is such that a human eye can detect the contrast.
35. The computer program product according to claim 34, wherein the computer code for mapping is performed such that the mapping is to data values which when displayed by a display device are observable to the human eye.
36. The computer program product according to claim 29, wherein the computer code for increasing the first object data increases the first object data at least by increasing brightness.
37. The computer program product to claim 29, wherein the first object and the second object have contrasting colors.
38. The computer program product according to claim 29, wherein the first object and the second object data represent contrasting colors which average more than 25 levels on a 256 level scale.
39. The computer program according to claim 29, wherein the image data is associated with a scale and the first object and the second object have contrasting colors such that the first object data and the second object data vary substantially by ten percent of the scale.
40. The computer program product according to claim 29, wherein the first object data and the second object data contain color information and the computer program product further includes computer code for converting the color information from the first object data and the second object data to greyscale.
41. computer code on a computer readable medium for displaying first object data representative of a first object in a digital image on a display device wherein the first object is obstructed by a second object, visible light passing through the second object, the first object and the second object having contrasting colors, the computer code comprising:
computer Bode for receiving digital data representative of the digital image into a computer system; and computer code for filtering the first and second object data such that the first object data is visibly displayed on the display device and the second object data is clipped.
42. The computer program product according to claim 41, wherein the computer code for filtering includes:
computer code for mapping the first object data to a scale such that contrasts between data values when displayed are visible to a human eye.
43. The computer program product according to claim 41, further comprising:
computer code for displaying the first object data on a display device.
44. The computer program product according to claim 41, further comprising:
computer code for providing user adjustment of the filtering after displaying the first object data on the display device.
45. The computer program product according to claim 41 wherein the computer code for filtering includes:
computer code for creating a visually displayable histogram of the digital data representative of the digital image;
computer code for providing user selection of a group of the digital data representative of the digital image;
computer code for removing the selected group from the digital data representative of the digital image; and computer code for adjusting data values of the remaining digital data by applying gamma correction to the digital data by increasing a gamma correction value until the first object is displayable when the digital data is displayed on a display device.
46. The computer program product according to claim 41, wherein the first object data and the second object data contain color information and wherein the computer program product further comprises:

computer code for converting the color information of the first object data and the second object data to greyscale.
47. The computer program product according to claim 46, further comprising computer code for adjusting at least the second object data until substantially all of the second object data produces zero output when represented on a display if the second object is darker than the first object.
48. The computer program product according to claim 46, further comprising computer code for inverting the first and second object data such that the second object appears darker when represented on a display device if the second object is substantially lighter than the first object.
49. The computer program product according to claim 48, further comprising computer code for adjusting at least the second object data until substantially all of the second object data produces zero output when represented on a display.
50. The computer program product according to claim 48 further comprising:
computer code for mapping the first object data to data values such that the data values when presented on a display are visible and scaling between data values is increased over scaling between the first object data.
51. The computer program product according to claim 49 further comprising:
computer code for mapping the first object data to data values such that the data values when presented on a display are visible and scaling between data values is increased over scaling between the first object data.
52. A system for detecting a first object concealed by a second object, the system comprising:
a camera for taking a digital image using visible light, the digital image containing data representative of the first object and of the second object;
a processor for receiving the digital image and processing the digital image to filter out the second object data and increasing the values of the first object data until visibly displayable.
53. The system according to claim 52 wherein the processor receives a user activated signal and the processor increases a gamma,correction value.
54. The system according to claim 53 wherein the gamma correction value is increased at or above 5.
55. The system according to claim 51 wherein the processor both increases the values of the first object data, wherein the first object data is on a scale and increases the scaling between first object data.
CA2459732A 2001-09-07 2002-09-06 Concealed object recognition Expired - Fee Related CA2459732C (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US31816401P 2001-09-07 2001-09-07
US60/318,164 2001-09-07
PCT/US2002/028351 WO2003030103A1 (en) 2001-09-07 2002-09-06 Concealed object recognition

Publications (2)

Publication Number Publication Date
CA2459732A1 true CA2459732A1 (en) 2003-04-10
CA2459732C CA2459732C (en) 2017-07-11

Family

ID=23236945

Family Applications (3)

Application Number Title Priority Date Filing Date
CA2459821A Expired - Lifetime CA2459821C (en) 2001-09-07 2002-09-06 Image stabilization using color matching
CA2459823A Expired - Fee Related CA2459823C (en) 2001-09-07 2002-09-06 Method, device and computer program product for demultiplexing of video images
CA2459732A Expired - Fee Related CA2459732C (en) 2001-09-07 2002-09-06 Concealed object recognition

Family Applications Before (2)

Application Number Title Priority Date Filing Date
CA2459821A Expired - Lifetime CA2459821C (en) 2001-09-07 2002-09-06 Image stabilization using color matching
CA2459823A Expired - Fee Related CA2459823C (en) 2001-09-07 2002-09-06 Method, device and computer program product for demultiplexing of video images

Country Status (15)

Country Link
US (6) US6654049B2 (en)
EP (3) EP1428378B1 (en)
KR (3) KR100919340B1 (en)
CN (4) CN101072364B (en)
AU (3) AU2002336445B2 (en)
BR (3) BRPI0212377B1 (en)
CA (3) CA2459821C (en)
CO (3) CO5650190A2 (en)
HK (3) HK1069470A1 (en)
IL (4) IL160760A0 (en)
MX (3) MXPA04002210A (en)
NZ (3) NZ531976A (en)
RU (3) RU2308816C2 (en)
WO (3) WO2003024112A1 (en)
ZA (2) ZA200401972B (en)

Families Citing this family (65)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003024112A1 (en) * 2001-09-07 2003-03-20 Intergraph Hardware Technologies Company Method, device and computer program product for demultiplexing of video images
US7489802B2 (en) * 2002-09-10 2009-02-10 Zeev Smilansky Miniature autonomous agents for scene interpretation
JP4307910B2 (en) 2003-03-07 2009-08-05 富士フイルム株式会社 Moving image clipping device and method, and program
JP3768968B2 (en) 2003-03-28 2006-04-19 株式会社エヌ・ティ・ティ・ドコモ Communication terminal and program
JP4017578B2 (en) * 2003-09-19 2007-12-05 三洋電機株式会社 Camera shake correction device, camera shake correction method, and recording medium recording camera shake correction program
US7840067B2 (en) * 2003-10-24 2010-11-23 Arcsoft, Inc. Color matching and color correction for images forming a panoramic image
US7506267B2 (en) * 2003-12-23 2009-03-17 Intel Corporation Compose rate reduction for displays
US7616220B2 (en) * 2003-12-23 2009-11-10 Intel Corporation Spatio-temporal generation of motion blur
KR100562119B1 (en) * 2004-04-29 2006-03-21 삼성탈레스 주식회사 Image Alignment Method by Using Characteristics of Moving Image
US7671916B2 (en) * 2004-06-04 2010-03-02 Electronic Arts Inc. Motion sensor using dual camera inputs
WO2006023863A1 (en) 2004-08-23 2006-03-02 Intergraph Software Technologies Company Real-time image stabilization
US20060120615A1 (en) * 2004-12-06 2006-06-08 Huiqiong Wang Frame compensation for moving imaging devices
US8245313B2 (en) * 2005-03-31 2012-08-14 Honda Motor Co., Ltd Document management system, document management program, document management system configuration method, and server computer
US7952612B2 (en) * 2006-06-22 2011-05-31 Nokia Corporation Method and system for image construction using multiple exposures
US20090010617A1 (en) * 2007-07-05 2009-01-08 International Business Machines Corporation Method and Apparatus for Optimizing Space Allocations for Digital Video Recordings
KR101392732B1 (en) 2007-08-20 2014-05-08 삼성전자주식회사 Apparatus and method for estimating motion by hand trembling, and image pickup device using the same
KR101268987B1 (en) * 2007-09-11 2013-05-29 삼성전자주식회사 Method and apparatus for recording multimedia data by automatically generating/updating metadata
US7800652B2 (en) * 2007-12-12 2010-09-21 Cyberlink Corp. Reducing video shaking
CN101622869B (en) * 2007-12-18 2012-03-07 松下电器产业株式会社 Image reproduction device, image reproduction method, and image reproduction program
US8493313B2 (en) 2008-02-13 2013-07-23 Dolby Laboratories Licensing Corporation Temporal filtering of video signals
US8260395B2 (en) * 2008-04-18 2012-09-04 Medtronic, Inc. Method and apparatus for mapping a structure
US8164655B2 (en) * 2008-05-19 2012-04-24 Spatial Cam Llc Systems and methods for concurrently playing multiple images from a storage medium
US20100026822A1 (en) * 2008-07-31 2010-02-04 Itt Manufacturing Enterprises, Inc. Multiplexing Imaging System for Area Coverage and Point Targets
JP5233708B2 (en) * 2009-02-04 2013-07-10 ソニー株式会社 Information processing apparatus, information processing method, and program
US20100235314A1 (en) * 2009-02-12 2010-09-16 Decisive Analytics Corporation Method and apparatus for analyzing and interrelating video data
US8458105B2 (en) * 2009-02-12 2013-06-04 Decisive Analytics Corporation Method and apparatus for analyzing and interrelating data
EP2355037A1 (en) * 2009-12-18 2011-08-10 Nxp B.V. Method of and system for determining an average colour value for pixels
US9183560B2 (en) 2010-05-28 2015-11-10 Daniel H. Abelow Reality alternate
CN102457701B (en) * 2010-11-02 2014-03-12 华为终端有限公司 Image display processing method and device for same
JP4932027B1 (en) * 2010-11-15 2012-05-16 株式会社ナナオ Image display method and liquid crystal display device using the same
JP6122269B2 (en) 2011-12-16 2017-04-26 キヤノン株式会社 Image processing apparatus, image processing method, and program
US10715817B2 (en) * 2012-12-19 2020-07-14 Nvidia Corporation Apparatus and method for enhancing motion estimation based on user input
US9530310B2 (en) 2013-11-01 2016-12-27 Xerox Corporation Method and system for detecting and tracking a vehicle of interest utilizing a network of traffic image-capturing units
US9892337B1 (en) 2014-06-27 2018-02-13 Blinker, Inc. Method and apparatus for receiving a refinancing offer from an image
US10572758B1 (en) 2014-06-27 2020-02-25 Blinker, Inc. Method and apparatus for receiving a financing offer from an image
US10540564B2 (en) 2014-06-27 2020-01-21 Blinker, Inc. Method and apparatus for identifying vehicle information from an image
US10867327B1 (en) 2014-06-27 2020-12-15 Blinker, Inc. System and method for electronic processing of vehicle transactions based on image detection of vehicle license plate
US9773184B1 (en) 2014-06-27 2017-09-26 Blinker, Inc. Method and apparatus for receiving a broadcast radio service offer from an image
US9607236B1 (en) 2014-06-27 2017-03-28 Blinker, Inc. Method and apparatus for providing loan verification from an image
US9558419B1 (en) 2014-06-27 2017-01-31 Blinker, Inc. Method and apparatus for receiving a location of a vehicle service center from an image
US9760776B1 (en) 2014-06-27 2017-09-12 Blinker, Inc. Method and apparatus for obtaining a vehicle history report from an image
US10515285B2 (en) 2014-06-27 2019-12-24 Blinker, Inc. Method and apparatus for blocking information from an image
US9589202B1 (en) 2014-06-27 2017-03-07 Blinker, Inc. Method and apparatus for receiving an insurance quote from an image
US10579892B1 (en) 2014-06-27 2020-03-03 Blinker, Inc. Method and apparatus for recovering license plate information from an image
US9818154B1 (en) 2014-06-27 2017-11-14 Blinker, Inc. System and method for electronic processing of vehicle transactions based on image detection of vehicle license plate
US9754171B1 (en) 2014-06-27 2017-09-05 Blinker, Inc. Method and apparatus for receiving vehicle information from an image and posting the vehicle information to a website
US9563814B1 (en) 2014-06-27 2017-02-07 Blinker, Inc. Method and apparatus for recovering a vehicle identification number from an image
US9594971B1 (en) 2014-06-27 2017-03-14 Blinker, Inc. Method and apparatus for receiving listings of similar vehicles from an image
US9779318B1 (en) 2014-06-27 2017-10-03 Blinker, Inc. Method and apparatus for verifying vehicle ownership from an image
US9589201B1 (en) 2014-06-27 2017-03-07 Blinker, Inc. Method and apparatus for recovering a vehicle value from an image
US10733471B1 (en) 2014-06-27 2020-08-04 Blinker, Inc. Method and apparatus for receiving recall information from an image
US9600733B1 (en) 2014-06-27 2017-03-21 Blinker, Inc. Method and apparatus for receiving car parts data from an image
CN104599226B (en) * 2015-02-14 2017-05-10 安徽大学 Large-capacity steganography method
US10019737B2 (en) * 2015-04-06 2018-07-10 Lewis Beach Image processing device and method
US10033926B2 (en) 2015-11-06 2018-07-24 Google Llc Depth camera based image stabilization
CN105282400B (en) * 2015-11-20 2018-07-13 北京理工大学 A kind of efficient video antihunt means based on geometry interpolation
US10679477B2 (en) 2016-05-09 2020-06-09 Herbert S Kobayashi Multicamera video alarm system for remote monitoring and method
US11386759B2 (en) 2016-05-09 2022-07-12 Herbert S Kobayashi Three level detector signal for multicamera video alarm system for remote monitoring and method
KR101818129B1 (en) * 2017-04-25 2018-01-12 동국대학교 산학협력단 Device and method for pedestrian recognition using convolutional neural network
WO2019071594A1 (en) 2017-10-09 2019-04-18 华为技术有限公司 Display processing method and electronic device
US11526970B2 (en) * 2019-09-04 2022-12-13 Samsung Electronics Co., Ltd System and method for video processing with enhanced temporal consistency
JP7209653B2 (en) * 2020-02-05 2023-01-20 株式会社ソニー・インタラクティブエンタテインメント Information processing device, information processing method, and program
CN111312127B (en) * 2020-02-24 2023-07-28 北京京东方光电科技有限公司 Display picture adjusting method and device of rotary display screen and rotary display screen
US11620967B2 (en) * 2020-03-16 2023-04-04 Lewis S. Beach Image processing device and method
CN113534281A (en) * 2020-04-14 2021-10-22 深圳市博利凌科技有限公司 Scanner and method for sensing position of hidden object behind surface of object

Family Cites Families (108)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3652790A (en) * 1970-01-28 1972-03-28 Itt Search and tracking television system and method
US3891798A (en) * 1971-03-19 1975-06-24 Rockwell International Corp Tracker unit
US3952151A (en) 1973-08-13 1976-04-20 Trw Inc. Method and apparatus for stabilized reproduction of remotely-sensed images
FR2302004A5 (en) * 1973-08-31 1976-09-17 Thomson Csf AUTOMATIC TARGET TRACKING SYSTEM, PROCESSED BY VIDEO CONTRAST ANALYSIS
US3988533A (en) * 1974-09-30 1976-10-26 Video Tek, Inc. Video-type universal motion and intrusion detection system
US4120004A (en) * 1976-01-26 1978-10-10 Coutta John M Surveillance system
US4405940A (en) * 1977-05-31 1983-09-20 Westinghouse Electric Corp. Apparatus and method for preprocessing video frame signals
NL7801727A (en) * 1978-02-16 1979-08-20 Hollandse Signaalapparaten Bv THRESHOLD SELECTION CIRCUIT.
EP0028933A3 (en) 1979-11-09 1981-06-03 Ascotts Ltd. Surveillance system
US4603430A (en) * 1984-09-21 1986-07-29 Hughes Aircraft Company Target discrimination utilizing median filters
JPS61166289A (en) * 1985-01-18 1986-07-26 Hitachi Ltd Picture transmission system
JP2528789B2 (en) * 1985-06-26 1996-08-28 中央電子 株式会社 Video information management device
EP0242618B1 (en) * 1986-03-25 1994-11-02 Nippon Oil And Fats Company, Limited Transparent optical article and process for preparing same
DE3634414C2 (en) * 1986-10-09 1994-12-08 Thomson Brandt Gmbh TV camera with a target
JPH0695008B2 (en) 1987-12-11 1994-11-24 株式会社東芝 Monitoring device
KR920010034B1 (en) * 1988-03-31 1992-11-13 Toshiba Co Ltd Image track display apparatus
GB2220319B (en) 1988-07-01 1992-11-04 Plessey Co Plc Improvements in or relating to image stabilisation
JP2563567B2 (en) * 1989-03-20 1996-12-11 松下電器産業株式会社 Shake correction device
KR910004009A (en) * 1989-07-27 1991-02-28 강진구 Video camera automatic shooting device
US5175694A (en) * 1990-02-08 1992-12-29 The United States Of America As Represented By The Secretary Of The Navy Centroid target tracking system utilizing parallel processing of digital data patterns
KR100204101B1 (en) * 1990-03-02 1999-06-15 가나이 쓰도무 Image processing apparatus
US5030984A (en) 1990-07-19 1991-07-09 Eastman Kodak Company Method and associated apparatus for minimizing the effects of motion in the recording of an image
JP2863818B2 (en) * 1990-08-31 1999-03-03 工業技術院長 Moving image change point detection method
US5243418A (en) * 1990-11-27 1993-09-07 Kabushiki Kaisha Toshiba Display monitoring system for detecting and tracking an intruder in a monitor area
DE69215733T2 (en) * 1991-04-12 1997-04-10 Victor Company Of Japan Device for detecting the relative movement between contents of successive fields of a video signal
EP0520765B1 (en) * 1991-06-25 1999-05-12 Canon Kabushiki Kaisha Movement vector detecting method/apparatus and encoding method/apparatus using such method/apparatus
US5164827A (en) * 1991-08-22 1992-11-17 Sensormatic Electronics Corporation Surveillance system with master camera control of slave cameras
US5657402A (en) * 1991-11-01 1997-08-12 Massachusetts Institute Of Technology Method of creating a high resolution still image using a plurality of images and apparatus for practice of the method
US6469746B1 (en) * 1992-12-28 2002-10-22 Sanyo Electric Co., Ltd. Multi-vision screen adapter
CA2148231C (en) 1993-01-29 1999-01-12 Michael Haysom Bianchi Automatic tracking camera control system
US5473369A (en) * 1993-02-25 1995-12-05 Sony Corporation Object tracking apparatus
JP3734829B2 (en) * 1993-06-04 2006-01-11 サーノフ・コーポレーション Electronic image stabilization system and method
JPH0773308A (en) * 1993-09-03 1995-03-17 Matsushita Electric Ind Co Ltd Digital image processor
JP3123587B2 (en) 1994-03-09 2001-01-15 日本電信電話株式会社 Moving object region extraction method using background subtraction
CA2125300C (en) * 1994-05-11 1999-10-12 Douglas J. Ballantyne Method and apparatus for the electronic distribution of medical information and patient services
DE4417128A1 (en) 1994-05-16 1995-12-14 Elnic Gmbh Closed-circuit video surveillance system with image processing
US5635982A (en) 1994-06-27 1997-06-03 Zhang; Hong J. System for automatic video segmentation and key frame extraction for video sequences having both sharp and gradual transitions
US20020186874A1 (en) * 1994-09-07 2002-12-12 Jeffrey H. Price Method and means for image segmentation in fluorescence scanning cytometry
JP2902966B2 (en) * 1994-12-16 1999-06-07 三洋電機株式会社 Camera shake correction device and video camera using the same
KR960028217A (en) 1994-12-22 1996-07-22 엘리 웨이스 Motion Detection Camera System and Method
JPH08186760A (en) * 1994-12-28 1996-07-16 Philips Japan Ltd Image shake correcting device
JP3268953B2 (en) 1995-02-27 2002-03-25 三洋電機株式会社 Tracking area setting device, motion vector detection circuit, and subject tracking device using the same
US5821985A (en) * 1995-02-28 1998-10-13 Nec Corporation Multi-point videoconference system having a fixed control station for data transfer
JP3892059B2 (en) 1995-03-07 2007-03-14 松下電器産業株式会社 Moving body tracking device
US5973733A (en) * 1995-05-31 1999-10-26 Texas Instruments Incorporated Video stabilization system and method
DE69628282T2 (en) * 1995-09-15 2004-03-11 Interval Research Corp., Palo Alto METHOD FOR COMPRESSING SEVERAL VIDEO IMAGES
US6018562A (en) * 1995-11-13 2000-01-25 The United States Of America As Represented By The Secretary Of The Army Apparatus and method for automatic recognition of concealed objects using multiple energy computed tomography
KR100188116B1 (en) * 1995-12-28 1999-06-01 김광호 Image stabilization circuit
US6108576A (en) * 1996-03-18 2000-08-22 The Research Foundation Of City College Of New York Time-resolved diffusion tomographic 2D and 3D imaging in highly scattering turbid media
CA2173677C (en) 1996-04-09 2001-02-20 Benoit Sevigny Processing image data
US5940139A (en) * 1996-08-07 1999-08-17 Bell Communications Research, Inc. Background extraction in a video picture
US5953055A (en) 1996-08-08 1999-09-14 Ncr Corporation System and method for detecting and analyzing a queue
US5915038A (en) * 1996-08-26 1999-06-22 Philips Electronics North America Corporation Using index keys extracted from JPEG-compressed images for image retrieval
US5974111A (en) * 1996-09-24 1999-10-26 Vivid Technologies, Inc. Identifying explosives or other contraband by employing transmitted or scattered X-rays
US5751378A (en) 1996-09-27 1998-05-12 General Instrument Corporation Scene change detector for digital video
US5822542A (en) 1996-10-31 1998-10-13 Sensormatic Electronics Corporation Electronic and structural components of an intelligent video information management apparatus
GB2321358A (en) 1997-01-20 1998-07-22 O E M Project Management Ltd Video security system coupled to a fax modem
JP3317176B2 (en) * 1997-02-06 2002-08-26 松下電器産業株式会社 Video transmission equipment
US6360022B1 (en) * 1997-04-04 2002-03-19 Sarnoff Corporation Method and apparatus for assessing the visibility of differences between two signal sequences
WO1998056176A1 (en) * 1997-06-03 1998-12-10 Koninklijke Philips Electronics N.V. Navigating through television programs
JP4240554B2 (en) * 1997-07-11 2009-03-18 ソニー株式会社 Image encoding apparatus, image encoding method, image decoding apparatus, and image decoding method
US6128108A (en) * 1997-09-03 2000-10-03 Mgi Software Corporation Method and system for compositing images
JPH1185654A (en) * 1997-09-12 1999-03-30 Matsushita Electric Ind Co Ltd Virtual www server device and camera controllable www server device
GB2329541B (en) * 1997-09-17 2002-05-29 Sony Uk Ltd Security control system
US5973723A (en) * 1997-12-12 1999-10-26 Deluca; Michael Joseph Selective commercial detector and eliminator apparatus and method
US6363380B1 (en) * 1998-01-13 2002-03-26 U.S. Philips Corporation Multimedia computer system with story segmentation capability and operating program therefor including finite automation video parser
US6151336A (en) * 1998-02-11 2000-11-21 Sorrento Networks, Inc. Time division multiplexing expansion subsystem
US6847737B1 (en) * 1998-03-13 2005-01-25 University Of Houston System Methods for performing DAF data filtering and padding
US6795112B1 (en) * 1998-03-13 2004-09-21 General Instrument Corp. Composite video multiplexing scheme
US6211913B1 (en) * 1998-03-23 2001-04-03 Sarnoff Corporation Apparatus and method for removing blank areas from real-time stabilized images by inserting background information
GB9813632D0 (en) 1998-06-24 1998-08-26 Robinson Alan W Outside televised broadcast shadow compensator
US6493041B1 (en) * 1998-06-30 2002-12-10 Sun Microsystems, Inc. Method and apparatus for the detection of motion in video
US6714909B1 (en) * 1998-08-13 2004-03-30 At&T Corp. System and method for automated multimedia content indexing and retrieval
US6459822B1 (en) * 1998-08-26 2002-10-01 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Video image stabilization and registration
US6741656B1 (en) * 1998-12-16 2004-05-25 Matsushita Electric Industrial Co., Ltd. Image communication apparatus
CN1116649C (en) * 1998-12-23 2003-07-30 皇家菲利浦电子有限公司 Personalized video classification and retrieval system
KR20040018395A (en) * 1999-01-29 2004-03-03 미쓰비시덴키 가부시키가이샤 Method of image feature encoding and method of image feature decoding
DE60037485T2 (en) 1999-01-29 2008-12-04 Sony Corp. Signal processing method and video signal processor for determining and analyzing a picture and / or audio pattern
KR100442991B1 (en) * 1999-02-01 2004-08-04 주식회사 팬택앤큐리텔 Searching device using moving picture index descripter and method thereof
AU759849B2 (en) 1999-03-16 2003-05-01 Central Research Laboratories Limited Closed circuit television (CCTV) camera and system
FR2794880B1 (en) * 1999-06-10 2001-12-14 Philippe Crochat AUTOMATIC PROCESS FOR TRACKING A MOVING TARGET BY AN ELECTRONIC CAMERA AND DEVICE FOR IMPLEMENTING IT
US6476858B1 (en) * 1999-08-12 2002-11-05 Innovation Institute Video monitoring and security system
US6694064B1 (en) * 1999-11-19 2004-02-17 Positive Systems, Inc. Digital aerial image mosaic method and apparatus
US6683995B2 (en) * 1999-12-23 2004-01-27 Eastman Kodak Company Method and apparatus for correcting large defects in digital images
US6461061B2 (en) * 1999-12-30 2002-10-08 Applied Science Fiction, Inc. System and method for digital film development using visible light
US6943920B2 (en) * 2000-02-03 2005-09-13 Eastman Kodak Company Method, system, and software for signal processing using pyramidal decomposition
US6599036B2 (en) * 2000-02-03 2003-07-29 Applied Science Fiction, Inc. Film processing solution cartridge and method for developing and digitizing film
US6786655B2 (en) * 2000-02-03 2004-09-07 Eastman Kodak Company Method and system for self-service film processing
JP2001243477A (en) * 2000-02-29 2001-09-07 Toshiba Corp Device for analysis of traffic volume by dynamic image
ATE423392T1 (en) 2000-03-10 2009-03-15 Infotech Ag METHOD AND DEVICE FOR ADJUSTING A COMPONENT ON A SUBSTRATE USING DIGITAL FEATURE SEPARATION
US20040125877A1 (en) 2000-07-17 2004-07-01 Shin-Fu Chang Method and system for indexing and content-based adaptive streaming of digital video content
DE20104329U1 (en) 2001-03-14 2001-05-31 Vtq Videotronik Gmbh Monitoring device with sensors
TWI236294B (en) * 2001-05-11 2005-07-11 Ulead Systems Inc Method and device for capturing digital video
US20020168091A1 (en) * 2001-05-11 2002-11-14 Miroslav Trajkovic Motion detection via image alignment
US6525658B2 (en) * 2001-06-11 2003-02-25 Ensco, Inc. Method and device for event detection utilizing data from a multiplicity of sensor sources
US6805501B2 (en) * 2001-07-16 2004-10-19 Eastman Kodak Company System and method for digital film development using visible light
WO2003024112A1 (en) * 2001-09-07 2003-03-20 Intergraph Hardware Technologies Company Method, device and computer program product for demultiplexing of video images
US20050007452A1 (en) * 2001-09-07 2005-01-13 Mckay Therman Ward Video analyzer
US6831590B1 (en) * 2001-09-12 2004-12-14 Cyterra Corporation Concealed object detection
US6816622B2 (en) * 2001-10-18 2004-11-09 Microsoft Corporation Generating resized images using ripple free image filtering
US6944331B2 (en) * 2001-10-26 2005-09-13 National Instruments Corporation Locating regions in a target image using color matching, luminance pattern matching and hue plane pattern matching
EP1316933B1 (en) 2001-11-28 2006-08-09 Matsushita Electric Industrial Co., Ltd. Home security system
US6999613B2 (en) 2001-12-28 2006-02-14 Koninklijke Philips Electronics N.V. Video monitoring and surveillance systems capable of handling asynchronously multiplexed video
EP1586202B1 (en) 2003-01-23 2009-03-18 Intergraph Software Technologies Company Video content parser with scene change detector
WO2004114677A1 (en) * 2003-06-23 2004-12-29 Sony Corporation Image processing method and device, and program
US7840067B2 (en) * 2003-10-24 2010-11-23 Arcsoft, Inc. Color matching and color correction for images forming a panoramic image
US7587086B2 (en) * 2004-06-04 2009-09-08 Microsoft Corporation Identifying selected pixels in a digital image
WO2006023863A1 (en) * 2004-08-23 2006-03-02 Intergraph Software Technologies Company Real-time image stabilization

Also Published As

Publication number Publication date
US8233044B2 (en) 2012-07-31
MXPA04002209A (en) 2005-02-17
US6654049B2 (en) 2003-11-25
MXPA04002210A (en) 2005-02-17
CO5580855A2 (en) 2005-11-30
HK1073038A1 (en) 2005-09-16
MXPA04002211A (en) 2005-02-17
CA2459821A1 (en) 2003-03-20
US20030048947A1 (en) 2003-03-13
US7436437B2 (en) 2008-10-14
WO2003024095A1 (en) 2003-03-20
CA2459732C (en) 2017-07-11
CO5650190A2 (en) 2006-06-30
IL160758A0 (en) 2004-08-31
KR100888095B1 (en) 2009-03-11
BRPI0212377B1 (en) 2015-10-27
CN100559873C (en) 2009-11-11
CA2459821C (en) 2011-01-11
EP1428378B1 (en) 2013-03-20
CO5570718A2 (en) 2005-10-31
US20030048282A1 (en) 2003-03-13
US7477797B2 (en) 2009-01-13
RU2004110722A (en) 2005-09-27
NZ531975A (en) 2007-11-30
KR20040048408A (en) 2004-06-09
ZA200401972B (en) 2004-09-15
CA2459823C (en) 2012-04-03
IL160759A (en) 2010-04-15
CA2459823A1 (en) 2003-03-20
HK1074726A1 (en) 2005-11-18
AU2002335713B2 (en) 2008-08-28
AU2002336445B2 (en) 2007-11-01
BR0212375A (en) 2006-03-01
EP1428378A1 (en) 2004-06-16
EP1430444A1 (en) 2004-06-23
RU2004110723A (en) 2005-09-27
CN1593061A (en) 2005-03-09
IL160760A0 (en) 2004-08-31
KR20040054679A (en) 2004-06-25
NZ531974A (en) 2005-08-26
WO2003024112A1 (en) 2003-03-20
KR100919340B1 (en) 2009-09-25
BR0212546A (en) 2006-05-23
US20080095437A1 (en) 2008-04-24
CN1554073A (en) 2004-12-08
ZA200401970B (en) 2007-01-31
CN1299240C (en) 2007-02-07
KR100927918B1 (en) 2009-11-19
WO2003030103A1 (en) 2003-04-10
BRPI0212375B1 (en) 2016-05-24
CN100359923C (en) 2008-01-02
RU2308761C2 (en) 2007-10-20
HK1069470A1 (en) 2005-05-20
IL160759A0 (en) 2004-08-31
BR0212377A (en) 2006-10-03
US7310110B2 (en) 2007-12-18
EP1428389A1 (en) 2004-06-16
US20040061786A1 (en) 2004-04-01
NZ531976A (en) 2007-07-27
US7079701B2 (en) 2006-07-18
CN101072364A (en) 2007-11-14
KR20040047814A (en) 2004-06-05
US20030048359A1 (en) 2003-03-13
CN1554184A (en) 2004-12-08
US20060215926A1 (en) 2006-09-28
WO2003030103A8 (en) 2005-01-06
RU2004110724A (en) 2005-09-27
RU2308816C2 (en) 2007-10-20
RU2310289C2 (en) 2007-11-10
AU2002327612B2 (en) 2008-01-17
CN101072364B (en) 2012-05-09
EP1430444B1 (en) 2013-05-08

Similar Documents

Publication Publication Date Title
CA2459732C (en) Concealed object recognition
AU2002335713A1 (en) Concealed object recognition
US11092507B2 (en) Method of indicating gas movement in a scene
US9996913B2 (en) Contrast based image fusion
CN107295248B (en) Image display apparatus and image display method for displaying image, and storage medium
US7006688B2 (en) Histogram adjustment features for use in imaging technologies
US8515196B1 (en) Systems and methods for processing infrared images
EP1758058A1 (en) System and method for contrast enhancing an image
CN101142812A (en) Image processing device and method, program and recording medium
US7130463B1 (en) Zoomed histogram display for a digital camera
WO2008102296A2 (en) Method for enhancing the depth sensation of an image
JP4359662B2 (en) Color image exposure compensation method
JP2005173879A (en) Fused image display device
EP1976273B1 (en) Image processing apparatus, imaging apparatus and image processing program
US11114045B1 (en) Method of enhancing the visibility of screen images
JP5050141B2 (en) Color image exposure evaluation method
JP4677924B2 (en) Imaging system, moving image processing method, and moving image processing apparatus
JP3912063B2 (en) Detection method of image shading unevenness
JP6270448B2 (en) Image processing apparatus, image processing method, and program
KR100305888B1 (en) Real-time correction device and method of digital image
KR100267997B1 (en) Apparatus and method for scanning conversion
JP2005176171A (en) Device for correcting luminance
JPH11297259A (en) Display control device with shading correcting function
JP2004234133A (en) Image processor

Legal Events

Date Code Title Description
EEER Examination request
MKLA Lapsed

Effective date: 20190906