Search Images Maps Play YouTube News Gmail Drive More »
Sign in
Screen reader users: click this link for accessible mode. Accessible mode has the same essential features but works better with your reader.

Patents

  1. Advanced Patent Search
Publication numberUS20060017829 A1
Publication typeApplication
Application numberUS 10/895,129
Publication dateJan 26, 2006
Filing dateJul 21, 2004
Priority dateJul 21, 2004
Also published asCN101023680A, EP1772019A2, US7880785, WO2006014641A2, WO2006014641A3
Publication number10895129, 895129, US 2006/0017829 A1, US 2006/017829 A1, US 20060017829 A1, US 20060017829A1, US 2006017829 A1, US 2006017829A1, US-A1-20060017829, US-A1-2006017829, US2006/0017829A1, US2006/017829A1, US20060017829 A1, US20060017829A1, US2006017829 A1, US2006017829A1
InventorsPaul Gallagher
Original AssigneeGallagher Paul K
Export CitationBiBTeX, EndNote, RefMan
External Links: USPTO, USPTO Assignment, Espacenet
Rod and cone response sensor
US 20060017829 A1
Abstract
An imager having color sensitivity in any lighting condition is provided by replacing at least one of the green color filtered pixels in a Bayer pattern array with an non-color filtered pixel and providing a different integration period for each color channel. When replacing a color filtered pixel with a non-color filtered pixel, the benefits of both color sensitivity and light sensitivity may be harnessed by providing the color filtered pixels with a longer integration period than the non-color filtered pixels. Color information may be interpolated by performing subtraction logic using non-color and color filter information from adjacent pixels in back-end processing. Integration times for each color channel can be adapted to minimize filter transmission and sensor absorption differences within each frame of data. Temporal displacement of colors of moving objects by having different integration periods may be corrected in back-end processing. Back-end processing may also correct motion blur.
Images(10)
Previous page
Next page
Claims(48)
1. A pixel array comprising:
a plurality of pixels, each comprising a photosensor for producing photo-generated charges during an integration period, wherein a first group of said plurality of pixels are color filtered and a second group of said plurality of pixels are non-color filtered.
2. The pixel array of claim 1, wherein the pixels of said second group have a shorter charge integration period than pixels of said first group.
3. The pixel array of claim 1, wherein said non-color filtered pixels are also not filtered for infrared wavelengths.
4. The pixel array of claim 1, wherein said non-color filtered pixels are filtered for infrared wavelengths.
5. The pixel array of claim 1, wherein said plurality of pixels are grouped with each group having more than one pixel, at least one of the pixels being non-color filtered.
6. The pixel array of claim 5, wherein each group comprises one each of a red, blue, green, and non-color filter.
7. The pixel array of claim 5, wherein each group of pixels further comprises one pixel each having respective associated red, green, and blue color filters.
8. The pixel array of claim 5, wherein each group comprises a red, blue, and two non-color filtered pixels.
9. The pixel array of claim 5, wherein each group of four pixel cells comprises one each of a pixel cell having cyan, magenta, and yellow color filters.
10. The pixel array of claim 5, wherein each group of four pixel cells shares one readout circuit.
11. The pixel array of claim 1, wherein pixels of said first group have a different integration from the pixels of said second group.
12. The pixel array of claim 7, wherein said non-color filtered pixel has the shortest integration period, said red color filtered pixel has a longer integration period than said non-color filtered pixel, said green color filtered pixel has a longer integration period than said red color filtered pixels, and said blue color filtered pixel has the longest integration period of said group of four pixels.
13. The pixel array of claim 8, wherein said red and blue color filtered pixels have a first integration period and said two non-color filtered pixels have a second integration period, wherein said first integration period is longer than said second integration period.
14. The pixel array of claim 5, wherein said group of four pixels comprises two pixels on each of two rows.
15. A pixel array comprising a plurality of pixel cells comprising:
a photosensor for producing photo generated charges during a plurality of integration periods, wherein a first group of said plurality of pixels are red color filtered, a second group of said plurality of pixels are blue color filtered, and a third group of said plurality of pixels are non-color filtered.
16. The pixel array of claim 15, wherein said first group has a first integration period, said second group has a second integration period, and said third group has a third integration period.
17. The pixel array of claim 16, wherein said first and second integration periods have the same duration, which is different from said third integration period.
18. The pixel array of claim 17, wherein said first and second integration periods are longer than said third integration period.
19. The pixel array of claim 16, wherein said first, second, and third integration periods have different durations.
20. The pixel array of claim 15, wherein non-color filtered pixels are also not filtered for infrared wavelengths.
21. The pixel array of claim 15, wherein said plurality of pixels are grouped with each group having one red color filtered pixel, one blue color filtered pixel, and one non-color filtered pixel.
22. The pixel array of claim 19, further comprising a fourth group of group of pixels that are green color filtered having a fourth integration period with a different duration from said first, second, and third integration periods.
23. The pixel array of claim 22, wherein said third integration period has the shortest duration.
24. The pixel array of claim 21, wherein each group shares a common readout circuit.
25. The pixel array of claim 23, wherein pixels are grouped in four-pixel sets with each set having four pixels, one each of said first, second, third, and fourth group of pixels, each set sharing a common readout circuit.
26. An imaging device comprising:
a pixel array comprising a plurality of pixels, each comprising a photosensor for producing photo-generated charges during an integration period, wherein a first group of said plurality of pixels has a first integration period and a second group of said plurality of pixels has a second integration period;
a timing and control unit for controlling the integration periods of said pixels; and
an image processor for processing signals read out by said readout circuitry.
27. The imaging device of claim 26, wherein said first integration period is shorter than said second integration period.
28. The imaging device of claim 27, wherein said first group of said plurality of pixels are non-color filtered and said second group of said plurality of pixels are color filtered.
29. The imaging device of claim 28, further comprising at least a third group of said plurality of color filtered pixels having a third integration period.
30. The imaging device of claim 29, wherein said third group of pixels have a different color filter than said second group of pixels.
31. The imaging device of claim 27, wherein said image processor corrects temporal displacement of colors due to the different integration periods.
32. The imaging device of claim 27, wherein said image processor corrects motion blur by interpolating color data based on data from said first group of pixels having a shorter integration period.
33. A method of operating a pixel array comprising:
generating charges with a first plurality of photosensors during a first integration period;
generating charges with a second plurality of photosensors during a second integration period;
transferring charges from said first plurality of photosensors to readout circuitry at the end of said first integration period;
transferring charges from said second plurality of photosensors to readout circuitry at the end of said second integration period, wherein said first and second integration periods partially overlap.
34. The method of claim 33, wherein said first integration period begins before said second integration period.
35. The method of claim 33, wherein said first integration period ends at a different time than said second integration period.
36. The method of claim 35 further comprising:
generating charges with a third plurality of photosensors during a third integration period;
generating charges with a fourth plurality of photosensors during a fourth integration period;
transferring charges from said third plurality of photosensors to readout circuitry at the end of said third integration period; and
transferring charges from said fourth plurality of photosensors to readout circuitry at the end of said fourth integration period.
37. The method of claim 36, wherein said first, second, third, and fourth integration periods partially overlap.
38. The method of claim 37, wherein said first, second, third, and fourth integration periods end at different times.
39. The method of claim 33, wherein one of said plurality of photosensors does not have a color filter or near-infrared filter over it.
40. The method of claim 39, wherein said unfiltered photosensors are integrated during the shortest integration period.
41. A method of operating an imaging device comprising:
generating charges with a first plurality of photosensors in a pixel array during a first integration period;
transferring charges from said first plurality of photosensors to readout circuitry at the end of said first integration period;
generating charges with a second plurality of photosensors in a pixel array during a second integration period;
transferring charges from said second plurality of photosensors to readout circuitry at the end of said second integration period;
controlling the integration periods of said pixel array such that said integration periods end at different times; and
processing signals read out by said readout circuitry.
42. The method of claim 41, wherein said integration periods partially overlap.
43. The method of claim 41, wherein said first integration period is longer than said second integration period.
44. The method of claim 41, wherein said step of generating charges with said second plurality of photosensors comprises generating charges from non-color filtered incident light.
45. The method of claim 44, wherein said step of generating charges with said first plurality of photosensors comprises generating charges from color filtered incident light.
46. The method of claim 45, wherein said step of processing signals includes interpolating color information by subtracting signals from said first integration period out of signals from said second integration period.
47. The method of claim 45, wherein said step of processing signals corrects temporal displacement of colors due to the different integration periods.
48. The method of claim 45, wherein said step of processing signals corrects motion blur.
Description
FIELD OF THE INVENTION

This invention relates to imager devices and, in particular, to improved light sensitivity in pixel arrays.

BACKGROUND OF THE INVENTION

Humans perceive light energy of varying wavelengths as color, detected by two types of light-sensitive receptors in the retina of the eye, rod cells and cone cells. The rod cells detect only the quantity of light, working at even low light levels with fewer photons, also known as night-vision. They are stimulated by the intensity of light and are responsible for perceiving the size, shape, and brightness of visual images, but do not perceive color and fine detail. The cone cells, of which there are three types, each capable of detecting a different range of wavelength of light being received, require hundreds of photons to activate them. They are less sensitive to low illumination levels but provide color information. Of the three types of cones, each one contains a distinctive type of pigment, absorbing red light, green light, and blue light.

By contrast, digital imagers comprise an array of pixel cells, each one of the pixel cells including a photoconversion device, e.g., a photodiode gate, photoconductor, or a photodiode, for converting light into an electrical charge. In a CMOS imager, a readout circuit is connected to each pixel cell, which typically includes a source follower output transistor. The photoconversion device converts photons to electrons which are typically transferred to a floating diffusion region connected to the gate of the source follower output transistor. A charge transfer device (e.g., transistor) can be included for transferring charge from the photoconversion device to the floating diffusion region. In addition, such imager cells typically have a transistor for resetting the floating diffusion region to a predetermined charge level prior to charge transference. The output of the source follower transistor is gated as an output signal by a row select transistor.

In color imagers, the pixel cells also have a color filter over the surface of the sensor, which limits the specific wavelengths of light that may be permitted to enter the photoconversion device. A Bayer pattern filter of alternatively red and green filters in a first row, and green and blue filters in a second row is most commonly used in a typical 2×2 square of pixel cells, which is repeated for the entire array, as illustrated in FIG. 1. There is an emphasis on green filters due to human visual response, which reaches a maximum sensitivity in the green wavelength region (550 nm) of the visible spectrum. Hence, when processed, the green data provides not only chrominance information, but as its peak response is close to the peak response of the human eye, it is also used for luminance information.

Color filtered pixels, like cone cells, require a greater number of photons, relative to unfiltered pixel cells or rod cells, in order to produce an appreciable signal. This is largely due to the color filters themselves, which diminish transmissivity. Color imagers must also have a near-infrared (NIR) blocking filter in the optical path to assure that NIR energy is not absorbed by the color filtered pixels. However, in low light conditions, not only does the imager suffer from the color filter transmissive losses, but it also cannot take advantage of the NIR energy present without an additional mechanism to remove the NIR filter from the pixel array.

A monochromatic imager is able to take advantage of the NIR energy. The pixel cells in a monochromatic imager have neither color nor NIR blocking filters, and are therefore more sensitive to lower levels of incident light, including NIR energy. However, monochromatic imagers are incapable of capturing color information of images.

There is a desire and need for a color imager that can provide better images in low light conditions.

BRIEF SUMMARY OF THE INVENTION

Exemplary embodiments of the present invention provide an imager having color sensitivity and low light response sensitivity by replacing at least one of the color filtered pixels in a Bayer pattern array with a pixel without any color filter. A different integration period for the non-color pixel compared with the color filtered pixels may also be provided. For example, one or both of the green color filtered pixels of a Bayer pattern array may be replaced by an unfiltered pixel.

Since the photoconversion device of the unfiltered pixel, such as a photodiode, will likely saturate before any of the color filtered pixels, exemplary embodiments of the invention provide the unfiltered channels with a shorter integration period than those of the colored channels.

These and other features of the invention will be more clearly understood from the following detailed description which is provided in connection with the accompanying drawings.

BRIEF DESCRIPTION OF THE DRAWINGS

FIG. 1 depicts a pixel array of prior art;

FIG. 2 depicts a pixel array according to one embodiment of the present invention;

FIG. 3 depicts a pixel array according to another embodiment of the present invention;

FIG. 4 is a block diagram of an imaging device;

FIG. 5 is a schematic diagram of a four-transistor (4T) pixel cell of prior art;

FIG. 6 is a schematic diagram of a four-pixel cell group according to the embodiment of FIG. 2;

FIG. 7 is a timing diagram illustrating the operation of the pixel cell of FIG. 6;

FIG. 8 shows a camera device incorporating at least one imager device constructed in accordance with an embodiment of the invention; and

FIG. 9 shows a processor system incorporating at least one imager device constructed in accordance with an embodiment of the invention.

DETAILED DESCRIPTION OF THE INVENTION

In the following detailed description, reference is made to the accompanying drawings which form a part hereof, and in which is shown by way of illustration specific embodiments in which the invention may be practiced. These embodiments are described in sufficient detail to enable those skilled in the art to practice the invention, and it is to be understood that other embodiments may be utilized, and that structural, logical and electrical changes may be made without departing from the spirit and scope of the present invention.

The term “substrate” is to be understood as a semiconductor-based material including silicon, silicon-on-insulator (SOI) or silicon-on-sapphire (SOS) technology, doped and undoped semiconductors, epitaxial layers of silicon supported by a base semiconductor foundation, and other semiconductor structures. Furthermore, when reference is made to a “substrate” in the following description, previous process steps may have been utilized to form regions or junctions in or over the base semiconductor structure or foundation. In addition, the semiconductor need not be silicon-based, but could be based on silicon-germanium, germanium, or gallium arsenide.

The term “pixel” refers to a picture element unit cell containing a photosensor and transistors for converting light radiation to an electrical signal. For purposes of illustration, a representative pixel is illustrated in the figures and description herein and, typically, fabrication of all pixels in an imager will proceed simultaneously in a similar fashion. The following detailed description is, therefore, not to be taken in a limiting sense, and the scope of the present invention is defined by the appended claims.

Referring now to the drawings, where like elements are designated by like reference numerals, FIG. 4 illustrates a schematic diagram of a conventional four-transistor (4T) pixel cell. Each of the conventional pixels comprises a photosensor 26, for example, a photogate, photoconductor or a photodiode overlying a substrate for producing a photo-generated charge in a doped region of the substrate. A readout circuit is provided for each pixel cell and includes at least a source follower transistor 34 and a row select transistor 36 for coupling the output of the source follower transistor 34 to a column output line. The pixel cell also typically has a floating diffusion node 28, connected to the gate of the source follower transistor 34. Charge generated by the photosensor 26 is sent to the floating diffusion region 26. The imager may also include a transfer gate 30 for transferring charge from the photosensor 26 to the floating diffusion node 27 and a reset transistor 32 for resetting the floating diffusion region node to a predetermined charge level prior to charge transference.

FIG. 5 illustrates a block diagram of a CMOS imager device 108 having a pixel array 100 with each pixel cell being constructed as described above. The pixel array 100 comprises a plurality of pixels arranged in a predetermined number of columns and rows. The pixels of each row in the array 100 are all turned on at the same time by a row select line, and the pixels of each column are selectively output by respective column select lines. A plurality of row and column lines are provided for the entire array 100. The row lines are selectively activated by a row driver 145 in response to a row address decoder 155. The column select lines are selectively activated by a column driver 160 in response to a column address decoder 170. Thus, a row and column address is provided for each pixel.

The CMOS imager 100 is operated by a control circuit 150 that controls address decoders 155, 170 for selecting the appropriate row and column lines for pixel readout, and row and column driver circuitry 145, 160 that apply driving voltage to the drive transistors of the selected row and column lines. The pixel column signals, which typically include a pixel reset signal Vrst and a pixel image signal Vsig for each row selected pixel in a column are read by sample and hold circuitry 161 associated with the column device 160. A differential signal Vrst−Vsig is produced for each pixel, which is amplified and digitized by analog-to-digital converter 175. The analog-to-digital converter 175 converts the analog pixel signals to digital signals that are fed to an image processor 180 to form a digital image output.

Color pixel arrays are often comprised of groups of four pixels having different color filters over them, as in a Bayer pattern, in order to capture light information from the visible spectrum. A portion of a conventional Bayer pattern pixel array is depicted in FIG. 1. In a Bayer pattern, each group of four pixels has one pixel with a red color filter, one pixel with a blue color filter, and two pixels with green color filters. When processed, the green data provides not only chrominance information, but as its peak response is close to the peak response of the human eye, it is also used for luminance information. Other color filter patterns may contain cyan, magenta, and yellow color filters, which will also capture light information from the visible spectrum.

According to one exemplary embodiment of the present invention, at least one of the pixels with a color filter in a Bayer pattern array may be replaced with a pixel without any color filter, as shown in FIG. 2. In the embodiment illustrated, one of the two green color filtered pixels is replaced by a pixel without a color filter. A red color filtered pixel AA and a green color filtered pixel BB share a first row and an unfiltered pixel CC and a blue color filtered pixel DD share a second row, such that the red color filtered pixel AA shares a first column with unfiltered pixel CC and the green color filtered pixel BB shares a second column with blue color filtered pixel DD.

Because the color filters diminish transmissivity, the color filtered pixels in such an array will saturate slower than the unfiltered pixels. Accordingly, the non-color filtered pixel may be provided with a shorter integration time than the color-filtered pixels. Additionally, when a pixel is provided to capture light intensity, size, shape and brightness information, rather than color, it may be possible to take advantage of near-infrared (NIR) energy. As such, when replacing one of the color filtered pixels with a non-color filtered pixel, a NIR filter may also be eliminated from the non-pixel in order to capture light information from the non-visible spectrum. Therefore, there will be three pixels capturing light from the visible spectrum and one pixel capturing light from the visible and non-visible spectrum.

In one exemplary embodiment, the integration times for readout of each pixel may be coordinated by the timing and control circuit 150, addressing appropriate row and column lines for pixel readout after the end of each pixel's integration time.

In another exemplary embodiment, each group of four pixels AA-DD in FIG. 2 shares one readout circuit. The readout circuit typically comprises floating diffusion node, reset gate, source follower transistor, and row select transistor. As shown in FIG. 6, pixel AA has a photosensor 26 a and transfer gate 30 a, pixel BB has a photosensor 26 b and transfer gate 30 b, pixel CC has a photosensor 26 c and transfer gate 30 c, and pixel DD has a photosensor 26 d and transfer gate 30 d. Unlike the pixel cells of prior art, each of these pixels AA-DD do not have their own readout circuits. Rather, all four pixels share a floating diffusion node 28′, reset gate 32′, source follower transistor 34′, and row select transistor 36′ in, for example, a shared-row, shared-column architecture.

Once the charges are generated by the photosensors 26 a-d, respective transfer gates 30 a-d transfer the charges one at a time to the floating diffusion node and are applied to the readout circuit one at a time. FIG. 7 illustrates exemplary charge integration and transfer gate timing diagrams for the pixel cells AA-DD according to the present invention. The signals in FIG. 7 can be provided by timing and control circuitry similar to circuitry 150 in FIG. 5. The timing diagram of FIG. 7 begins at a point in time of readout of previously accumulated charges in the pixels AA, BB, CC, and DD.

At T1, the row select transistor 36′ (RS) is turned on. Each of the pixels AA-DD have accumulated charges that are sampled individually for readout. The order of readout can be in any order desired. For illustration, FIG. 7 shows readout in the order of pixels CC, AA, DD, and BB. At time T2, the reset transistor 32′ (RST) is pulsed and the sample and hold reset circuit (SHR) takes a reset signal sample Vrst-CC at time T3. Then the non-color filtered pixel CC is sampled after its transfer gate 30 c (TGU) is pulsed at time T4, followed by a pulse of the sample and hold signal circuit (SHS) taking the output signal Vsig-CC at time T5. The reset transistor 32′ (RST) is pulsed again at T6 to begin the cycle for sampling the red color filtered pixel AA (TGR). Sample and hold reset circuit (SHR) and sample and hold signal circuit (SHS) are also pulsed similarly in the process of sampling the reset level Vrst-AA and signal level Vsig-AA of the red color filtered pixel AA. The cycle repeats for the DD and BB pixels, each time pulsing a different transfer gate, until all pixels AA-DD have been sampled, in terms of a reset signal Vrst and pixel output signal Vsig. At time T7, the row select transistor 36′ (RS) is turned off as another row gets sampled. It should be noted that in this architecture, two rows of pixels are actually read out, with the red and green color filtered pixels being in a first row and the non-color filtered and blue color filtered pixels being in a second row. After these two rows are read, the next two rows are read. Also in this architecture, two adjacent pixels in a row share a column line.

At time T8, all of the transfer gates and the reset transistor are turned on to dump any residual photodiode charges and reset the circuit in preparation for charge integration. Each pixel then begins its own charge integration period when the transfer gate turns off, i.e., TGU, TGR, TGB, TGG go low. The integration period ends when the transfer gate of a pixel turns on again to transfer charge to the floating diffusion region. The beginning of each pixel's integration period is controlled by the timing and control for the pixel array, which may be hard coded based on the transmissivity of each color filter or dynamically controlled for different lighting conditions.

Thus, in FIG. 7, the blue color filtered pixel DD starts its integration period first, at time T9, when the transfer gate 30 d (TGB) is closed. The blue color filtered pixel DD starts its integration period at time T9, when the transfer gate 30 b (TGB) is closed. At time T1, the red color filtered pixel AA starts its integration period when the transfer gate 30 a (TGR) is closed, the non-color filtered pixel CC starts its integration period at time T11 when the transfer gate 30 c (TGU) is closed, and the green color filtered pixel BB starts its integration period at time T12 when the transfer gate 30 b (TGG) is closed. After T12, all of the pixels AA-DD are integrating. That is, they are all collecting incident light and converting the photons into charge and storing them in their respective photosensors 26 a-d. At time T13, the row select 36′ (RS) is turned on, to begin the readout period for the pixels which proceeds as discussed above with respect to time periods T1 to T7.

As shown in FIG. 7, the charge integration period of pixel AA is (T10-T15), the charge integration period of pixel BB is (T12-T17), the charge integration period of pixel CC is (T11-T14), and the charge integration period of pixel DD is (T9-T16). The charge integration periods may be set based on the sensitivity of the pixel and its ability to capture light given the transmissivity of the color filter over it. In the example described above, blue color filtered pixels tend to be less sensitive to light and therefore require a longer integration time to accumulate charge. Therefore, a blue color filtered pixel DD may have the longest charge integration period. Similarly, an non-color filtered pixel CC may have a shorter charge integration time, since non-color filtered pixels saturate faster. It is also possible to set all the color filtered pixels to the same charge integration period, while setting the non-color filtered pixel to the shortest integration time. It should be noted that the order in which the pixels are sampled is not limited to the order described above. The timing of operation of the four pixel group, as described above, is only an illustration of how the integration and readout periods can be arranged.

For simplicity, the circuit shown in FIG. 6 illustrates one group of four pixels. However, this four-pixel group would be repeated throughout a pixel array 100 in FIG. 5. By replacing a color filtered pixel with an non-color filtered pixel throughout the array, the benefits of both color sensitivity and light sensitivity may be achieved simultaneously in the pixel array.

FIG. 3 illustrates another embodiment of the present invention, where both green color filtered pixels are replaced by an non-color filtered pixel. Therefore, the 2×2 pixel group will have a red color filtered pixel and an non-color filtered pixel in the first row and an non-color filtered pixel and a blue color filtered pixel in the second row. Because two pixels in the four-pixel grouping are color filtered pixels and two pixels are non-color filter, there may effectively be two different integration times: one short integration time for the non-color filtered pixels and one longer integration time for the color filtered pixels. However, individual charge integration times can also be set for each pixel, if desired.

Since, in this embodiment, there is no green color filtered pixel, the image processor 180 of FIG. 5 will perform a subtraction logic function to determine what portion of charge from the non-color filtered pixels was from the green spectrum, based on the charge collected from adjacent red and blue color filtered pixels.

It should be noted that the four-pixel groups of this embodiment (FIG. 3) would be repeated throughout a pixel array 100 in FIG. 5.

Prior art techniques for color response balance involves applying electronic gain to weaker output signals. However, when electronic gain is applied, noise is amplified along with the signal, which is not desirable. According to the present invention, the color response can be balanced, by varying the integration times for each color channel, to correct for the differences in filter transmissivity and silicon absorption by wavelength within the pixel, rather than by applying electronic gain.

In the exemplary embodiments of the present invention, color channels may be sampled separately, with each having its own charge integration period. When pixel arrays have various integration times for groups of pixels within the array, there is a potential problem with temporal displacement of the colors of an object moving in the scene. This may also be corrected for in back-end processing, for example, in the image processor 180 of FIG. 5. Using the same back-end processing as for color temporal displacement, motion blur may also be corrected since the non-color filtered pixels have a very short integration time with respect to the other pixels, providing a frame of reference for interpolating from earlier moments in the motion.

As noted, four-pixel groups as described above may be repeated and used in a pixel array 100 of the imager device 108 illustrated in FIG. 5. FIG. 8 shows a camera device 300, a typical processor-based camera device modified to include an imager device 108 (FIG. 5) employing pixels of the present invention as an input device to the device 300. The imager device 108 may also receive control or other data from camera device 300 as well.

Camera device 300 includes a central processing unit (CPU) 302 that communicates with various devices over a bus 304. Some of the devices connected to the bus 304 provide communication into and out of the system 300, illustratively including an input/output (I/O) device 306 and imager device 108. Other devices connected to the bus 304 provide memory, illustratively including a random access memory system (RAM) 310, and a peripheral memory device such as FLASH or hard drive memory 312.

The camera device 300 may also be included in a processor-based camera system 400, as illustrated in FIG. 9. Examples of processor-based camera systems 400 which may employ the camera device 300, include, without limitation, computer systems, camera systems, scanners, machine vision systems, vehicle navigation systems, video telephones, surveillance systems, auto focus systems, star tracker systems, motion detection systems, image stabilization systems, and others.

System 400 includes a central processing unit (CPU) 402 that communicates with various devices over a bus 404. Some of the devices connected to the bus 404 provide communication into and out of the system 400, illustratively including an input/output (I/O) device 406 and camera device 300. Other devices connected to the bus 404 provide memory, illustratively including a random access memory system (RAM) 410, FLASH or hard drive 412, and one or more peripheral memory devices such as a floppy disk drive 414 and compact disk (CD) drive 416. The camera device 300 may be combined with a processor, such as a CPU, digital signal processor, or microprocessor, in a single integrated circuit.

In another embodiment (not illustrated), it may be desirable to provide the pixel array of the present invention on an SOC imager device such that the processing may be performed on the imager.

The above description and drawings are only to be considered illustrative of exemplary embodiments which achieve the features and advantages of the invention. Modification of, and substitutions to, specific process conditions and structures can be made without departing from the spirit and scope of the invention. Accordingly, the invention is not to be considered as being limited by the foregoing description and drawings, but is only limited by the scope of the appended claims.

Referenced by
Citing PatentFiling datePublication dateApplicantTitle
US7468504 *Mar 9, 2006Dec 23, 2008Northrop Grumman CorporationSpectral filter for optical sensor
US7531373Sep 19, 2007May 12, 2009Micron Technology, Inc.Methods of forming a conductive interconnect in a pixel of an imager and in other integrated circuitry
US7602430 *Apr 18, 2007Oct 13, 2009Foveon, Inc.High-gain multicolor pixel sensor with reset noise cancellation
US7663679 *Mar 5, 2007Feb 16, 2010Fujifilm CorporationImaging apparatus using interpolation and color signal(s) to synthesize luminance
US7728895 *Jul 11, 2006Jun 1, 2010Panasonic CorporationSolid-state image sensing device having shared floating diffusion portions
US7741210Mar 31, 2009Jun 22, 2010Aptina Imaging CorporationMethods of forming a conductive interconnect in a pixel of an imager and in other integrated circuitry
US7773137 *Jun 13, 2007Aug 10, 2010Fujitsu Semiconductor LimitedImaging apparatus, imaging element, and image processing method
US7773138 *Sep 12, 2007Aug 10, 2010Tower Semiconductor Ltd.Color pattern and pixel level binning for APS image sensor using 2×2 photodiode sharing scheme
US7864233May 23, 2007Jan 4, 2011Samsung Electronics Co., Ltd.Image photographing device and method
US7876956 *Nov 10, 2006Jan 25, 2011Eastman Kodak CompanyNoise reduction of panchromatic and color image
US7978240 *Oct 2, 2006Jul 12, 2011Konica Minolta Photo Imaging, Inc.Enhancing image quality imaging unit and image sensor
US8045024 *Apr 15, 2009Oct 25, 2011Omnivision Technologies, Inc.Producing full-color image with reduced motion blur
US8068153 *Mar 27, 2009Nov 29, 2011Omnivision Technologies, Inc.Producing full-color image using CFA image
US8068159 *Apr 18, 2007Nov 29, 2011Samsung Electronics Co., Ltd.Image capturing device and method with infrared detection thorugh plural filter regions
US8111307Oct 25, 2008Feb 7, 2012Omnivision Technologies, Inc.Defective color and panchromatic CFA image
US8119435Nov 5, 2010Feb 21, 2012Omnivision Technologies, Inc.Wafer level processing for backside illuminated image sensors
US8125546Jun 5, 2009Feb 28, 2012Omnivision Technologies, Inc.Color filter array pattern having four-channels
US8164651 *Apr 29, 2008Apr 24, 2012Omnivision Technologies, Inc.Concentric exposure sequence for image sensor
US8203633May 27, 2009Jun 19, 2012Omnivision Technologies, Inc.Four-channel color filter array pattern
US8224082Mar 10, 2009Jul 17, 2012Omnivision Technologies, Inc.CFA image with synthetic panchromatic image
US8224085Jan 3, 2011Jul 17, 2012Omnivision Technologies, Inc.Noise reduced color image using panchromatic image
US8237831May 28, 2009Aug 7, 2012Omnivision Technologies, Inc.Four-channel color filter array interpolation
US8253832Jun 9, 2009Aug 28, 2012Omnivision Technologies, Inc.Interpolation for four-channel color filter array
US8432466 *Sep 29, 2011Apr 30, 2013International Business Machines CorporationMultiple image high dynamic range imaging from a single sensor array
US8456555 *Nov 16, 2007Jun 4, 2013Samsung Electronics Co., Ltd.Method and apparatus for improving dynamic range of image
US8514322Jun 16, 2010Aug 20, 2013Aptina Imaging CorporationSystems and methods for adaptive control and dynamic range extension of image sensors
US20110032395 *Oct 19, 2010Feb 10, 2011Konica Minolta Photo Imaging, Inc.Imaging unit and image sensor
WO2008049342A1 *May 30, 2007May 2, 2008Kar Ki ChanImage sensor and method of using the same
WO2013090922A1 *Dec 17, 2012Jun 20, 2013Tenebraex CorporationSystems and methods for creating full-color image in low light
WO2013183886A1 *May 28, 2013Dec 12, 2013Samsung Techwin Co., LtdCamera system with multi-spectral filter array and image processing method thereof
Classifications
U.S. Classification348/276, 348/281, 382/300, 348/E09.01, 358/525
International ClassificationH04N9/07, H04N5/3745, H04N5/353, H04N1/46, H04N3/14, H04N5/225, H04N9/083, H04N9/04
Cooperative ClassificationH04N5/332, H04N9/045
European ClassificationH04N5/33D, H04N9/04B
Legal Events
DateCodeEventDescription
Aug 28, 2009ASAssignment
Owner name: APTINA IMAGING CORPORATION, CAYMAN ISLANDS
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICRON TECHNOLOGY, INC.;REEL/FRAME:023245/0186
Effective date: 20080926
Owner name: APTINA IMAGING CORPORATION,CAYMAN ISLANDS
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICRON TECHNOLOGY, INC.;US-ASSIGNMENT DATABASE UPDATED:20100304;REEL/FRAME:23245/186
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICRON TECHNOLOGY, INC.;US-ASSIGNMENT DATABASE UPDATED:20100225;REEL/FRAME:23245/186
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICRON TECHNOLOGY, INC.;US-ASSIGNMENT DATABASE UPDATED:20100318;REEL/FRAME:23245/186
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICRON TECHNOLOGY, INC.;US-ASSIGNMENT DATABASE UPDATED:20100323;REEL/FRAME:23245/186
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICRON TECHNOLOGY, INC.;US-ASSIGNMENT DATABASE UPDATED:20100325;REEL/FRAME:23245/186
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICRON TECHNOLOGY, INC.;US-ASSIGNMENT DATABASE UPDATED:20100329;REEL/FRAME:23245/186
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICRON TECHNOLOGY, INC.;US-ASSIGNMENT DATABASE UPDATED:20100413;REEL/FRAME:23245/186
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICRON TECHNOLOGY, INC.;US-ASSIGNMENT DATABASE UPDATED:20100420;REEL/FRAME:23245/186
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICRON TECHNOLOGY, INC.;US-ASSIGNMENT DATABASE UPDATED:20100427;REEL/FRAME:23245/186
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICRON TECHNOLOGY, INC.;US-ASSIGNMENT DATABASE UPDATED:20100504;REEL/FRAME:23245/186
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICRON TECHNOLOGY, INC.;REEL/FRAME:23245/186
Jul 21, 2004ASAssignment
Owner name: MICRON TECHNOLOGY, INC., IDAHO
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:GALLAGHER, PAUL K.;REEL/FRAME:015610/0200
Effective date: 20040719