WO1996020454A1 - Method and apparatus for reading images without need for self-generated illumination source - Google Patents

Method and apparatus for reading images without need for self-generated illumination source Download PDF

Info

Publication number
WO1996020454A1
WO1996020454A1 PCT/US1995/016828 US9516828W WO9620454A1 WO 1996020454 A1 WO1996020454 A1 WO 1996020454A1 US 9516828 W US9516828 W US 9516828W WO 9620454 A1 WO9620454 A1 WO 9620454A1
Authority
WO
WIPO (PCT)
Prior art keywords
bar code
lens
ccd
signal
code reader
Prior art date
Application number
PCT/US1995/016828
Other languages
French (fr)
Inventor
Bryan L. Olmstead
Michael J. Ahten
Bruce E. Paris
Jorge L. Acosta
James W. Ring
Paul R. Huss
Jon P. C. Williams
Alexander M. Mcqueen
Randy L. Person
Original Assignee
Spectra-Physics Scanning Systems, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US08/363,258 external-priority patent/US5770847A/en
Application filed by Spectra-Physics Scanning Systems, Inc. filed Critical Spectra-Physics Scanning Systems, Inc.
Priority to JP8520561A priority Critical patent/JPH09512372A/en
Priority to DE19581524T priority patent/DE19581524B4/en
Priority to GB9617173A priority patent/GB2301691B/en
Publication of WO1996020454A1 publication Critical patent/WO1996020454A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K7/00Methods or arrangements for sensing record carriers, e.g. for reading patterns
    • G06K7/10Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
    • G06K7/10544Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum
    • G06K7/10821Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum further details of bar or optical code scanning devices
    • G06K7/10851Circuits for pulse shaping, amplifying, eliminating noise signals, checking the function of the sensing device
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K7/00Methods or arrangements for sensing record carriers, e.g. for reading patterns
    • G06K7/10Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
    • G06K7/10544Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum
    • G06K7/10554Moving beam scanning
    • G06K7/10594Beam path
    • G06K7/10683Arrangement of fixed elements
    • G06K7/10702Particularities of propagating elements, e.g. lenses, mirrors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K7/00Methods or arrangements for sensing record carriers, e.g. for reading patterns
    • G06K7/10Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
    • G06K7/10544Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum
    • G06K7/10712Fixed beam scanning
    • G06K7/10722Photodetector array or CCD scanning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K7/00Methods or arrangements for sensing record carriers, e.g. for reading patterns
    • G06K7/10Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
    • G06K7/10544Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum
    • G06K7/10712Fixed beam scanning
    • G06K7/10722Photodetector array or CCD scanning
    • G06K7/10752Exposure time control
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K7/00Methods or arrangements for sensing record carriers, e.g. for reading patterns
    • G06K7/10Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
    • G06K7/10544Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum
    • G06K7/10792Special measures in relation to the object to be scanned
    • G06K7/10801Multidistance reading
    • G06K7/10811Focalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K7/00Methods or arrangements for sensing record carriers, e.g. for reading patterns
    • G06K7/10Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
    • G06K7/10544Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum
    • G06K7/10821Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum further details of bar or optical code scanning devices
    • G06K7/10881Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum further details of bar or optical code scanning devices constructional details of hand-held scanners

Definitions

  • the field of the present invention relates to image detection and, in particular, to a method and apparatus for reading images such as bar code labels while reducing or eliminating the need for a self-generated illumination source.
  • Labels bearing information in any of a number of different encoded formats are commonly affixed to products, packaging, or other items and are used in many applications. It is common, for example, to encode retail product identification information in a bar code format on a product package or a label affixed to a product package. Bar code symbols are also used on a broad range of retail packages for check-out and inventory purposes.
  • a bar code reading device located at the check-out station of a retail establishment, may be used by a clerk to read product identification data into an associated point-of- sale computer system.
  • a number of existing bar codes consist of a series of parallel light and dark rectangular areas of varying widths. The light areas are often referred to as "spaces" and the dark areas as "bars”.
  • a bar code label may be read by a device (e.g., a scanner) which detects reflected and/or refracted light from the bars and spaces comprising the characters.
  • a device e.g., a scanner
  • Such a device typically includes a light source for illuminat- ing the bar code label to be read.
  • One common method of illuminating the bar code label is by use of a scanning laser beam, in which case a spot of light is swept across the bar code label and the intensity of returned light is detected by an optical detector.
  • the optical detector generates an electrical signal having an amplitude deter ⁇ mined by the intensity of the detected light.
  • Another method for illuminating the bar code label is by use of a uniform light source with an array of optical detectors connected to an analog shift register (commonly called a charge-coupled device or CCD) .
  • a uniform light source with an array of optical detectors connected to an analog shift register (commonly called a charge-coupled device or CCD) .
  • CCD charge-coupled device
  • an electrical signal is generated having an amplitude determined by the intensity of the collected light.
  • the amplitude of the electrical input signal has one population of levels for dark bars and a second population of levels for light spaces.
  • positive-going transitions and negative-going transitions in the electrical signal occur, signifying transitions between bars and spaces.
  • the electrical signal may be converted to a binary scan signal which is analyzed to determine the arrangement of bars and spaces of the scanned label.
  • the bar and space information is provided to a decoding unit to determine whether the bar code is recognized and, if so, to decode the information contained in the bar code.
  • An essential element of virtually every existing bar code reader system is a means for illuminating the bar code.
  • Laser based scanners typically use a flying laser spot to illuminate the bar code and detect the returned light.
  • Wand devices generally use a light emitting diode (LED) to illuminate a small portion of the bar code and a detector to image the bar code while the operator physically sweeps the illuminated beam across the bar code.
  • Video based readers such as readers utilizing charge-coupled devices to receive input, typically use a plurality of LED's to illuminate the entire bar code with uniform intensity, and then image the entire bar code with a CCD sensor.
  • Each of the above systems relies on self-generated illumination for the purpose of distinguishing the bars and spaces of the bar code or for increasing the amount of collected light to enable reasonable performance.
  • the requirement of self-generated illumination may, however, lead to additional complexity, a larger or bulkier device, and/or higher power consumption.
  • a scanning device using a flying laser spot may require a patterned mirror structure, bifocal lens, laser light source, and motor to sweep the laser spot across a target.
  • the motor used for sweeping the laser spot tends to be one of the least reliable compo- nents of a scanner, followed by the laser illumination source.
  • the use of a motor and laser illumination source tends to make a system less reliable and more costly.
  • the requirement of a self-generated illumination generally makes an image reading system more complex, costly and unreliable.
  • CCD readers often have significant depth of field limitations. For example, CCD readers using red LED's for illumination typically have a depth of field of less than one inch. This limits the types of applications in which CCD readers may be used, and limits the versatil ⁇ ity of such CCD readers in general. Moreover, due to low sweep speeds, use of CCD readers has also generally been limited to handheld devices.
  • the invention provides ir ⁇ various aspects a method and apparatus for reading bar codes or other images while reducing or eliminating the need for self-generated illumination.
  • a preferred embodiment of the invention utilizes a CCD based imaging system.
  • the imaging system comprises an adaptive exposure circuit which is used to prevent satura ⁇ tion of a CCD detector at high illumination levels, thereby increasing the system's dynamic range.
  • the imag ⁇ ing system further comprises an efficient optical collec ⁇ tion system, which may include an aperture with a high aspect ratio and/or a multi-plane imaging technique.
  • the imaging system thereby may have an increased depth of field at a particular light level, or an increased read rate, or both.
  • a correlated double sampling circuit is preferably used in conjunction with the CCD detector to allow lower light levels to be detected by the system.
  • the correlated double sampling circuit may also provide an increased depth of field, an increased read rate, or both.
  • the imaging system further takes advantage of derivative based signal processing for enhanced edge detection.
  • Various other embodiments may utilize some but not all of the above elements, or may include additional refinements, while obtaining the benefit of reducing or eliminating the need for self-generated illumination.
  • Fig. 1 is diagram showing a perspective partial cut ⁇ away view of a preferred handheld bar code reader using ambient light.
  • Fig. 1A is a block diagram of an optical reading system in accordance with one or more aspects of the present invention.
  • Fig. 2 is a diagram of a CCD detector receiving a focused beam of light.
  • Fig. 3A is a potential diagram illustrating the processes of charge accumulation and charge transfer from a photodiode to a CCD shift register location.
  • Fig. 3B is a physical illustration of a charge coupled device having five potential wells.
  • Fig. 3C is a potential diagram illustrating a transfer of charge along a CCD shift register.
  • Fig. 4 is a diagram of a simplified output stage of a CCD detector.
  • Fig. 5 is a diagram of a bar code and a corresponding waveform after filtering.
  • Fig. 6 is a circuit diagram of a driver circuit for driving a CCD chip.
  • Fig. 7 is a block diagram of a multiple-CCD detector system.
  • Fig. 8A is a block diagram of a correlated double sampling circuit
  • Fig. 8B is a graph of waveform diagrams associated with the sampling circuit of Fig. 8A.
  • Fig. 9 is a circuit diagram of a CCD output stage.
  • Fig. 10A is a block diagram of a dual sampler processor
  • Fig. 10B is a graph showing timing diagrams for the dual sampler processor of Fig. 10A
  • Fig. 11A is a circuit block diagram of a dual slope processor
  • Fig. 11B is a graph showing timing diagrams for the dual slope processor of Fig. 11A.
  • Fig. 12 is a graph of transfer functions for the dual sampler processor of Fig. 10A and the dual slope processor of Fig. 11A.
  • Fig. 13A is a diagram of an optical system based on a single lens model
  • Fig. 13B is a diagram of a single lens model showing creation of a blur spot when the target image is moved away from the focal point.
  • Figs. 14A-14D are diagrams of various types of aberrations and distortions that may be associated with an optical system.
  • Fig. 15 is a graph illustrating a cosine law effect.
  • Fig. 16 is a diagram illustrating light collection.
  • Fig. 17 is a diagram comparing the point spread function shapes through focus for a round aperture and an oblong/elliptical aperture.
  • Fig. 18 is a diagram illustrating change in elliptical spot size with respect to tilt angle.
  • Fig. 19 is a graph showing illustrative depth of field plots for a round aperture and for an elliptical aperture collecting the same amount of light.
  • Fig. 20A is a diagram of an unapertured lens interposed between a target and an imager
  • Fig. 2OB is a diagram of an apertured lens in a similar relative position.
  • Figs. 21A and 21B are diagrams showing an oblique view and a front view, respectively, of a symmetric ulti- focal lens.
  • Figs. 22A-22C are diagrams showing various views of a multi-focal lens system combined with an elliptical aperture.
  • Figs. 23A, 23B and 23C are graphs showing a minimum renderable bar width as a function of distance from a single-lens, a poly-optic lens with no interference, and a poly-optic lens with inter-zonal interference, respectively.
  • Figs. 24A and 24B are diagrams of point spread functions for a center zone of the Fig. 21A-21B lens through focus and away from focus, respectively.
  • Figs. 25A and 25B are diagrams of point spread functions for a ring zone of the Fig. 21A-21B lens through focus and away from focus, respectively.
  • Figs. 26A through 30C are approximated point spread function plots of various combinations of center and ring zones of various focal lengths.
  • Figs. 31A-31C are diagrams of a symmetric multi-focus lens shown from several viewpoints with a plurality of apertures.
  • Fig. 32 is a diagram of a cut-away symmetric lens.
  • Fig. 33 is a diagram of an asymmetric lens having its optical axis in the center of the lens.
  • Figs. 34A is a diagram showing off-axis tilt when the center zone focuses to the closest depth and the outside ring zone focuses to the farthest depth
  • Fig. 34B is a diagram showing the reverse situation.
  • Figs. 35A and 35B are diagrams showing different embodiments of a lens system having an array of lenses arranged in a pattern around an imaging device.
  • Figs. 36A-36C are diagrams of various embodiments of a lens system for planar coverage having an array of lenses arranged in a pattern around an imaging device.
  • Figs. 37A-37C are diagrams showing other embodiments of a lens system for planar coverage having an array of lenses arranged in a pattern around an imaging device and including a collimating lens.
  • Figs. 38A-38B are diagrams of alternate embodiments of a poly-optic lens system.
  • Figs. 39A-39B are diagrams of alternate embodiments of an optical system using curved mirrors to collect light. 8
  • Figs. 40A-40D are diagrams of a particular poly-optic lens from different angles illustrating light collection.
  • Figs. 41A-41C are diagrams of a preferred poly-optic lens.
  • Figs. 42A-42C are diagrams of other embodiments of optical systems using shutters to reduce inter-zonal interference.
  • Fig. 43A is a circuit diagram of a preferred adaptive exposure circuit
  • Fig. 43B is a graph of waveform patterns for the circuit of Fig. 43A.
  • Figs. 44A and 44B are block diagrams showing alternative configurations of an optical system employing an adaptive exposure circuit.
  • Fig. 45A is a block diagram of a signal processing circuit for determining transitions in a video signal
  • Fig. 45B is a graph of waveform patterns for the circuit of Fig. 45A.
  • Figs. 46A-46D are diagrams showing various views of a CCD detection module incorporating various aspects of the present invention.
  • Fig. 47 is a diagram of an embodiment of a handheld imager incorporating the CCD detection module of Figs. 46A-46D.
  • Fig. 48 is a diagram of an embodiment of a portable data terminal incorporating the CCD detection module of Figs. 46A-46D.
  • Fig. 49 is a diagram of a multiple-CCD omnidirec ⁇ tional imaging device for reading over a multi-dimensional area.
  • Figs. 50A and SOB are diagrams illustrating the effect on an output signal of rotating a bar code with respect to a CCD array.
  • Fig. 51 is a graph comparing pixel width versus roll angle for a CCD array with square pixels and one with rectangular pixels (having a 4:1 aspect ratio) .
  • Figs. 52A and 52B are diagrams showing an oblique view and a side view, respectively, of a masked CCD array.
  • Fig. 52C is a diagram of a masked CCD array embodi ⁇ ment including an additional lens for reducing the effect of vignetting.
  • Fig. 53 is a diagram of an imaging system including an anamorphic lens system.
  • Figs. 54A and 54B are diagrams of apodized poly-optic lenses for reducing inter-zonal interference.
  • Figs. 55A and 55B are line spread functions associated with the apodized lenses of Figs. 54A and 54B.
  • Figs. 56A and 56B are diagrams illustrating a multi ⁇ element lens system for enhancing depth of field while minimizing interference among lens elements.
  • Fig. 57 is a diagram illustrating a multiple-focus optical system with a poly-optic lens having some of the aberrations of the zone uncorrected so as to narrow the angular field of view for the longer focal length zones.
  • Fig. 58 is a diagram of a multi-directional optical system using cylindrical optics to allow variously oriented images (e.g., bar codes) to be read.
  • Fig. 59 is a diagram of a one-axis filter for determining the orientation of a bar code.
  • Figs. 60A-60C are diagram of a multi-directional optical system using a beam splitter and mirrors, with various alternative area and linear sensor configurations depicted in Figs. 60D-60F.
  • Fig. 61A is a diagram of another embodiment of a multi-directional optical system employing a mirror basket to divide a single projected image and independently rotate the different segmented images
  • Fig. 61B is a particular application of a mirror basket with a handheld bar code reader.
  • Figs. 62A-62F are diagrams associated with another embodiment of a multi-direction reader employing a kaleidoscope mirror for projecting multiple rotated images of a bar code onto a detector array.
  • Fig. 63 is a diagram of an optically transmissive spacer extending forward from a bar code reader housing to allow ambient light to reach a bar code or other image to be read.
  • Fig. 64A is a diagram of an optical diffuser
  • Fig. 64B is a diagram illustrating the effect of a system employing the optical diffuser of Fig. 64A
  • Fig. 64C is a diagram comparing the operation of a system without an optical diffuser.
  • Fig. 65A and 65B are side view and front view diagrams, respectively, of an LED pointer system
  • Fig. 65C is a diagram illustrating various pointer shapes at different distances.
  • Fig. 66 is a diagram of a signal processor circuit including an infinite impulse response high pass filter.
  • Fig. 67 is a diagram of a signal processor circuit including a finite impulse response high pass filter.
  • Figs. 68A and 68B are graphs of exemplary impulse response coefficients for the signal processor of Fig. 67
  • Figs. 69A and 69B are exemplary waveform diagrams also associated with the signal processor of Fig. 67
  • Figs. 70A and 70B are different depth cross-sectional views of the scan head of the bar code reader of Fig. 1.
  • Fig. 70C is a top sectional view of the optical chas ⁇ sis mounted in the scan head shown in Figs. 70A and 70B.
  • Fig. 71A is an electrical block diagram of a pre- ferred electronics and control system for the Fig. 1 bar code reader
  • Figs. 71B through 71D are exemplary waveform diagrams associated with the Fig. 71A electrical block diagram.
  • Fig. 72 is a block diagram of a bar code reader having rescan capabilities.
  • Figs. 73A and 73B are diagrams of a mounted multi- focal lens showing preferred dimensions thereof, and Fig. 73C is a table showing preferred characteristics of the multi-focal lens of Figs. 73A-73B.
  • Figs. 73A and 73B are diagrams of a mounted multi- focal lens showing preferred dimensions thereof
  • Fig. 73C is a table showing preferred characteristics of the multi-focal lens of Figs. 73A-73B.
  • FIG. 1 is a cut-away view of a preferred embodiment of a bar code reader.
  • the bar code reader 1501 depicted in Fig. 1 is a handheld unit with relatively large depth of field that can operate in ambient light by employing a preferred combination of the techniques and components as will be described in detail herein.
  • the handheld bar code reader 1501 comprises a housing 1502 encasing most or all of the optical system and electronics.
  • the bar code reader 1501 further comprises a digital board 1503 on which is mounted an optics chassis 1506, and an analog board 1504 connected to the back of the optics chassis 1506 and also connected to the digital board 1503.
  • a handle board 1505 is also encased in a portion of the housing 1502 forming a handle below the scan head 1522.
  • the scan head 1522 and optics chassis 1506 are shown in more detail in Figs. 70A-70C, along with other optical system and related components, and described in more detail later herein.
  • an imaging lens 1509 such as imaging lens 1749 shown in Figs. 73A-73B.
  • the imaging lens 1509 is prefer ⁇ ably a multi-focus lens providing an enlarged depth of field. Details regarding a variety of suitable multi- focus lenses and other optical schemes, as well as preferred electronics for facilitating operation in ambient light, are described in more detail later herein.
  • FIG. 1A is a block diagram of an image reading system 100 in accordance with one or more aspects of the present invention.
  • the image reading system 100 of Fig. 1A comprises an optical collection system 103, a CCD detector 104, a pre- amplification block 105, a signal processing block 106, and a decoding block 107.
  • the optical collection system 1 03 operates in the presence of ambient illumination 101, gathering ambient light reflecting off a target (such as a bar code label 102) and focusing the gathered light on the CCD detector 104.
  • the CCD detector 104 interacts with the pre-amplification block 105 to obtain a CCD input signal comprised of imaging data.
  • the CCD input signal may be processed by the signal processing block 106 and provided to the decoding block 107 for decoding, as further explained herein.
  • FIG. 2 is a diagram of a CCD detector 110 receiving a focused beam of light 115.
  • Light is reflected from a target located at a target line 114.
  • the reflected light is passed through a lens 113 and focused onto the CCD detector 110.
  • the CCD detector 110 may be embodied as a linear integrated chip comprising an array of photodiodes 112. In operation, the CCD detector 110 acts in a manner similar to a video camera, with the array of photodiodes
  • 113 generally define the depth of field and the minimum bar code size for an imager incorporating such a CCD detector 110.
  • Linear CCD detectors have four basic elements. These four basic elements relate to photon to electron conver ⁇ sion, charge storage, charge transfer, and charge to voltage conversion. Photon to electron conversion is per ⁇ formed by multiple photodiodes (e.g., pixels 112). Typical photodiodes have a quantum efficiency of 50%, which means that one electron charge is generated for every two photons that hit the photodiode. The electron charges are stored in a potential well, which acts like a capacitor. When an exposure period is complete, a read out gate pulse received from a controller allows the accumulated charges to be dumped into a CCD shift regis ⁇ ter. A potential diagram illustrating the processes of charge accumulation and charge transfer from a photodiode to a CCD shift register location is shown in Fig. 3A.
  • the charge coupled device may be embodied as a distributed gate metal-oxide silicon field-effect transistor (MOSFET) 130, a physical illustration of which is shown in Fig. 3B.
  • the distributed gate MOSFET 130 has a long source-drain channel 131 and contains a plurality of gates 132 (one gate 132 per pixel 112) .
  • Charge is injected at each gate location (e.g., shift register loca ⁇ tion) by a photodiode, by the process depicted in Fig. 3A.
  • Charge is transferred along a chain of shift register locations by controlling the gate voltages.
  • FIG. 3B A potential diagram illus ⁇ trating the described charge transfer process is shown in Fig. 3C.
  • an amplifier converts the charge to voltage. This conversion is typic ⁇ ally accomplished by allowing a capacitor to integrate the charge and then buffering the voltage to an output pin. A MOSFET is used to clear the charge from the capacitor between transfers of each pixel . The conversion process may also involve amplification of the converted voltage level.
  • the output voltage generally has a DC bias level equal to about half of the power supply voltage. The higher the input light intensity, the lower the DC bias voltage, as the photoelectric charge is in the form of injected electrons.
  • FIG. 4 is a diagram of a simplified output stage of a CCD detector 110 such as shown in Fig. 2.
  • a flow of electrons 150 is accumulated by an integration capacitor 152 (also referred to a sense capacitor) .
  • a transistor 154 such as a MOSFET, clears the stored charge between each pixel.
  • the MOSFET 154 may be referred to as a reset transistor or a reset MOSFET.
  • the MOSFET 154 is connected to a reference voltage source 156.
  • the voltage level of the sense capacitor 152 may be read through a buffer stage 155, which also may comprise a MOSFET.
  • the buffer MOSFET 155 is connected at one end to a positive bias voltage 160, at the other end through a current sink 158 to a negative bias voltage 159.
  • the current sink 158 comprises a resistor
  • the negative bias voltage 159 is set to ground level.
  • the CCD detector 110 provides data in the form of a discrete time analog signal.
  • the analog signal appears similar to "stair-steps", with data from each pixel contributing to one of the steps.
  • Most CCD chips have an output signal that is in a low state when the image is bright (white) and a high state when the image is dark
  • the signal output from the CCD detector 110 may be low pass filtered to arrive at a smooth signal, such as shown in Fig. 5.
  • a smooth signal such as shown in Fig. 5.
  • FIG. 5 an exemplary portion of a bar code is shown having darker bars 170 and lighter spaces 171.
  • a corresponding CCD output signal 172 is shown, after low pass filtering, above the bar code portion.
  • the filtered CCD output signal 172 is shown in a high state when the darker bars 170 are read, and a low state when the lighter spaces 171 are read.
  • the CCD output signal 172 may have a sizable DC component (e.g., 4.5 volts).
  • the amplitude of the CCD output signal 172 depends mainly on the sensitivity of the photodiode array and the exposure time.
  • the photo- detectors in the CCD detector 110 collect charge. Periodically, the accumulated charge is dumped into an analog shift register. The information in the analog shift register is serially read out to form an analog CCD signal which, when filtered, corresponds to the CCD output signal 172. All of the accumulated data must be transferred from the photodetectors and read out between exposures, so that residual charge does not distort the data from the next exposure period. If it is required to have an exposure time faster than the data read out time, a shuttered CCD device may be used, thereby allowing the exposure time to fill some fraction of the read out time.
  • the CCD detector 110 may particularly be embodied as a CCD chip of the ILX series as manufactured by Sony, such as an ILX-503A chip (having 2048 pixels) or ILX-505A chip
  • the Sony ILX-703A may be used in applications requiring an electronic shutter.
  • FIG. 6 is a circuit diagram of a driver circuit suitable for driving an ILX-503 or ILX-703 CCD chip 204.
  • the Fig. 6 driver circuit generates a pixel clock signal ( ⁇ CLK) 201 and a read out gate clock signal ( ⁇ ROG) 202 to facilitate transfer of charge from the photodetectors to the analog shift register.
  • the pixel clock signal 201 is generated by a resonator oscillator circuit 203 as depicted in Fig. 6.
  • the read out gate clock signal 202 is generated by a read out gate clock generator circuit 205 as also depicted in Fig. 6.
  • the read out gate clock generator circuit 205 comprises an oscillator circuit 207 followed by a pair of one-shot circuits. Another one-shot circuit 208 acts to temporarily disable the pixel clock signal 201 when the read out gate clock signal 202 is generated.
  • the CCD chip 204 generates a CCD video output signal 206.
  • the CCD output signal 206 may, as noted, be provided to a low-pass reconstruction filter and thereafter to other circuitry for signal processing, decoding and other functions as further described herein.
  • the low-pass reconstruction filter may be embodied as part of the pre- amplification block 105 or the signal processing block 106.
  • the low pass filter preferably is constructed so as to block out signals having a frequency higher than one- half the frequency of the pixel clock signal 201.
  • the low pass filter preferably has a steep filtering character ⁇ istic, since a gradual filter may increase the number of pixels per bar required for reading a bar code.
  • the low pass filter may also provide gain, if necessary, depending on the light level and the exposure time.
  • multiple linear CCD arrays may be used for imaging. Sensitivity tends to decrease with larger imagers, making the use of multiple linear CCD arrays preferable in certain applications. Moreover, cost tends to increase geometrically with larger CCD arrays.
  • FIG. 7 is a block diagram of a multiple-CCD detector system, which may be configured, for example, to implement a low-cost multiple scan line system.
  • a plurality of CCD detectors 230 are clocked by a single CCD driver circuit 231, such as shown previously in Fig. 6.
  • the CCD driver circuit 231 provides a pixel clock signal and a read out gate clock signal 233 to each of the CCD detectors 230.
  • the CCD detectors 230 are coupled to a signal processor 232 via a high speed analog multiplexer 235, which selects the CCD detectors 230 in sequence for processing.
  • a counter 237 may be coupled to the CCD driver circuit 231 so as to count clock pulses and switch the high speed analog multiplexer 235 at appropriate times.
  • the signal processor 232 receives a CCD output signal 238 from each CCD detector 230 in sequence, as controlled by the analog multiplexer 235, and thereby processes the CCD output signals 238 repetitively, in a sequential fashion.
  • the CCD detectors 230 may be physically placed in a variety of different patterns so as to generate a desired imaging pattern.
  • the CCD detectors 230 may be completely separate, mounted on a common circuit board, or constructed on the same chip substrate.
  • I d is the dark current (in amps) of the photo- detector
  • B is the bandwidth of the system (in hertz)
  • e is the charge on an electron (i.e., 1.6xl0- 19 C)
  • Dark current I d is a strong function of temperature, doubling every 25°C (in the case of the ILX-503A chip) .
  • the effect of photodetector shot noise becomes more pronounced the longer the photodetector integrates. Thus for longer exposure times, the photodetector shot noise will increase. The effect of this shot noise, however, tends to be offset by the fact that longer exposure leads to more signal.
  • the signal-to-noise ratio due to shot noise is proportional to the square root of the exposure time.
  • a second source of noise is associated with the process of charge transfer. This source of noise results from surface states (i.e., 1/f) and thermal noise gener ⁇ ated by the charge transfer MOSFET.
  • Charge transfer noise manifests as a linear growth of noise from the starting pixel to the ending pixel. While the first pixel to be transferred stays in the charge transfer register for only a short time, the last pixel remains much longer, there ⁇ fore accumulating much more noise than the first pixel. Because charge transfer efficiencies are generally high, this noise is usually not significant enough to degrade performance.
  • a third source of noise is known as kTC noise.
  • This type of noise is caused by a switched capacitor effect related to the reset transistor at the end of the CCD chain. A similar effect may also occur in switched capacitor filters.
  • This source of noise is typically the dominant noise source in the CCD chip, but can be miti ⁇ gated or removed by other techniques discussed herein, such as correlated double sampling.
  • FIG. 9B A graph of kTC noise power as a function of frequency is shown in Fig. 9B for a CCD output stage shown in Fig.
  • the CCD output stage in Fig. 9A corresponds to that of Fig. 4, except that the reset MOSFET 154 in Fig. 4 is modelled by its internal resistances in an OFF state
  • one graph 260 shows kTC noise power when the reset MOSFET
  • MOSFET 154 is turned on and thus has a resistance of R on .
  • a fourth source of noise results from the charge-to-voltage conversion process.
  • the charge-to-voltage ampli ⁇ bomb is subject to surface states (i.e., l/f) and thermal noise of the FET, as well as thermal noise due to other resistances.
  • This noise source dominates after the kTC noise and is not affected by exposure time. This noise source has only a weak temperature dependance.
  • Figure 8A is a block diagram of a particular type of correlated double sampling circuit such as may be employed to reduce kTC noise.
  • Figure 8B is a diagram of waveforms illustrating some of the noise reduction properties of the circuit of Fig. 8A.
  • a correlated double sampling circuit such as shown in Fig. 8A may advantageously be employed to remove reset noise generated by the periodic resetting of the sense node by the reset MOSFET in the CCD output stage (shown, e.g., in Fig. 9A) , thereby improving signal-to-noise ratio.
  • reset noise may be expressed in terms of the following relationship:
  • the correlated double sampling circuit generally operates by sampling the CCD output voltage once after the reset pulse has been switched off and once again after signal charge has been read out, thereby providing a direct measurement of charge. This charge measurement is generally independent of reset noise, since the reset noise will be "correlated" over the duration of a pixel period. In other words, the noise voltage due to the channel resistance will not change over the period when the two samples are taken, allowing the reset noise to be eliminated externally from the CCD chip. As shown in Fig.
  • an input signal 240 (such as output from CCD chip 204 of Fig. 6) is buffered by an amplifier 241.
  • the output of the amplifier 241 is con ⁇ nected to a DC restoration circuit 239 comprising an analog switch 243, a capacitor 242, and a bias voltage 244 (which may be ground) , configured as shown.
  • a clamp sig ⁇ nal 255 controls the analog switch 243.
  • the DC restora ⁇ tion circuit 239 is connected to another buffer 245.
  • Buffer 245 is connected to another analog switch 246.
  • the analog switch 246 is controlled by a sample signal 259.
  • the other end of the analog switch 246 is connected to a capacitor 247 and to another buffer 248. Buffer 248 outputs an output signal 249.
  • the input signal 240 may appear, for example, like the illustrative input waveform 250 of Fig. 8B.
  • the wave- forms of Fig. 8B have been idealized somewhat for purposes of explanation.
  • the heavier shaded lines in the sample input waveform 250 illustrate relatively low frequency noise 257 appearing in the input signal 240.
  • a feed- through pulse 254 occurs in the input waveform 250 when the reset MOSFET is active. During the interval of the feedthrough pulse 254, relatively high bandwidth noise may occur. However, once the feedthrough pulse 254 is ended and the reset MOSFET is inactive, a generally low fre ⁇ quency, constant noise pattern will be expected. This low frequency noise 257 occurring in the interval 255 between feedthrough pulses 254 may be removed or substantially reduced by the correlation operation of the circuit of Fig. 8A.
  • the input signal 240 is applied to the input buffer 241.
  • the clamp signal 255 (shown as waveform 251 in Fig. 8B) is active, the capacitor 242 is coupled to the bias voltage 244 through the analog switch 243. The capacitor 242 is thereby charged to the reset voltage level 256, as shown in relation to waveform 250 of Fig. 8B.
  • the clamp signal 255 is then deactivitated, releasing the clamp (i.e., opening the analog switch 243) but allowing the capacitor 242 to remain charged at the reset voltage level 256.
  • the input waveform 250 then switches to a level 265 determined by the amount of charge collected by a particu- lar photodiode in the CCD detector.
  • the output of the buffer 241 follows the input waveform 250, which forces the opposite end of the capacitor 242 to a new voltage level, since the capacitor 242 maintains the same voltage across it that it had prior to the changing levels in the input waveform 250.
  • the voltage at the input of the buffer 245 therefore comprises the difference between the voltage level 265 of the input waveform and the reset voltage level 256. Because the reset voltage level 256 of the input waveform 250 contains noise 257 that is cor- related over a pixel period, common mode noise is thereby removed from the output of the DC restoration circuit 239.
  • the voltage level of the capacitor 242 is sampled when the sample signal 259 (shown as waveform 252 in Fig.
  • FIG. 8B An output waveform 253 associated with input waveform 250 is also shown in Fig. 8B.
  • FIG. 10A A more detailed diagram of a particular type of correlated double sampling circuit, known as a dual sampler processor, is shown in Fig. 10A. Timing diagrams for the correlated double sampling circuit of Fig. 10A are shown in Fig. 10B.
  • a CCD signal 270 is connected to a pre-amplification stage 273.
  • a resistance R L represents the output resistance of the CCD detector.
  • the pre-amplification stage 273 is connected to a low pass filter stage 274 comprising an amplifier 275, two resist- ors 276 and 277, and a capacitor 278, configured as shown.
  • the transfer function of the low pass filter stage is:
  • the low pass filter stage 274 is connected to a capacitor 280.
  • Capacitor 280 is connected to a buffer stage 282 and one end of a switch SI. The other end of the switch SI is connected to ground 286.
  • the buffer stage 282 is connected to a second switch S2.
  • the other end of the switch S2 is connected to a capacitor 283 and is buffered by another buffer stage 284.
  • Buffer stage 284 outputs a correlated signal 288.
  • the CCD signal 270 may be divided into a plurality of pixel periods, each of which comprises input from a particular pixel, as noted.
  • Timing of the dual sampling processor may be divided into five non-exclusive time intervals within a pixel period. These time periods may be referred to as (1) reset, (2) reference, (3) clamp, (4) video dump, and (5) sample. Each of these time periods are explained more fully below.
  • the reset period 310 shown with respect to waveform 300 in Fig. 10B is commenced when the reset switch (e.g., reset MOSFET as shown in Fig. 4) is turned on, which starts a pixel readout sequence.
  • the reset switch e.g., reset MOSFET as shown in Fig. 4
  • Waveform 300 in Fig. 10B is an example of a CCD video output signal starting from the time the reset switch is turned on.
  • the sense capacitor 152 in the CCD output stage (e.g., as shown in Fig. 4) is set to the voltage level of the reference voltage source 156, thereby setting a positive charge packet in sense capacitor 152 to be eroded by signal electrons.
  • the time constant associated with this operation is preferably very small.
  • the time constant may be 800 picoseconds, where the sense capacitor 152 has a value of 0.4 pF and an "on" resistance of 2 k ⁇ .
  • the reset pulse width applied to the reset MOSFET 154 to control clearing the charge on the sense capacitor 152 can be made quite small if desired (e.g., a few nanoseconds) .
  • Application of a reset pulse to the reset MOSFET 154 generates a feed-through pulse seen in the CCD signal 270.
  • This feed-through pulse is caused by an effective capaci ⁇ tor divider network comprising the sense capacitor 152 and a gate-to-source capacitance C gs of the reset MOSFET 154.
  • a reference period 312 occurs after the reset switch (i.e., MOSFET 154) is turned off.
  • the voltage level at the sense node e.g., at the sense capacitor 152 quickly decays to the reference potential 156 during this time.
  • the sense node retains the reference potential until signal charge is dumped onto the sense capacitor 152.
  • a clamp period 314 occurs during the reset period 3 1 2.
  • the clamp switch SI is turned on, thereby forcing the voltage level at the input to the first buffer stage 282 to ground level or, alterna ⁇ tively, a non-zero reference level if desired. This action establishes a precise reference level for each pixel before video dump occurs. It also eliminates reset noise introduced during the time the reset switch 154 was on.
  • the clamped level is considered the first "sample" taken by the dual sampler circuit.
  • the clamp is released (i.e., switch SI turned off) at the start of a following video dump period 316.
  • the clamp period 314 follows the video dump period 316.
  • charge from the summing well is transferred onto the sense node (i.e., at the sense capacitor 156) when the output transfer gate becomes active. This causes the sense node to be dis ⁇ charged by an amount proportional to the signal level in that summing well of the CCD.
  • a sample period 318 During the video-dump period 316 occurs a sample period 318.
  • Switch S2 is turned on before video dump to assure that sampling begins when the clamp switch is "on" at ground potential. If sampling were not to start until after video dump, the response would exhibit a slight non-linearity resulting from the finite bandwidth and response time of the sample- and-hold circuit. Because switch S2 is turned on before video dump, sampling begins at the same potential for each pixel.
  • the time period when the clamp is released to when the video is held is referred to as the clamp-to-sample time or sample-to-sample time t s .
  • a preferred timing guideline is to position the reset clock far from the video dump edge due to its large ampli ⁇ tude. This positioning allows sufficient settling time of the reference level before the video signal is sampled.
  • the reset clock feed-through pulse is eliminated as a result of the clamp switch, as the video signal is clamped over the entire pixel period except during the sample-to- sample time t s .
  • the sample-to-sample time t s and electrical bandwidth T D are preferably selected to achieve low noise perform ⁇ ance.
  • the correlated double sampling processor acts as a band-pass filter in that low frequency noise is rejected when t a is made smaller (because these low frequencies become correlated) , while high frequency noise is rejected when the bandwidth is made smaller.
  • the center frequency of the effective band-pass filter is tuned by adjusting t s and T D while maintaining a high signal amplification. The optimum center frequency is selected when the noise is minimized for a given CCD noise input spectrum.
  • the correlated double sampling circuit is configured so as to completely eliminate kTC noise and to optimally filter white and 1/f noise generated by the on- chip CCD amplifier, thereby improving overall signal-to- noise ratio.
  • a dual slope processor Another type of correlated double sampling circuit that may be used is a dual slope processor.
  • the perform ⁇ ance of the dual slope processor when operating at a slow data rate is very similar to that of the dual sampler processor, but the dual sampler processor is preferred at high pixel rates.
  • a diagram of a dual slope processor is shown in Fig. 11A.
  • Timing diagrams for the dual slope processor of Fig. 11A are shown in Fig. 11B.
  • a resistor 352 represents the output resistance of a CCD detector.
  • a CCD output signal 350 is connected to a pre-amplifier 353.
  • the output of the pre-amplifier 353 is connected to an inverting amplifier 355 and a non-inverting amplifier 354, each of which have a gain of unity.
  • the output of the inverting amplifier 355 and the output of the non- inverting amplifier 354 are each connected to a three- position switch 360.
  • the output of the non-inverting amplifier 354 is connected through a resistor 356 to one position 361 of the three-position switch, while the output of the inverting amplifier 355 is connected through a resistor 357 to another position 363 of the three- position switch 360.
  • a third position 362 of the three- position switch 360 is left open.
  • a common end 364 of the three-position switch 360 is connected to an integrator 370.
  • the integrator 370 comprises an operational ampli ⁇ bomb 366 and a capacitor 367.
  • the input and the output of the operational amplifier 366 are connected to opposite ends of the capacitor 367 as shown in Fig. 11A.
  • Another switch 365 is connected from the input of the integrating amplifier 366 to ground 369.
  • the CCD video signal 350 is amplified by the pre-amplifier 353.
  • the pre-amplifier 353 may have a gain of, for example, 2 V/V.
  • the output of the pre- amplifier 353 is fed into the integrator 370 by way of the three-position switch 360.
  • Processing of the CCD video signal 350 begins by first resetting the integrator 370 to ground potential by turning on switch 365 and setting switch 364 to position 362. After resetting the inte- grator 370, switch 365 opens, and the three-way switch 360 switches to the first position 361 so as to allow the reference voltage level to be integrated onto the integrating capacitor 367 for a sample time t s .
  • the effect of integrating is to subtract the reset potential and thereby eliminate common mode reset noise.
  • charge is dumped onto the sense node (e.g., the sense capacitor 152) of the CCD.
  • the three-position switch 360 is switched to the second position 363 to connect the integrator 370 to the output of the inverting amplifier 355.
  • the refer ⁇ ence level plus the video level are integrated for t s seconds when the switch 360 opens to the third position 362.
  • the signal is then held by the integrator 370.
  • Output signals from either the dual sampler or dual slope processors may be digitized by suitable means, such as and analog-to-digital converter, and processed digit ⁇ ally, or may be processed using some or all of the analog techniques described further herein.
  • CCD detectors place special requirements on the optical system.
  • a CCD-based bar code reader designed to read fast moving bar code labels must have a relatively short exposure time.
  • the optical system of such a bar code reader is thus preferably configured so as to gather the maximum light available.
  • the invention contemplates a number of different optical structures that are suitable for use in the pre ⁇ ferred embodiments. These optical structures include, for example, poly-optic and multi-focus lenses, graded index elements, optical flats, prisms, and others. These struc ⁇ tures may be used alone or in combination to increase the light gathering power and depth of field of the various embodiments described herein.
  • a simple optical system for a linear CCD is a single lens system.
  • a single lens model is diagrammed in Fig. 13A, which shows some of the basic relationships between reading distance and length of scan line.
  • the single lens model of Fig. 13A shows the gathering of light from a target 400 focused through a lens 401 and projected onto an imager (e.g., CCD detector) 402. As shown in Fig.
  • the lens 401 can be thought of as projecting from the imager 402 a virtual scan line of length W F at a distance X F where the lens 401 is focused.
  • the ratio of the virtual scan line length W F to the imager length W A is the ratio of the focal distance X F to the imager distance I:
  • the width W F corresponds to the field of view of the imager 402 at a distance X F .
  • Design of an optical system may involve considera ⁇ tions of focus blur, diffraction, aberrations, variable light intensity, light scattering, label reflectivity, and label print contrast ratio. These topics are discussed briefly below before a more complete description is pro ⁇ vided of the preferred embodiments of optical structures in accordance with various aspects of the present invention.
  • the phenomenon of focus blur relates to the distance of the target image from the focal spot of the lens. If a bar code is located at a distance X F from the lens and within the field of view width W F , then the bar code will be in focus on the imager 402. If, as shown in Fig. 13B, the bar code is moved closer to or farther away from the lens 401, the image will become blurry. This blurriness results from the bar code being imaged by a "virtual spot" 405 or blur spot that is larger than the focal spot at the optimum focal point 406. The size of the blur spot 405 is a linear function of distance from the focal point 406, as can be seen in Fig. 13B.
  • the blur spot 405 grows more rapidly with distance for a wider lens aperture (opening) than for a narrower lens aperture.
  • the ratio of the aperture size to the focal length is called the f-number or f-stop and is written as f/#.
  • an f/4 lens will have half the depth of field of a f/8 lens, because the diameter of the aperture of the f/8 lens is half the aperture diameter of the f/4 lens.
  • An analogy can be drawn between a CCD imaging system and a laser optic system, where the waist of the laser beam corresponds to the focal point of the imaging system, and the size of the laser spot corresponds to the image of the point spread function at the target (e.g., bar code) .
  • the spot at the focal point 406 may not be of negligibly small size, but rather may be somewhat larger due to the effects of diffraction and lens aberrations.
  • the size of the lens aperture affects the spot size resulting from diffraction, which may occur when only a portion of a wavefront of light is allowed through the aperture. Larger apertures generally result in a smaller spot size, while smaller apertures result in a larger spot size.
  • the effects of diffraction also depend on the wavelength of the light, whether ambient or self- generated. For example, blue light tends to create a smaller spot size than red light.
  • the effects of focus blur and diffraction spot size may estimated as a root mean square (i.e., the square root of the sum of the squares of both spot sizes) .
  • lens aberrations may lead to larger waist spot sizes.
  • the resulting waist spot size may be expressed as D(l+K), where D is the diffraction limited spot size and K is an aberration factor.
  • D the diffraction limited spot size
  • K an aberration factor.
  • spherical aberration is the primary type of aberration with large aperture lenses (larger than f/3) .
  • Spherical aberration arises from the difference between a perfect shaped lens and a spherical shaped lens.
  • a spherical wavefront is required.
  • a spherical lens shape is not the correct shape to create a spherical wavefront.
  • the correct lens shape depends on the object-to-image distance ratio (referred to as the magnification ratio or conjugate ratio) .
  • the difference between this perfect shaped lens and a spherical shaped lens causes spherical aberration.
  • the effect of spherical aberration, as shown in Fig. 14A, is an effective shift in focal length depending on where radially on the lens the light passes (i.e., ray height) .
  • the region of the lens near the center has one focus, while concentric rings of the lens radiating outward from the center will have differing focal lengths depending on their distance from the center.
  • Spherical aberration can be minimized or eliminated by using a properly shaped spherical lens (e.g., plano-convex, double-convex, meniscus) , shaping the lens correctly (such as with aspheric lenses) , or by using multiple lenses (e.g., doublets or triplets).
  • a properly shaped spherical lens e.g., plano-convex, double-convex, meniscus
  • shaping the lens correctly such as with aspheric lenses
  • multiple lenses e.g., doublets or triplets.
  • Chromatic aberration relates to the variation of focal length as a function of wavelength.
  • the focal length variation occurs because the index of refraction of an optical substance varies with wavelength (an effect called dispersion) .
  • Chromatic aberration is the primary type of aberration with apertures smaller than f/2, when astigmatism is not present. Chromatic aberration can be corrected by using two different lenses with different dispersion properties.
  • Coma relates to a variation in magnification as a function of ray height when imaging objects not on the optical axis.
  • the effect of coma, illustrated in Fig. 14B, for example, is typically overshadowed by the effect of astigmatism, but can nevertheless be corrected by proper placement of the aperture.
  • Astigmatism is the difference in focus between the horizontal and vertical axis.
  • the effect of astigmatism, illustrated in Fig. 14C, for example, is to cause the image of tangential lines (i.e., vertical axis) to appear at one distance and radial lines (i.e., horizontal axis) to appear at a different distance.
  • Lenses with or without astigmatism will generally exhibit field curvature, which is the bending of the image plane. Field curvature is characterized by different tangential or sagittal lines, as shown in the comparative diagram of Fig. 14D.
  • Cosine law effect a term which describes the variation of light intensity across an imager. Cosine law effect is most pronounced with a wide angle field of view. Several factors each having a "cosine" component combine to form an intensity distribu- tion that varies as the cosine to the fourth power as a function of field angle. Wider field angles thus produce larger variations in signal intensity over the imager, while narrower field angles produce smaller signal intens ⁇ ity variations. Cosine law effect should be taken into account so as to collect enough light (i.e., obtain a suf ⁇ ficient signal level) at the edges of a virtual scan line. A graph of the cosine law effect is shown in Fig. 15.
  • the amount of light collected by the imager 402 depends in part on the intensity of the light (whether ambient or not) and the reflectivity of the bar code label.
  • a light collection model is illustrated in Fig. 16. If a target label 410 with reflectivity r is illumi ⁇ nated with light 411 having an irradiance Ii lumens/m 2 , then Ii*r lumens/m 2 will generally be re-radiated from the label.
  • a Lambertian (perfectly diffuse) source that emits N lumens/m 2 into a hemisphere emits N/ ⁇ lumens/steradian-m 2 at normal incidence to the source.
  • the radi ⁇ ance of the target label (at normal incidence) having irradiance Ii-r may be expressed as lumens/ steradian-m 2 .
  • An imaging system will receive an,irradiance at the imager 402 equal to ⁇ TN'sin 2 ⁇ , where N' is the object radiance in lumens/steradian-m 2 , T is the trans ⁇ mission of the system, and ⁇ is the half angle from the center of the imager to the edge of the lens.
  • I C I X - (rT/4A 2 ) • (X F - f) 2 /X F 2
  • l t Incident Light (lumens/m 2 )
  • I c Collected Light (lumens/m 2 )
  • T Lens Transmission (percentage)
  • Aperture Half Angle (degrees)
  • A f/# of Aperture (unitless)
  • f Focal Length of Lens (meters)
  • X F Distance to Focus (meters)
  • the amount of light collected at the imager 402 is independent of distance of the target label from the lens 401.
  • the collected light is, however, proportionally affected by the intensity of the light (ambient or other- wise) illuminating the label, the reflectance of the label, and the lens transmission.
  • the collected light is inversely proportional to the f-number (f/#) squared, and depends to some degree on the position of focus (X F ) relative to the focal length.
  • the collected light I c may be multiplied by the print contrast ratio of the target label. It should be noted that the above calculations do not take into account any signal modulation reduction resulting from modulation transfer function (MTF) effects or cosine law effects.
  • MTF modulation transfer function
  • Focus Blur (mils) V [(Foe) 2 +(Diff) 2 ] where Foe is geometric focus blur, and Diff is diffraction limited minimum spot size;
  • Foe (mils) 1000- [f/(f/#)] • [(R - R 0 )/R 0 ] where R is distance from lens (inches) , R 0 is focal point (inches) , f is focal length of lens (inches) , and f/# is relative aperture (f-stop) ;
  • Diff (mils) 9.61xl0 "5 ⁇ R- (f/#)/f where R is distance from the lens (inches) , ⁇ is nominal wavelength (nanometers) ; f is focal length of the lens (inches) , and f/# is relative aperture (f-stop) ;
  • Resolution (mils) 1000WR/ [I (f,R 0 )N] - 1000WR/ (fN) where R is distance from the lens (inches) , W is width of the CCD array (inches), I(f,R 0 ) is CCD image plane (inches) , and N is the number of pixels in CCD array,
  • - Field of View (mils) 8.333 WR/[I(f,R 0 )] - 8.333 WR/f where R is distance from the lens (inches) , W is width of the CCD array (inches), and I(f,R 0 ) is CCD image plane (inches) .
  • Astigmatism ⁇ 2 -X/(2A) where ⁇ is the wavelength in meters, A is the relative aperture (f/# or f-stop) , V is the Abbe number, and ⁇ is the half field angle in radians.
  • At least 2.5 pixels per minimum element width are needed to resolve bar codes. Focus blur can frequently be 1.9 times bigger than the minimum element width. Spots as big as 2.4 times the minimum element width may still provide workable results in low noise systems.
  • Certain of the preferred embodiments make use of shaped apertures and specialized lenses in order to increase light gathering capability and allow operation in reduced light, such as in ambient light conditions. Since in many applications a bar code reader images only bars/ spaces and not necessarily true two dimensional images, some optimization of the lens aperture is possible. Increasing the size of the aperture in the direction of the bars generally provides greater collection with minimal effect on the depth of field characteristics.
  • Figure 17 is a diagram comparing the point spread function shapes through focus for a round aperture 430 and an oblong or elliptical aperture 431, both having the same light collection. As shown in Fig. 17, at a focal point 435, the round aperture through focus point spread spot 436 and the elliptical aperture through focus point spread spot 437 are each at a minimum. However, the elliptical aper ⁇ ture through focus point spread spot 437 is larger in the lateral direction than the round aperture through focus point spread spot 436.
  • w c represents the narrowest width of the spot
  • represents the tilt angle (e.g., with respect to the vertical bars of a bar code)
  • w represents the projected width.
  • the area of the elliptical spot is given by the following equation:
  • a narrower aperture such as the elliptical aperture 431 (or other oblong aperture, such as a rectangular aper ⁇ ture) shown in Fig. 17, generally tends to increase the depth of field.
  • the height of the aperture 431 in the direction parallel to the bars is enlarged to equalize the collection.
  • Figure 19 is a graph showing illustrative depth of field plots for a round aperture 430 and for an elliptical aperture 431 collecting the same amount of light.
  • the graph of Fig. 19 plots the minimum resolvable bar width on one axis 440 against the distance from the lens on the other axis 441.
  • the narrower aperture creates a larger depth of field than the wider aperture, but has a larger minimum spot size due to diffraction.
  • An optical system employing a small, round aperture generally has a relatively good depth of field but suffers from the need for a long exposure time.
  • a small, round aperture e.g., aperture 430
  • Such a system is shown, for example, in Fig. 20A.
  • a target 444 e.g., a bar code, shown from a back view
  • a round lens 443 or lens having a round aperture
  • CCD detector 442 In the optical system of Fig. 20A, as little as 0.02% of the reflected light may reach the CCD detector 442.
  • An oblong aperture (e.g., aperture 431) speeds up exposure time by allowing more light collection.
  • a system comprising and oblong aperture is shown, for example, in Fig. 20B.
  • a target 458 e.g., a bar code, shown from a back view
  • an apertured lens 457 or an oblong or similarly shaped lens
  • a narrow axis 446 of the aperture 445 corresponds to an f- stop for the desired depth of field (e.g., f/20)
  • a long axis 447 of the aperture 445 has a larger f-stop (e.g., f/5) and therefore gathers more light.
  • the range of angles of bar codes that the optical system of Fig. 20B will detect may be reduced in comparison to the Fig. 20A system by use of a oblong aperture, the collected light can be substantially increased (e.g., by as much as ten times or more) .
  • a preferred embodiment of a CCD imaging device employs a CCD detector 1079 as shown in Fig. 50A whose pixels 1080 are rectangular in shape, with their long axes perpendicular to the line formed by the linear array of pixels 1080.
  • the Fig. 50A rectangular pixel structure enhances the detection efficiency of the CCD detector 1079 without significantly degrading spatial resolution, so long as the CCD detector 1079 is not rotated at too much of an angle with respect to the barcode (or other image) being read.
  • Figures 50A and 50B are diagrams illustrating the effect on an output signal 1082A, 1082B of rotating a bar code with respect to the CCD detector 1079.
  • the unrotated bars 1081A projected onto the CCD detector 1079 in Fig. 50A result in a relatively clean square-wave output signal 1082A.
  • the partially rotated bars 1082A projected onto the CCD detector 1079 in Fig. 50B result in an output signal 1082B with sloped rising and falling edges due in part to the spreading of the bars 1081A across a greater number of pixels 1080.
  • the pixel aspect ratio should be selected so that the resolution of the bar code (or other image being read) is not impaired with labels that have a tolerable amount of roll with respect to the imaging device. Moreover, if the pixels are too tall, the projected image of the pixels will provide too little resolution to properly image the barcode.
  • Figure 51 is a graph comparing pixel width versus roll angle for a CCD array with square pixels and with rectangular pixels (having a 4:1 aspect ratio). As shown in Fig.
  • the angle of bar code label rotation does not significantly affect the normalized projected pixel width for a CCD array with square pixels; however, it does affect the normalized projected pixel width for a CCD array with rectangular pixels having a 4:1 aspect ratio. Thus, it is possible that more rectangular pixels would be needed to obtain the same performance as square pixels for handling rotated bar code labels.
  • N W ef£ (0, K-f)-cos(0) (51.1)
  • W e£f the projected or effective width (i.e., how big the pixel width looks with a certain rotation)
  • K the aspect ratio.
  • the value f is a factor reflecting the tolerance of the signal processing electronics to a distorted output signals.
  • f is generally equal to about 0.8.
  • the maximum acceptable pixel aspect ratio is about 10 to 1.
  • Figs. 50A and 50B Use of rectangular pixels such as described above with respect to Figs. 50A and 50B is not limited to CCD devices, but may be applied to any pixel-based photodetection device. If an existing CCD detector 1079 has an aspect ratio that is considered larger than optimum to handle rotated bar code labels at all tolerable angles, there are several ways to modify the system to provide a more acceptable aspect ratio without requiring the CCD detector 1079 itself be modified. For example, as shown in Figs. 52A and 52B, the CCD detector 1079 may be transformed into a CCD detector 1090 having a more acceptable aspect ratio by use of a metal mask 1091 over the pixels. The metal mask 1091 has a slit 1093 which allows light to reach only the portions of the CCD pixels conforming to the height of the mask, thereby allowing relatively simple mechanical adjustment of the effective height of the CCD pixels.
  • a metal mask 1091 may prove to be more economical than use of a different CCD detector having the proper pixel aspect ratio.
  • the mask 1091 can be applied externally in front of the CCD detector 1090, as shown in Figs. 52A and 52B.
  • the mask 1091 is preferably physically close (e.g., touching) to the imager plane so as to reduce the effect of vignetting, which would cause light rays to spill over into regions of the CCD photodetectors not intended for light collection.
  • Figure 52B shows a flux density graph 1092 adjacent to the CCD detector 1090, which illustrates a gradual roll off of flux density caused by small amounts of vignetting even with a properly positioned mask 1091. If, however, the mask 1091 is too far away, it ceases to be useful for reducing the pixel aspect ratio and simply reduces light collection efficiency.
  • the top layer metal mask of the CCD chip itself is modified so as to provide an appropriate sized slit 1093 in front of the CCD photo ⁇ detectors.
  • a preferred mask slit on the top layer metal mask of the chip has a height dimension and a width dimension so as to provide a 10:1 pixel aspect ratio.
  • An alternative modified CCD imaging system with an external mask is shown in Fig. 52C, wherein extra optics are used to re-image the pixels to a new imaging plane.
  • an imaging lens 1097 is positioned in front of a mask 1096 (opposite the CCD array 1098) having a slit 1099.
  • the mask 1096 is positioned further away from a CCD detector 1098 than the Fig. 52A- 52B system.
  • a second lens 1095 In between the mask 1096 and the CCD detector 1098 is positioned a second lens 1095.
  • the second lens 1095 reduces or eliminates vignetting by re-directing light focused by the imaging lens 1097 onto the CCD detector 1098.
  • the corrective mask 1096 is more effective in this system than in the system of Figs. 52A- 52B, and the problems associated with vignetting are less of an issue.
  • the Fig. 52C system does, however, have the additional complexity of a second lens and a longer path length for the optical system.
  • FIG. 53 Another alternative for modifying pixels of a CCD detector with too large an aspect ratio is an anamorphic lens system, such as shown in Fig. 53.
  • the Fig. 53 system includes a CCD detector 1100 in front of which is posi ⁇ tioned a first aspheric cylinder lens 1102 in a vertical orientation with respect to the CCD detector 1100 and a second aspheric cylinder lens 1104 in front of the first aspheric cylinder 1102 and in a horizontal orientation with respect to the CCD detector 1100.
  • Each aspheric cylinder lens 1102, 1104 provides significant modulation along one axis only.
  • the horizontally oriented aspheric cylinder lens 1104 is provided with a lower magnification power than the vertically oriented aspheric cylinder lens 1102, thereby reducing the effective aspect ratio of the pixels 1106 of CCD detector 1100.
  • An element of height HI and width Wl i.e., aspect ratio of H1:W1 is therefore, by the operation of aspheric cylinder lenses 1102 and 1104, altered in apparent dimension at the CCD detector 1100 to have an aspect ratio of H2:W2.
  • the "projection" of the pixels 1106 of the CCD detector 1100 onto the bar code plane are reduced in height from H2:W2 to H1:W1, where H1/H2 ⁇ W1/W2.
  • the distance from the CCD detector 1100 to the vertically oriented cylinder lens 1102 is given as I l r and the distance from the CCD detector 1100 to the horizont- ally oriented cylinder lens is given as I 2 .
  • the distance from the bar code 1105 to the vertically oriented cylinder lens 1102 is given as 0 1# and the distance from the bar code 1105 to the horizontally oriented cylinder lens is given as 0 2 .
  • a possible drawback with the Fig. 53 system is that aspheric cylinders may be hard to manufacture.
  • a preferred embodiment of the present invention also makes use of a multi-focus lens system (e.g., a poly-optic system) .
  • a multi-focus lens allows an increase in depth of field without sacrificing the ability to collect light.
  • Each lens element (called a zone) of a multi-focal lens is focused at a different distance.
  • the field from each zone is simultaneously focused on the CCD detector.
  • each zone is multiplexed (with an LCD shutter, for example) on the CCD detector instead of simultaneously superimposing all zones.
  • each zone is imaged onto one of a plurality of separate CCD detectors.
  • FIG. 22A-22C An exemplary embodiment of a multi-focus lens system combined with an elliptical aperture is shown in various views in Figs. 22A-22C.
  • the symmetric multi-focal lens 450 has concentric zones and is depicted by itself in Figs. 21A and 21B.
  • Figure 21A shows the multi-focus lens 450 from an oblique view
  • Fig. 21B shows the same lens 450 from front view.
  • the lens 450 is divided into a plurality of concentric zones 451.
  • Such a lens 450 has a symmetric configuration and may be suitable for machining with a diamond turned lathe.
  • the lens and aperture system of Figs. 22A-2C are shown from three different views.
  • Figure 22A shows a side view
  • Fig. 22B shows a top view
  • Fig. 22C shows an oblique view.
  • the lens and aperture system of Figs. 22A- 22C comprises an elliptical (or rectangular) aperture 461 located in front of a multi-focus lens 460.
  • the ellipti ⁇ cal aperture 461 preferably has a high aspect ratio.
  • the elliptical aperture 461 and multi-focus lens 460 are placed before an imager 462, such as the CCD detector described earlier herein. In operation, light reflected from a bar code 463 passes through the elliptical aperture 462 and is focused by the multi-focus lens 460.
  • Each zone 465 of the multi-focus lens 460 focuses the reflected light 466 in a different manner on the imager 462, because each zone 465 focuses at a different distance (i.e., has a different focal length in front of the lens 460) .
  • the depth of field is thereby expanded over that of a single- focus lens.
  • the response of a multi-focus lens may be determined under principles of superposition by summing the responses of each lens zone 451 by itself. If a shutter is used to screen out zones 451 of the ulti- focus lens 450 (e.g. by use of an LCD) , only one zone 451 (or however many zones are unscreened) at a time need be analyzed, and the result of each individual zone super ⁇ imposed to obtain a composite response for the lens 450.
  • Figure 23A is a graph showing the minimum renderable (i.e., resolvable) element width as a function of distance from the lens in a single-lens system.
  • Figure 23B is a graph showing minimum renderable bar width as a function of distance from the lens for a poly-optic (i.e., multi- focus) lens system, assuming no interference.
  • Each parabolic-shaped plot 471 represents a particular zone 451 of the poly-optic lens 450.
  • a focal point 470 corresponds to the minimum point of the parabola.
  • the plots 471 are superimposed on one another to graphically show the crea ⁇ tion of an enlarged depth of field for the poly-optic system.
  • Fig. 23C is a similar plot to Fig. 23B but assumes that there is intra-lens interference from the poly-optic lens 450. Interference between zones 451 may be caused by lobes in the line spread function of the lens 450, by mis ⁇ alignment of the optical centers of the zones 451, or a difference in the field of view of each zone 451. Different lens configurations generate different lobe patterns. Some point spread functions for a few common poly-optic lenses are shown in the Figs. 24A through 30C and described below.
  • a simple multi-focus lens 450 configuration comprises a set of concentric zones 451, each with its own focal length, as shown previously in Figs. 21A-21B.
  • the point spread function through focus for the center zone 452 of the Fig. 21A lens is shown in Fig. 24A at focus and Fig. 24B away from focus.
  • the point spread function through focus for a particular ring zone 454 is shown in Fig. 25A at focus and Fig. 25B away from focus.
  • the diagrams of Figs. 24A and 25A are full scale, while the diagrams of Figs. 42
  • Figures 26A through 30C depict point spread and line spread function plots of various combinations of center and ring zones of various focal lengths.
  • the optical point spread function images appearing in Figs. 26A through 30C have been approximated for simplicity.
  • the point spread function far from focus can be well approximated by geometric optics.
  • the shape of the point spread function far from focus is similar to a cylinder.
  • the point spread function's shape is determined by aberrations
  • FIG. 26A through 30C are primarily diffraction limited (i.e., they assume only small aberrations) at or near focus.
  • Figure 26B is a diagram of one image 480 created from a center zone 452
  • Fig. 26A is a diagram of another image 481 created from a ring zone 454, where the center zone 452 and ring zone 454 have differing focal lengths.
  • Figure 26C is a diagram showing a combined image 482 created from the combination of the center zone 452 with the ring zone 454 with only the center zone 452 in focus.
  • the image from the ring zone 454 is not readily visible mainly due to its low flux density, since it is out of focus.
  • the plots of the line spread functions 483 and 484 clearly show the effect of the ring zone 454.
  • the line spread functions 483, 484 each have a main peak 488 and 489, respectively, corresponding to the center zone 452 and a pair of much lower peaks 490 and 491, respectively, corresponding to the ring zone 454.
  • the effect of the ring zone 454 will become considerably more pronounced.
  • Figures 27A-27C illustrate point spread functions in between the focus of both zones.
  • Figure 27B is a diagram of an image 500 created by the center zone 452
  • Fig. 27A is a diagram of an image 501 created by a ring zone 454, assuming differing focal lengths.
  • Figure 27C is a diagram of a combined image 502 created from both the two zones 452, 454, while in between the focus of both. It may be noted that, while there is a "dip" 503 or dead spot in the center of the combined image 502, the line spread functions 505, 506 are relatively free of the effect of this dip 503.
  • Figures 28A-28C illustrate point spread functions when the ring zone 454 is in focus.
  • Figure 28B is a diagram of an image 510 created by the center zone 452
  • Fig. 28A is a diagram of an image 511 created by a ring zone 454, assuming differing focal lengths.
  • Figure 28C is a diagram of a combined image 512 created from both the two zones 452, 454. The combined image 512 of Fig. 28C is very similar to that of Fig. 26C, wherein only the center zone 452 was in focus.
  • the line spread functions 515, 516 in Fig. 28C show the effect of the out- of-focus center zone 452.
  • Figures 29A-29C illustrate a more severe effect, wherein a ring zone 454 is not blurred enough with respect to the center zone 452.
  • the result is an image 522 having a ring structure.
  • Figure 29B is a diagram of an image 520 created by the center zone 452, while Fig. 29A is a dia ⁇ gram of an image 521 created by the ring zone 454.
  • Figure 29C is a diagram of a combined image 522 created from both the two zones 452, 454 contains the ring structure, as mentioned, which appears as diffuse light in a ring pat ⁇ tern around the center focal point.
  • the line spread func ⁇ tions 525, 526 indicate the effect of this ring structure.
  • the ring structure may cause excessive inter-symbol interference (ISI) when attempting to read bar codes.
  • ISI inter-symbol interference
  • FIGS. 30A- 30C are diagrams of images created under the same condi ⁇ tions as Figs. 29A-29C, but for a system where the ringed zone is apertured with two circular apertures.
  • Figure 3OB is a diagram of an image 530 created by the center zone 452
  • Fig. 30A is a diagram of an image 531 created by the apertured ring zone 454.
  • Figure 30C is a diagram of a combined image 532 created from both the two zones 452, 454.
  • the use of two circular apertures as shown in the horizontal line spread function 536 of Fig. 30C successfully removes the side lobes in the horizontal pattern. This form of aperturing can thus be used to read bar codes whose bars are vertically oriented without creating an undue amount of inter-symbol interference from the multi-focus lens.
  • Figure 31A-31C are diagrams illustrating a particular apertured multi-zone lens configuration providing a line spread function similar to that in Fig. 30C.
  • a symmetric multi-focus lens 601 is shown from several viewpoints with a plurality of apertures 602, one aperture 602 located in front of each one of a plurality of zones 603 the lens 601.
  • Fig. 30C shows an image resulting from two circular apertures
  • the apertured multi-focus lens 601 of Figs. 31A-31C provides a plurality of images from the multiple ring zones and a corresponding plurality of peaks in the vertical line spread function 535.
  • the apertured multi-focus lens 601 of Figs. 31A-31C would, like the configuration of Fig. 30C, also have a horizontal line spread function 536 relatively free from intersymbol interference.
  • the apertures 602 of Fig. 31A or 3IB need not be circular, but may be square, rectangular, or any other suitable shape. In a preferred embodiment, each of the apertures 602 are roughly square in shape. When physic- ally manufactured the square apertures can be melded into a continuous opening similar to that of the aperture 461 in Figs. 22A-22C. While such a continuous opening may be "rectangular" in overall shape, it functions as a series of discrete, individual apertures, one for each lens zone. The selection of the height of the apertures 602 in front of each zone may be a compromise balancing light collection requirements and roll response characteristics. The width of each aperture 602 controls the depth of field for its lens zone, and is preferably the same for each aperture 602 so as to minimize interzonal interference.
  • each of the apertured zones 603 may have a different focal length. This may be achieved in at least two different ways.
  • a symmetric lens 610 is physically cut immediately below the center zone 611, as shown in Fig. 32.
  • the cut lens 612 will have the characteristics of an asymmetric multi-focus lens.
  • the lens 620 may be molded in an asymmetric fashion, as shown in Fig. 33, such that the lower portions 621 of each ring zone have a different focal length than the upper portions 622 of each ring zone.
  • An advantage of the latter technique is that the lens could be constructed so as to have its optical axis 625 in the center of the lens 620, as shown in Fig. 33.
  • the off-axis tilt is typically smallest when the center zone 627 focuses to the closest depth and the outside ring zone 628 focuses to the farthest depth, as shown in Fig. 34A.
  • Figure 34B illustrates the reverse situation, which may lead to greater pointing problems.
  • An alternative optical collection system employs a "masked” or apodized multiple focus lens, such as shown in either of Fig. 54A or 54B.
  • a multiple-zone lens 1120 is shown having concentric ringed zones 1121a through 1121d, and further having darkened portions 1122 which are masked or apodized.
  • the apodized portions 1122 can be created by depositing silver or other reflective material on the lens 1120.
  • a multiple-zone lens 1130 is shown having concentric ringed zones 1131a through 1131d, and also having apodized portions 1132.
  • the Fig. 54A and 54B provide a multiple focal plane optical col ⁇ lection system which maintains high collection efficiency while reducing interference between light rays from the different focal planes that are simultaneously incident on a CCD detector.
  • the masking can therefore replace the aperture(s) of the embodiments shown in Figs. 22A-22C and 31A-31C, inasmuch as the masking has a similar effect as the aperture(s) in reducing inter-zonal interference.
  • the Fig. 54A and 54B lenses 1120, 1130 preferably each comprise a set of concentric annular zones 1121a-d and 1131a-d, respectively, having varying focal lengths.
  • each zone 1121a-d and 1131a-d is preferably substantially equal, so that substantially equal amounts of light are collected from each zone.
  • the radii required to maintain constant collection area from zone to zone will change in a straightforward fashion from this equation to account for the reduction in area due to the masking.
  • Figure 55A is a graph of individual line spread functions 1141 through 1144 for each lens zone 1121a-d or 1131a-d of the multi-focal lenses of either Fig. 54A or 54B.
  • Lobes that would otherwise exist in the horizontal line spread functions are reduced or eliminated by use of a mask (e.g., apodization on the lens itself) to maintain a constant vertical height for each zone 1121a-d or 1131a- d of the lens 1120 or 1130.
  • the resulting composite or vertically-integrated horizontal line spread function 1146 is shown in Fig. 55B.
  • the resulting vertically-integrated horizontal line spread function 1146 is relatively flat, without lobes.
  • the line spread function varies depending on the focal distance and the effect of aberrations and diffrac ⁇ tion.
  • the Fig. 54A and 54B masking will provide a composite line spread function as shown in Fig. 55B.
  • an optimal mask pattern is selected which minimizes sidelobes over a maximum range of distances.
  • the optimal mask pattern selection depends upon a variety of factors, including the number and position of focal planes present in a particu ⁇ lar optical collection system, the degree of aberration correction required, the angular field of view required, and/or the size of the barcodes to be read.
  • another method for constructing multi-focus lenses having the desired properties is to combine a plurality of indi- vidual lenses, one for each zone.
  • Each individual lens has an off-center circular aperture and is optimized for the desired focus condition of the corresponding zone.
  • Each lens zone in turn is optimized.
  • the final lens essentially comprises a piecewise continuous combination of each individual lens zone.
  • the lens can be optimized in at least two ways.
  • a global merit function can be made that takes into account the focal properties of each zone.
  • the entire aspheric deformation coefficients can then be solved at once.
  • the second method is to allow each zone a few deformation coefficients, giving the center zone the lowest order coefficients and the edge zone the highest order coefficients.
  • the system is optimized in order, from the center zone to the edge. This method involves an iterative process, as the edge zones may need to cause slight modifications to the lower order coefficients so as to minimize aberrations.
  • Fig. 35A One particular optical lens structure is shown in Fig. 35A.
  • Fig. 35A which is drawn from a top view, an array 640 of single-focus lenses 641 are arranged in an arc-like pattern around an imager (e.g., a CCD detector 642) .
  • Each lens 641 is a slightly different distance away from the CCD detector 642, so the pattern formed by the lens array is not a perfect arc.
  • a set of different focal length lenses could be aligned in an arc to achieve the same result.
  • the centers of the scan lines 643 associated with each of the lenses 641 fan out in a radial fashion from the CCD detector 642.
  • the focal points 644 of the lenses 641 are staggered, so that the top lens 641a has the closest field depth 645a (because it has the nearest focal distance) , the next lens 641b has the next closest field depth 645b, and so on, until the last lens 641e which has the farthest depth of field 645e.
  • the structure may provide for a seamless depth of field without using shutters, so long as the lenses 641 are arranged so as not to cause interference with one another.
  • Another advantage is that the structure mitigates the effect of coma, due to the radial fanning of the centers of the scan lines 643.
  • FIG. 35B A related embodiment of an optical structure is shown in Fig. 35B.
  • the Fig. 35B structure is similar to the Fig. 35A structure, except that the Fig. 35B embodiment is shown from a side view instead of a top view, so that the array 650 of lenses 651 expands the depth of field vertically rather than horizontally.
  • the Fig. 35B structure has similar properties of an arc-like pattern of lenses 651, radially fanning scan lines 653, staggered depth of fields 655, and a corresponding mitigation of the effect of coma and aggregate expansion of depth of field for the system.
  • FIG. 36A is a modification of the system of Fig. 35B.
  • the Fig. 36A embodiment depicts an array 660 of lenses 661 arranged in an arc-like pattern around a CCD detector 662. Scan lines 663 fan out radially from the CCD detector 662.
  • the Fig. 36A embodiment further adds a set of mirrors 667 arranged to form the radial scan lines 663 into a collimated set of scan lines 668.
  • the individual depths of field 665 are staggered for each lens 661 along the collimated scan lines 668.
  • a composite depth of field 669 is shown adjacent to the collimated scan lines 668.
  • FIG. 36A embodiment is that it provides a more dense coverage of a plane.
  • prisms or light-directing lenses may be used instead, or any combination of mirrors, prisms and/or light directing lenses.
  • Figure 36B shows the Fig. 36A embodiment implemented using prisms 677 instead of mirrors.
  • Figure 36C shows the Fig. 36A embodiment implemented using a single collimating lens 687 to perform a similar function.
  • FIG. 37A Another embodiment of an optical structure is shown in Fig. 37A.
  • the Fig. 37A structure is similar to that of Fig. 36C, but reverses positions of the large single lens 707 and the array of lenses 701.
  • a large single collimat ⁇ ing lens 707 is located in front of an imager (e.g., a CCD detector 702) .
  • a plurality of smaller lenses 701 are positioned in front of the larger lens 707, and are staggered slightly in distance from the larger lens 707.
  • Scan lines 708 from the smaller lenses 701 are directed in a collimated pattern to provide dense coverage of a plane.
  • the scan lines 708 are directed by the larger lens 707 to the CCD detector 702.
  • FIG. 37B Another embodiment of an optical structure is shown in Fig. 37B.
  • the structure of Fig. 37B is conceptually similar to that of Fig. 37A.
  • Figure 37B like the structure of Fig. 37A, comprises a large collimating lens 717, convex in shape, positioned in front of a CCD detector 712.
  • the Fig. 37B structure also comprises an array of smaller lenses 711. Rather than staggering the distance of the smaller lenses 711 from the larger lens 717, however, in Fig. 37B is added a set of staggered optical flats 719 positioned between the larger lens 717 and the array of smaller lenses 711.
  • the staggered optical flats 719 shift the focal point of the smaller lenses 711 and provide staggered of depths of field 715 along the collimated scan lines 718 associated with the array of lenses 711.
  • An advantage of the Fig. 37B system is that the smaller lenses 711 need not be staggered, and therefore may facilitate manufacture of the optical system.
  • FIG. 37C Another embodiment of an optical structure is shown in Fig. 37C.
  • the structure of Fig. 37C is similar to that of Fig. 37B, except that a graded index element (also called a "GRIN" element) 729 is used in place of the staggered optical flats 719.
  • An array of smaller lenses 721, as in Fig. 37B, is non-staggered.
  • the smaller lenses 721 are positioned in front of the GRIN element 729, which is positioned in front of a large collimating lens 727 set before the CCD detector 722.
  • Scan lines 728 form a colli ⁇ mated pattern, providing dense coverage of a plane.
  • the GRIN element 729 is formed of a material having a graded index, so that the path length differs along its axis.
  • the path is long, while towards the other end 731 of the GRIN element 729, the path is short. Accordingly, the field depths 725 are staggered along the collimated scan lines 728, as shown.
  • Figure 38A shows another embodiment of an optical structure.
  • the Fig. 38A structure is similar to that of 51
  • a poly-optic lens 751 incorporates aspects of the large collimating lens 717, the staggered optical flats 719, and the array of smaller lenses 711 shown in Fig. 37B into a single structure.
  • the back portion 753 of the poly-optic lens 751 is curved in the shape of the collimating lens 717 of Fig. 37B.
  • the array of smaller lenses 711 from Fig. 37B appear as individual lens elements 752 in Fig. 38A.
  • the individual lens elements 752 are staggered, so that varying amounts of optical material are located between each individual lens element 752 and the collimated lens backing 753.
  • This varying amount of optical material acts in a fashion similar to the staggered optical flats 719 of Fig. 37B, thereby providing for differing focal lengths and staggered field depths 755 along collimated scan lines 758.
  • the poly-optic lens 751 of Fig. 38A therefore has a composite depth of field similar to that of Fig. 37B.
  • Another embodiment of an optical structure is shown in Fig. 38B.
  • the Fig. 38B structure is similar to that of Fig. 37C, except that some of the individual elements of Fig. 37C have been incorporated into a single element.
  • a poly-optic lens 761 incorporates aspects of the large collimating lens 727, the GRIN element 729, and the array of smaller lenses 721 shown in Fig.
  • the back portion of the poly-optic lens 761 is curved in the shape of the collimating lens 727 of Fig. 37C.
  • the array of smaller lenses 721 from Fig. 37C appear as individual lens elements 762 in Fig. 38B.
  • a GRIN ele ⁇ ment 764 is embedded between the individual lens elements 762 and the collimated lens backing 763, thereby providing for differing focal lengths and staggered field depths 765 along collimated scan lines 768.
  • the poly-optic lens 761 of Fig. 38B therefore has a composite depth of field similar to that of Fig. 37C. While Figs.
  • FIG. 39A is shown from a side view of the CCD detector, the CCD detector may also be rotated 90 degrees so that the lenses or other light collection elements are horizontally rather than vertically stacked, and so that the photodiodes receive the collected light in a manner similar to that shown in Fig. 35A. In general, horizontal stacking results in greater light collection, but vertical stacking results in less interference between lenses or lens elements.
  • Figures 39A and 39B are alternative embodiments of optical structures using curved mirrors or similar ele ⁇ ments.
  • the optical system of Fig. 39A comprises an array of curved (i.e., focusing) mirrors 790 positioned in a diagonal line 791 at an angle ⁇ with respect to a CCD detector 792.
  • FIG. 39A The array of curved mirrors with different focal lengths or different optical path lengths provide staggered field depths 793 along collimated scan lines 794 as shown.
  • the structure of Fig. 39A minimizes the effects of chromatic aberration.
  • Figure 39B shows a similar optical structure to Fig. 39A, but utilizes a single multi-faceted curved surface 796 rather than array of curved mirrors 790.
  • the single multi-faceted curved surface 796 is positioned at a diagonal angle with respect to the CCD detector 792, and thereby provides staggered field depths 797 in a manner similar to the Fig. 39A structure along collimated scan lines 798 as shown.
  • the Fig. 39B structure also minimizes the effects of chromatic aberration.
  • FIG. 40A-40D Another embodiment of a poly-optic lens system is shown in Figs. 40A-40D.
  • Figure 40A is a diagram of a side view of a poly-optic lens 780 positioned in front of a CCD detector 782. As shown, light from a distant point is collected by the poly-optic lens 780 and focused on the CCD detector 782.
  • Figure 40B is a diagram showing a top view of the same poly-optic system. Light from distant points on each of the collimated scan lines is collected by each individual lens element 783, and focused on the CCD detector 782. Preferably, each of the lens elements 783 has a different focal length.
  • Figure 40C is a diagram showing a top view of the same poly-optic system of Fig.
  • FIG. 4OB shows a target scan line 784 for several of the individual lens elements 783.
  • Light reflected from a target scan line 784 passes through an axial point 785 in an individual lens element 783, as shown, and is projected onto the CCD detector 782.
  • the target scan line 784 for each individual lens element 783 is slightly offset from the other target scan lines 784, due to the relative positions of the lens elements 783.
  • Figure 40D shows an oblique view of the same target scan line capture as Fig. 40C.
  • a preferred multi-focal lens is depicted in Figs. 41A-41C.
  • An oblique view of a preferred lens 801 is shown in Fig. 1A.
  • the lens 801 may be circularly symmetric, as shown.
  • the back surface 801a of the lens 801 is preferably spherical, while the front surface 801b has four different aspheric curvatures as described below.
  • the lens 801 of Fig. 41A may be provided with a rectangular opaque mask which provides a similar function to an aperture. Alternatively, the effect of the aperture can be obtained by cutting the lens 801 to obtain a cut- away lens 810, as shown in Figs. 4IB and 41C.
  • Figure 4IB shows an oblique view of the cut-away lens 810, while the same cut-away portion is shown from a top view in Fig. 41C.
  • the cut-away angle is selected so as to preferably allow for a 45-degree field angle.
  • the lens 801 comprises four zones 802, including a center zone 803 (Zone 1) and three outer ring-shaped zones 804, 805 and 806 (Zones 2 through 4, respectively) .
  • Zone 1 is preferably 0.426 inches wide, while Zones 2 through 4 are preferably 0.213 inches wide.
  • the base radius i.e., radius of the back spherical portion of the lens
  • the distance from the lens 801 to an imager is preferably 34.85272 millimeters. Preferred focus distances appear in Table 41-1.
  • the material of the lens 801 preferably comprises a plastic such as polystyrene, with an index of refraction of 1.59453059. Following are preferred dimensions and other characteristics according to lens zone 802.
  • a preferred aperture width is 0.043 inches.
  • Another embodiment of a multiple-focus optical system comprises a poly-optic lens wherein the longer focal length zones have a restricted angular field of view, as may be accomplished by intentionally leaving some of the aberrations of the zone uncorrected, thereby restricting 56 the angular field of view for that zone over which a clear image of the barcode is produced on the CCD detector. The effect of this structure is shown in Fig. 57, which depicts a poly-optic lens 1160 having three zones and positioned in front of a CCD detector 1161.
  • a poly-optic lens 1160 By selectively leaving coma, astigmatism, and lateral chromatic aberrations uncorrected, a poly-optic lens 1160 is provided having three different fields of view 1162, 1163 and 1164, one for each zone.
  • the field of view 1162 for the first zone is wide to allow reading of wide images
  • the poly-optic lens 1160 of Fig. 57 offers the advantage of minimizing inadvertent reading of bar codes or other images far from the bar code reader, as well as improving the aberration correction in the regions where reading is desired thereby improving resolution.
  • Figures 42A-42C are alternative embodiments of optical structures using shutters to reduce or eliminate inter-zonal interference. These embodiments use either mechanical or electro-optic shutters so that only one (or several) lenses are active simultaneously.
  • Figure 42A shows an arrangement of lenses 830 in a circular pattern, forming a poly-optic array.
  • the lenses 830 are positioned in front of a CCD detector 832.
  • a mechanical shutter 834 is positioned in front of the lens array 830.
  • the shutter 834 may be rotated mechanically around a motor shaft (not shown) .
  • the shutter 834 has an opening 835 (or aperture) which revolves in a circular fashion in response to a motive force applied to the motor shaft. In operation, the shutter 834 screens out all of the lenses 830 except one lens 830 (or several lenses 830, if desired) which collects light through the opening 835 in the shutter 834.
  • FIG. 42B is a diagram of an alternate embodiment employing a shutter mechanism.
  • a poly-optic array of lenses 840 are arranged in a linear fashion.
  • the lenses 840 are positioned in front of a CCD detector 842.
  • a rectangular or other appropriately shaped shutter 844 is positioned in front of the array of lenses 840.
  • the shutter 844 has an opening 845 (i.e., aperture) enabling a single lens 840 (or several lenses 840, if desired) to operate at a given time.
  • the shutter 844 is movable along the axis 846 parallel to the array of lenses 840, and may be driven by a solenoid, a linear motor, or a mechanical cam (not shown) . In such a manner, the position of the opening 845 may be shifted from one lens 840 to the next. In operation, the shutter 844 is driven along the axis 846 parallel to the lenses 840 so as to position the opening 845 in front of the desired lens 840 or lenses 840 in sequence. In this manner, the lenses 840 are prevented from interfering with one another.
  • Figure 42C is a diagram of another embodiment employing a shuttering technique.
  • a poly- optic array of lenses 850 forms a circular cluster pattern as shown, or may form any other suitable pattern matched to a particular LCD shutter configuration.
  • the LCD array 854 is positioned in front of the array of lenses 850.
  • the lenses 850 are positioned in front of a CCD detector 852.
  • the LCD array 854 comprises a plurality of elec ⁇ tronically operated shutters 855. Each shutter 855 is individually responsive to an electronic shutter command from a shutter controller (not shown) .
  • An advantage of the LCD array 854 is that it has no moving parts, and therefore may increase reliability.
  • the above described shuttering techniques may be implemented with a multi-focal plane optical collection system comprising, for example, either a single poly-optic element or a multi-element optical system.
  • Figs. 23B and 23C which, as noted previously, are graphs plotting minimum renderable bar size against distance from a lens. If all the lenses in the previously described poly-optic arrays of Figs. 42A-42C operate simultaneously, then dead zones 472 such as shown in Fig. 23C may occur in the composite depth of field due to inter-zonal interference, such that a target image may be misidentified or missed entirely by the reading device.
  • the shuttering techniques of Figs. 42A-42C provide for sequential operation of the individual lenses, thereby minimizing or eliminating inter-zonal interference.
  • the depth of field for each lens as shown in Fig.
  • the composite depth of field becomes the aggregate of the individual depths of field, without the effect of inter-zonal interference.
  • the result is a much larger effective depth of field than is possible with only a single lens, or without shuttering.
  • FIGs. 56A and 56B Another embodiment of a multi-focus optical system in which inter-zonal interference is reduced is shown in Figs. 56A and 56B.
  • the optical system as shown in Fig. 56B comprises a multi-element lens system 1155 and a plurality of photo-sensitive detectors such as linear CCD arrays 1156.
  • the multi-element lens system 1155 comprises three lens elements 1155a-c, each having a different focal length and different field of view.
  • a comparison of the respective fields of views 1151 through 1153 are shown in Fig. 56A.
  • the closest field of view 1151 is associated with the first lens element 1155a; the middle field of view 1152 is associated with the second lens element 1155b; and the most distant field of view 1153 is associ ⁇ ated with the third lens element 1155c.
  • the lens elements 1155a-c may be embodied as a single poly-optic lens, or as three separate lenses adjacently positioned. While three lens elements 1155a-c are shown, more or fewer lens ele ⁇ ments may be used as needed depending on the total depth of field requirements. In operation, each lens element 1155a-c focuses light on a different one of the three linear CCD arrays 1156a-c.
  • the first lens element 1155a has a focal length Fl and focuses light on a first linear CCD array 1156a; the second lens element 1155b has a focal length F2 and focuses light on a second linear CCD array 1156b; and the third lens element 1155c has a focal length F3 and focuses light on a third linear CCD array 1156c.
  • Each linear CCD array 1156a-c thereby receives focused light from only one of the lens elements 1155a-c, and receives an image from only one focal plane.
  • This technique achieves a result of non-interference between lens elements 1155a-c similar to the shuttering techniques described above with respect to Figs. 42A-42C, with the advantage of shorter image reading times since all of the focal planes are imaged simultane- ously or in parallel by the linear CCD arrays 1156a-c.
  • the poly-optic and shuttering lens techniques described are preferably used in conjunction with the aperturing techniques described elsewhere herein. Some of the various embodiments may be compared in speed and depth of field characteristics as follows. Use of a small, oblong aperture in connection with a single-focus increases speed over a small, round aperture, as the oblong aperture collects more light. Both the oblong aperture and the round aperture result in a relatively good depth of field. Use of a poly-optic lens configura ⁇ tion with large apertures results in even faster opera ⁇ tion, but may result in a "perforated" field depth due to intra-lens or inter-lens interference.
  • the perforation in the depth of field may be eliminated by using shutters or multiple CCD arrays (each associated with an independent lens element, as described previously) .
  • Use of a poly- optic lens configuration with oblong apertures results in even faster operation, and also may result in a perforated field depth.
  • shutters or multiple CCD arrays with separate lenses may be used to reduce or eliminate the perforation in the depth of field.
  • a CCD detector used in conjunction with some of the above lens and/or aperture configurations may be used at a small fraction of saturation (e.g., 1% of full scale) in order allow it to run faster (i.e., achieve faster reading speeds) .
  • correlated double sampling may be used to lower the noise level in the system.
  • Correlated double sampling may also be beneficial where there is no self-generated illumination source being used because the incoming signal level may be reduced.
  • a CCD detector may use ambient light to read images such as bar code labels.
  • optical techniques described herein may be used in whole or part in conjunction with the adjustable-focus techniques described in U.S. Patent Nos. 5,317,211 and
  • the optical system can in a variety of embodiments be configured to allow multidirectional imaging using a single detector array. Multidirectional scanning is particularly desirable in fixed reader applications where the relative orientations of bar code and reader may vary widely.
  • two cylindrical optics such as cylindrical lenses 1172 and 1174 are used to image the bar code 1175.
  • a cylindrical optic also called a cylindrical rotational filter
  • the cylindrical lenses 1172, 1174 are preferably oriented so that their axes are sub ⁇ stantially perpendicular to one another (i.e., ortho ⁇ gonal) , so that one cylindrical lens 1174 produces an image well resolved in the horizontal direction while the other cylindrical lens 1172 produces an image well resolved in the vertical direction.
  • the light collected by the horizontal axis cylindrical lens 1174 is rotated approximately 90° by a plurality of image-rotating mirrors 1176 and then combined with the light from the vertical axis cylindrical lens 1172 by a beamsplitter 1178.
  • the number of cylindrical optics is not limited to two, but may be any suitable number; likewise, their angle of orientation with respect to one another need not be 90° but may be any suitable angle.
  • the cylindrical optics may be reflective rather than refractive elements, or combina- tions of cylindrical and non-cylindrical optical elements may be used to modify the acceptance angles of the rota ⁇ tional filters.
  • the focal distances of a multi- axis system may be offset such that different orientation bar codes may focus at different distances away from the reader for the purpose of enhancing depth of field.
  • a simple one-axis filter is shown in Fig. 59 which may be used singly or multiply to detect when a bar code is aligned in a particular orientation for the purpose of, e.g., controlling other bar code reading equipment which utilizes such information.
  • the one-axis filter of Fig. 59 comprises a vertical axis cylindrical lens 1181 positioned in front of a linear CCD detector 1180.
  • a bar code 1182 is oriented along the cylindrical axis of the cylin ⁇ drical lens 1181, the bar code image is at its best resolution.
  • the signal output from the CCD detector 1180 can be compared against a threshold value to determine when a bar code 1182 is properly oriented in front of the CCD detector 1180 so that it may be read.
  • FIG. 60A-60C Another embodiment of an optical system allowing variously oriented barcodes to be read is shown in Figs. 60A-60C, with various alternative area and linear sensor configurations shown in Figs. 60D-60F.
  • Figure 60A depicts an oblique view of the optical system
  • Fig. 60B depicts a top view
  • Fig. 60C depicts a side view.
  • the embodiment of Figs. 60A-60C comprises a beamsplitter cube 1202 (or alternatively a pellicle beamsplitter or other similar device), a pair of mirrors 1203 and 1204, and imaging regions 1215, 1216 on which are mounted one or more imaging elements (such as shown in any of Figs. 60D- 60F) .
  • the beamsplitter cube 1202 comprises a pair of triangular prism-shaped beamsplitters joined along their diagonal sides, so as to form a cube shape as shown in Fig. 60A.
  • the beamsplitter cube 1202 splits the incoming bar code image into two paths 1207 and 1208.
  • the first path 1208 continues in a vertical line and reaches imaging region 1216.
  • the second path 1207 reaches mirror 1203 which causes a 90° rotation of the image, and then reaches mirror 1204 which causes another 90° rotation of the image so that it also is directed in a vertical line 1209, parallel to the first path 1208.
  • the second vertical path 1209 reaches imaging region 1215.
  • the net effect of the optical structure is shown in Fig. 60B, which shows two versions of the image rotated at 90° with respect to one another, and each respectively being directed down one of the vertical paths 1208 and 1209.
  • Figure 60D depicts two CCD area sensors (i.e., two-dimensional CCD arrays) 1217, 1218 located on imaging regions 1215 and 1216, respectively.
  • the CCD area sensors 1217, 1218 allow detection of a bar code or other image by any of a number of manners well known in the art.
  • Figure 60E depicts an alternative arrangement wherein two CCD linear arrays 1219, 1220 are disposed on imaging regions 1215 and 1216, respectively.
  • the two CCD linear arrays 1219, 1220 are likewise used to detect the received image.
  • the CCD linear arrays 1219, 1220 are lined up in parallel because the image projected on the first CCD linear array 1219 will be oriented at 90° with respect to the image projected on the second CCD linear array 1220; placing them in parallel therefore increases the chance that the image will be properly oriented on at least one of the CCD linear arrays 1219, 1220.
  • Figure 60F depicts another alternative arrangement wherein a single CCD linear array 1221 straddles both imaging regions 1215, 1216 along their common diagonal.
  • the Fig. 60F embodiment generally requires a larger CCD linear array 1221 than the CCD linear arrays 1219, 1220 of the Fig. 60E embodiment, but may require less processing or fewer components because only one CCD linear array needs to be read out and processed.
  • FIG. 61A Another multi-directional optical system is shown in Fig. 61A.
  • the optical system 1225 receives light reflected from an image (e.g., a bar code) and rotates the image 90° by using an angle adjustment mirror 1226.
  • the rotated image is then reflected off a mirror basket 1228 which divides the single projected image into several segments, resulting in a multi-segment pattern projected onto a photodetector 1227 (such as one or more linear CCD array(s) or CCD area sensor) .
  • a photodetector 1227 such as one or more linear CCD array(s) or CCD area sensor
  • an imaging line from the photodetector 1227 is reflected off the mirror basket 1228 and divided into several segments, resulting in a star shaped pattern.
  • the star shaped pattern 1229 is rotated 90° by the angle adjustment mirror 1226 and projected into the imaging volume in front of the bar code reader to create multi ⁇ directional virtual imaging lines across an image to be read.
  • the mirror basket 1228 preferably has a first mirrored region 1230 and a pair of second mirrored regions 1231 attached to either side of the first mirrored region 1230 and bent therefrom by an angle of 30°.
  • the mirror basket 1228 is positioned at a 45° angle with respect to the horizontal in order to rotate the image 90° while at the same time generating the star shaped pattern 1229.
  • Fig. 61A involves several compo- nents in the collection optics, it offers the advantage that the images associated with the variously oriented virtual imaging lines are projected onto different seg- ments of the photodetector 1227, reducing or eliminating the possibility of interference between different imaging lines.
  • a mirror basket can be particularly useful in the embodiment of Fig. 6IB, whereby a handheld bar code reader 1235 which has a single virtual imaging line in normal handheld operation can be made to have multiple virtual imaging lines in a fixed configuration.
  • the handheld bar code reader 1235 is placed in a cradle (not shown) adjacent to a fixed optical projection unit 1236.
  • the optical projection unit 1236 comprises an angle adjustment mirror 1238 and a mirror basket 1237 (e.g., such as mirror basket 1228) positioned at a 45° angle for rotating an image (such as from an object held above the fixed optical projection unit 1236) and projecting it (in a star-shaped pattern) onto the imaging sensor within the handheld bar code reader 1235.
  • the optical projection unit 1236 thereby gives the handheld bar code reader 1235 effective multi-directional imaging capability, even if the handheld bar code reader 1235 is only a single-line imaging device.
  • a multi-direction reader which allows multi-directional reading employs a kaleido ⁇ scope mirror for projecting multiple rotated images of a bar code onto a detector array, such as shown in Figs. 62A-62F.
  • a kaleidoscope mirror 1250 comprises two mirror elements 1253, 1254 connected along one edge and oriented at a 45° angle with respect to one another, with the reflective portions of the mirror elements 1253 and 1254 facing one another.
  • An image 1251 projected into the "tunnel" formed by the kaleidoscope mirror 1250 causes a plurality of images to be generated at a detector 1252 at the opposite end of the kaleidoscope mirror, as shown in Figs.
  • a plurality of imaging "zones" 1259 are defined over the region of the detector 1252, each having a version of the original image 1256 oriented at a different direction.
  • the kaleidoscope mirror embodiment of Fig. 62A-62B has the advantage of allowing relatively simple signal processing of the CCD area sensor output without requiring complex algorithms for determining the image orientation.
  • the data from the CCD area sensor may be read out linearly in a fixed direction (i.e., line-by-line), and there is a high probability that one of the duplicate projected images 1257 will be at a sufficiently horizontal angle with respect to the CCD area array as to allow recognition of the image (such as, e.g., a bar code).
  • Figures 62D-62F are examples showing some of the various relative orientations that the duplicate projected images 1257 may take.
  • the straight arrow symbol "-->" in Figs. 62D-62F represents the angle of orientation of each duplicate projected image.
  • the original image 1263 is oriented in a vertical direction with respect to the CCD area sensor, but a duplicate image 1261 is projected at a horizontal orientation by the kaleidoscope mirror 1250.
  • the original image 1264 is oriented at a lesser vertical direction (e.g., about 67°), and another duplicate image 1262 from a different image zone 1259 is directed in a close to horizontal direction so as to allow recognition using the CCD sensor array.
  • Figure 62F shows the "worst case” scenario in which the original image is oriented at 45° with respect to the horizontal direction.
  • all of the duplicate projected images 1257 are about equally difficult to recognize using the CCD sensor array, and the ability to recognize any of the projected images 1257 will depend in part on the tolerances of the signal processing and recognition electronics and intelligence.
  • For a bar code reader to function properly with ambient illumination sufficient distance should prefer ⁇ ably be maintained between the bar code and the optical system to allow ambient light to reach the bar code. If the reader and bar code are brought too close together, the bar code may be shadowed to such an extent by either the reader or the object bearing the bar code that it becomes unreadable in ambient light.
  • Fig. 63 comprises an optically transmissive spacer 1303 extending forward from the housing 1301 which encases the optical system.
  • the optically transmissive spacer 1303 ensures that a minimum adequate distance is maintained between the reader and the bar code, while at the same time allowing ambient light to reach the bar code.
  • the optically transmissive spacer 1303 may be transparent, or may comprise opaque or solid portions with missing gaps (e.g., two opaque or solid walls and two missing walls) .
  • Ambient light 1311 entering through the transparent or opaque material, or the gaps, of the optically transmissive spacer 1303 reflects off an optically scattering surface 1306 (which is a more reflective surface) in front of the housing 1301 and thereby results in ambient light 1312 being reflected away from the housing 1301 and onto the bar code 1310.
  • the ambient light reflected from the bar code 1310 is received into the receiver aperture 1305 and collected by the optical system.
  • the optically transmissive spacer 1303 may optionally be configured with a transparent window 1304 across its front end (i.e., the end opposite the housing 1301). In hand-held readers, the optically transmissive spacer 1303 may take the form of a transparent "nose" on the front end of the reader.
  • an opaque "nose” may be employed in which sections of the nose are missing to allow entrance of ambient light into the volume between the reader and the bar coded object 1309.
  • the window 1304 is not required, but never ⁇ theless is optional to provide protection.
  • a transparent window 1304 is preferred so as to keep objects off of the optical front-end.
  • the optically transmissive spacer 1303 mechanically prevents the bar coded object 1309 from being brought too close to the reader to cut off all ambient light, and the transparent, translucent, or missing portions of the optically transmissive spacer 1303 allow ambient light 1311 to enter the space between the bar coded object 1309 and the optical collection system and illuminate the affixed bar code 1310.
  • the spacer 1303 also limits the required range for depth of focus.
  • a second structure for increasing ambient light at close range or for shaded bar codes is an optically diffusive front surface placed on the barcode reader (excluding the optical collection system) , such as shown for example in Fig. 64A.
  • the optically diffusive front surface 1320 shown in Fig. 64A may be fabricated from a variety of materials, including, but not limited to, roughened white plastic, roughened clear plastic, or brushed metal.
  • the optically diffusive front surface 1320 scatters a portion of the incident ambient light back in the direction of the bar code 1321, thereby increasing the level of illumination at the bar code 1321.
  • FIG. 64C compares a situation wherein an ambient light source 1330 hits the back of a bar coded object 1331, thereby causing the bar code 1332 to be shaded with respect to the bar code reader 1335 and preventing sufficient ambient light to allow the reader to properly operate.
  • the optical collection and imaging system can in one aspect be regarded as projecting one or more images of the detector array(s) into the imaged volume of space in front of the barcode reader, creating one or more virtual imag ⁇ ing lines.
  • one of these imaging lines is preferably projected across the bar code. This targeting can become increasingly difficult as the depth of field of the bar code reader is lengthened by, for example, use of the multi-focal imaging techniques described herein.
  • one or more LEDs or similar devices may be used as pointers.
  • a preferred LED pointer system is depicted in Figs. 65A and 65B, which are side view and front view diagrams, respectively.
  • the Fig. 65A-65B embodiment comprises an LED 1350
  • a housing wall 1352 with an aperture or slit 1351 a holographic (or other type of) diffuser 1353 positioned so as to cover the slit 1351, and a plastic lens 1355 positioned in front of the slit 1351 and in alignment with the LED 1350.
  • a holographic (or other type of) diffuser 1353 positioned so as to cover the slit 1351
  • a plastic lens 1355 positioned in front of the slit 1351 and in alignment with the LED 1350.
  • light from the LED 1350 is diffused by the holo ⁇ graphic diffuser 1353 and thereby uniformly illuminates the area of the slit 1351.
  • the plastic lens 1355 which serves as a pointer optic or optical system to image the light from the illuminated slit 1351 at a point which coincides in space with the projected imaging line.
  • the LED output may otherwise appear nonuniform and may generate an image of the LED chip (including the bond wires) used to support the LED 1350.
  • two pointer LEDs such as LED 1350 are employed, one on either side of the imager (as per Fig. 46A or 48) , or alternatively positioned above or below the imager.
  • the pointer LEDs are not used to illuminate the bar code for reading, but merely to assist the bar code reader operator in properly positioning the bar coded object with respect to the reader. The proper positioning can be recognized by the appearance of a more focused image of the pointer slit 1351 near the bar code.
  • Figure 65C is a diagram illustrating various pointer slit shapes at different distances. As shown in Fig.
  • the illumi- nated pointer slits appear as relatively small, spaced apart dots 1360 at near field, as larger oblong more closely spaced shapes 1361 at mid field, and as a single larger oblong shape 1362 at far field.
  • the pointer LED(s) 1350 are preferably cycled on and off corresponding to the image capture cycle, being off while an image of the bar code is being captured by the imager so the pointer beams do not interfere with the image detection.
  • an optical notch filter (not shown) may be used to prevent backscattered light from the LED pointer from reaching the CCD detector array.
  • one alternative embodiment employs at least two pointers focused at different points in space.
  • Another alternative embodiment employs a compound para ⁇ bolic concentrator optical assembly (not shown) instead of a diffuser 1353.
  • Several other alternative embodiments are particularly directed to increasing the efficiency of coupling of the LED light through the pointer optical system.
  • One such embodiment comprises a cylindrical lens (not shown) positioned in front of the LED 1350 but behind the diffuser 1353 and a slit shaped pointer aperture 1351, the cylindrical lens serving to direct a larger fraction of the light emitted by the LED 1350 onto the pointer aperture 1351 (thereby reducing the LED power needed) .
  • Another embodiment wherein a similar objective is achieved comprises a light pipe (not shown) covering the LED 1350 and shaping the light to match the shape of the pointer aperture 1351.
  • the light pipe may have a diffusing end so that it could also replace the diffuser 1353.
  • an area imager may be provided having increased speed in various lighting conditions, or being capable of operation in ambient light without a self- generated illumination source.
  • the multi ⁇ directional reading techniques using a multi-dimensional CCD area sensor and virtual imaging lines can be employed in the embodiments described herein using CCD area sensors.
  • the CCD signal is pro- Defined in a variety of manners.
  • the CCD output- signal itself may be used as part of an auto-exposure or adaptive exposure system for adjusting the exposure time of the CCD detector, as further described herein.
  • the CCD output signal may also be processed to determine the locations of bars and spaces and recognize bar codes or other symbols, as also discussed further herein.
  • Figure 43A depicts one embodiment of an adaptive exposure system in accordance with one or more aspects of the present invention.
  • Figure 43B is a diagram of exemp- lary waveforms in accordance with the Fig. 43A embodiment.
  • a pulse train oscillator 900 comprising a Schmitt trigger U7A having an input node 901 connected through a capacitor C9 to ground 902.
  • the input node 901 of the Schmitt trigger U7A is connected through a diode D3 and a resistor R26 positioned in series to the output node 903 of the Schmitt trigger U7A.
  • the input node 901 is also connected to another diode D2 configured in the opposite direction of diode D3, and thereafter to a resistor R22 and then to the output node 903 of the Schmitt trigger U7A.
  • the output of the pulse train oscillator 900 at point A comprises a periodic waveform 920 such as shown in Fig. 43B.
  • the periodic waveform 920 may have a period of, for example, 30 milliseconds and a 200 microsecond negative- going pulse 921.
  • Capacitor C9 may have a capacitance of 0.1 ⁇ F
  • resistor R22 may have a resistance of 910 ohms
  • resistor R26 may have a resistance of 4.3 k ⁇ .
  • the output pulse of waveform 920 is generally positive when capacitor C9 is charging through resistor R22, and is negative when the capacitor C9 discharges through resistor R26.
  • the pulse train oscillator 900 is connected to a monostable multivibrator 905.
  • the monostable multi ⁇ vibrator 905 comprises a capacitor C2 connected to one end of a photodiode Dl as shown.
  • the other end of the photo ⁇ diode Dl is connected to ground 902.
  • the photodiode Dl is preferably located adjacent to the CCD detector or other ⁇ wise nearby so as to be able to detect the amount of light being received from the optical system.
  • the capacitor C2 is connected to the input node 906 of a Schmitt trigger U7B. Capacitor C2 may have a capacitance of 0.02 ⁇ F.
  • the pulse train oscillator 900 drives the capacitor C2 and the photodiode Dl.
  • the waveform 924 at point B and the resulting output waveform 926 from the monostable multivibrator 905 at point C are shown in Fig. 43B.
  • the conductance of the photodiode Dl controls the output pulse width 927 of the monostable multivibrator 905, and therefore controls the duty cycle of oscillation at point C.
  • the output waveform 926 at point C will have a relatively narrow negative pulse (e.g, almost 100% duty cycle) .
  • the output waveform 926 will have a relatively wide negative pulse (e.g., near 0% duty cycle) .
  • Monostable multivibrator 908 comprises a capacitor C13 at its input .
  • the capacitor C13 is shunted by a resistor Rll to ground, and is connected through another resistor R13 to a Schmitt trigger U7F.
  • Capacitor C13 may have a capacitance of 0.001 ⁇ F, and each of resistors Rll and R13 may have a resistance of 1 k ⁇ .
  • the other monostable multivibrator 910 is configured in a similar manner with capacitor CIO, resistors R23 and R25, and Schmitt trigger U7D, except that resistor R25 has a larger resistance than resistor Rll, such as 3.3 k ⁇ .
  • monostable multivibrator 908 is con ⁇ nected to another monostable multivibrator 912, configured in a similar manner with capacitor C7, resistors R8 and R9, and Schmitt trigger U7C.
  • the output of monostable multivibrator 912 at point F is read-out gate (ROG) signal 913, the function of which is explained below.
  • Monostable multivibrators 908 and 912 preferably each produce a fixed pulse of about 1 microsecond in duration, while monostable multivibrator 910 preferably produces a fixed pulse of about 3 microseconds in duration.
  • the shutter signal 915 at point H is a function of the outputs of both monostable multivibrators 905 and 910.
  • the output of multivibrator 905 at point C and the output of multivibrator 910 at point G are diode OR'ed together through diodes D4 and D5, respectively, to a shutter out ⁇ put pulse stage 914.
  • the shutter output pulse stage 914 comprises a transistor network, including a transistor Ql whose base 916 is coupled to each of diodes D4 and D5 through a resistor R27.
  • Diodes D4 and D5 are connected to the reference voltage 917 (e.g., 5 volts) through a resistor R28 and a capacitor C3 connected in parallel.
  • the collector 918 of transistor Ql which produces the shutter signal 915 at point H, is connected through resistor R30 to the reference voltage 917.
  • the base 916 of the transistor is connected to ground 902 through resistor R29, while the emitter 919 of the transistor Ql is connected directly to ground 902.
  • the shutter signal 915 (shown as waveform 938 in Fig. 43B) comprises a series of pulses 939 which transition from a low state to a high state when the negative pulse of monostable multivibrator 905 begins, and stay high until the negative pulse of monostable multivibrator 910 times out.
  • the effect is that of a variable duty cycle oscillator, wherein the duty cycle varies according to the light received by the photodiode Dl.
  • the actual exposure period 940 as shown in Fig. 43B is somewhat less than the duration of the pulses 939 in the shutter signal 938.
  • the exposure period 940 is commenced at the start of each pulse 939 of the shutter signal 938, but is terminated by the ROG signal pulse 935 as shown in waveform 934 of Fig. 43B.
  • the ROG pulse 935 terminates the exposure period 940 by causing a transfer of charge from the pixels in the CCD array into the CCD shift registers.
  • the shutter signal 938 is high, the pixels collect light.
  • the shutter signal 938 is low, the pixels are shorted; however, the ROG pulse 935 generally terminates the collection of light before the shutter signal 938 transitions from high to low.
  • Monostables 908, 910 and 912 ensure that the ROG pulse 935 terminating the exposure period 940 falls cleanly just before the end of the shutter pulse 939 and at a time when clocking to the CCD array is disabled.
  • Monostable 910 establishes a 3 microsecond clock disable period 937, at the end of which the shutter pulse 939 generally terminates.
  • monostable 908 establishes a 1 microsecond delay 933, after which the ROG pulse 935 is generated by monostable 912.
  • the output of monostable 910 disables clocking to the CCD array at the time when the ROG pulse 934 occurs.
  • the adaptive exposure circuit of Fig. 43A thus provides an exposure period which varies based on the amount of light. When there is more light, the exposure period 940 is smaller, and when there is less light the exposure period 940 is larger (up to, e.g., 30 milliseconds) .
  • Figures 44A and 44B show alternative configurations of an optical system employing an adaptive exposure circuit.
  • a CCD chip 950 produces a video output signal 956 which is connected to a pre-amplifier 951.
  • the pre-amplified signal 952 is connected to a peak detector 954, which measures the peak-to-peak value of the pre- amplified video signal 952.
  • the peak-to-peak signal 957 is provided to an exposure control circuit 955 which adjusts the exposure period for the CCD chip 950.
  • the exposure control circuit 955 is configured in a similar manner to the exposure control circuit of Fig. 43A, but uses the peak-to-peak signal 957 to control the exposure time rather than the collected light as measured by a photodiode Dl.
  • Figure 44B shows an alternative adaptive exposure technique that is similar to the technique of Fig. 44A, but the peak-to-peak signal level of the first derivative of the video signal is used to control the exposure time of the CCD chip instead.
  • a CCD chip 960 produces a video output signal 967 which is connected to a pre-amplifier 961.
  • the pre-amplified signal 966 is connected to a signal processing block 962.
  • the signal processing block 962 generates a first derivative of the pre-amplified signal 966.
  • the first derivative signal 963 is connected to a peak detector 964, which measures the peak-to-peak value of the first derivative signal 963.
  • the peak-to-peak signal 968 is provided to an exposure control circuit 965 which adjusts the exposure period for the CCD chip 960.
  • the exposure control circuit 965 is configured in a similar manner to the exposure control circuit of Fig. 43A, but uses the peak-to-peak signal 968 generated from the first derivative of the video signal to control the exposure time rather than the collected light as measured by a photodiode Dl.
  • transitions in the video signal may be detected by any of a number of means as known in the art. For example, transitions in the video signal may be detected by taking a second derivative of the video signal, and finding zero crossings of the second derivative signal during selected timing intervals.
  • One such technique is described, for example, in U.S. Patent No. 4,000,397 entitled "Signal Processor Method and Apparatus" issued in the name of Hebert et al., which patent is hereby incorporated by reference as if fully set forth herein.
  • a second-derivative detection circuit is shown in Fig. 45A.
  • Corresponding waveforms for the Fig. 45A circuit are shown in Fig. 45B.
  • a low pass reconstruction filter 971 receives a video input signal 972 from a CCD chip 973.
  • the video input signal 972 may comprise a plurality of stepped voltage levels, like stair steps, corresponding to the contents of the CCD array, such as shown in waveform 990 of Fig. 45B.
  • the video input signal 972 may be passed through an optional gain stage 974, before being fed into the low pass filter 971.
  • the low pass reconstruction filter 971 smooths the video input signal 972 into a smoother, more continuous waveform, such as waveform 992 of Fig. 45B.
  • the low-pass filtered signal 975 is coupled to a differentiator circuit 976, which generates a first derivative 979 of the low-pass filtered signal 975.
  • the first derivative 979 is coupled to another differentiator circuit 978, which takes another derivative, thereby gen ⁇ erating a second derivative 980 of the low-pass filtered signal 975.
  • the first derivative signal 979 corresponds to waveform 994 of Fig. 45B, while the second derivative signal 980 corresponds to waveform 996 of Fig. 45B.
  • the first derivative signal 979 is also coupled to a peak detector circuit 981, which detects the peak value of the first derivative signal 979 and generates a threshold signal 977 thereby.
  • the threshold signal 977, the first derivative signal 979, and the second derivative signal 980 are provided to a comparator logic block 982.
  • the comparator logic block 982 detects zero crossings of the second derivative signal 996 (as shown in Fig. 45B) by comparing the second derivative signal with a zero voltage level. The zero crossings are considered valid transi ⁇ tions if they occur when the first derivative signal 979 exceeds the level (positive or negative) of the threshold signal 977.
  • the comparator logic block 982 compares the first derivative signal 979 with the threshold levels (positive and negative) 987, 988 defined by threshold signal 977 and determines a valid transition when the second derivative signal 996 crosses the zero level during a time when the first derivative signal exceeds the level of either threshold signal 987, 988.
  • a reconstructed bar signal waveform 998 such as may be obtained by the method described above is shown in Fig. 45B. The polarity of the bar signal waveform 998 depends on the sign of the first derivative signal 994 at the time of the second derivative zero crossing. Further details regarding second derivative detection techniques are described in U.S. Patent No. 4,000,397, cited above. Transitions in the video signal may also be detected by use of techniques described in U.S. Patent No.
  • Derivative-based signal processing performance can be enhanced by employing AC coupled signal processing tech ⁇ niques as described further below. Because derivatives are noisier than the original signal, derivative-based signal processors sometimes may require a signal to noise ratio on the order of 10 to 1 for optimal operation. Moreover, if the bar code is blurred or is misaligned with respect to the reader, the transitions between low and high signal levels may be gradual ("soft edges") rather than abrupt (“hard edges”), resulting in low amplitude derivative signals despite adequate signal levels.
  • a preferred embodiment of the signal processing electronics comprises an AC coupled sig ⁇ nal processing unit based on (1) a high pass FIR (finite impulse response) filter with a relatively low cutoff frequency; and (2) a comparator with hysteresis.
  • a block diagram of such a signal processor is shown in Fig. 67.
  • the Fig. 67 embodiment comprises a high pass FIR filter 140 1 and a hysteretic comparator 1402 which includes an operational amplifier 1405 with resistors 1406, 1407 and 14 08 having values selected to provide a suitable amount of hysteresis.
  • An input signal 1410 received from a detector array is connected to the FIR filter 1401 that removes the low frequency and DC bias (i.e., non-barcode signals) .
  • the comparator 1402 trips to the BAR (high) state if the output of the FIR filter 1401 is positive, and trips to the SPACE (low) state if the output of the FIR filter 1401 is negative.
  • a hysteresis band is used to keep noise from triggering the comparator 1402.
  • An optimal FIR filter 1401 consists of a Gaussian impulse response high pass filter, whose coefficients for an llth-order filter are tabulated and graphed in Figure 68B.
  • a simplified FIR filter may also be implemented in the form of a moving average filter. Coefficients for an exemplary llth-order moving average high pass FIR filter are tabulated and graphed in Figure 648.
  • the Fig. 68A filter has the advantage of being easily implemented in software.
  • the operation of the Fig. 67 signal processor depends in part on the speed of the bar code signal (i.e., time between transitions) , and is illustrated in Fig. 69A for bar widths approaching the time constant of the FIR filter 1401 (i.e., fast bar code signals).
  • the FIR filter 1401 acts as a DC restoration circuit so as to AC-couple the input signal 1410 (shown as waveform 1421 in Fig. 69A) around zero volts.
  • the amplitude of the coupled signal 1422 i.e., the output of the FIR filter 1401
  • the amplitude of the coupled signal 1422 is nearly as large as that of the input waveform 1421; therefore, there is no additional noise introduced by the action of the FIR filter 1401.
  • the hysteresis band 1424 of the comparator 1402, comprising positive hyster ⁇ esis threshold 1425 and negative hysteresis threshold 1426, is set to be just larger than the noise floor.
  • the noise floor is usually constant and predictable for a constant signal level and can therefore be set appropriately if the expected input signal level is known in advance, or otherwise may be set by experiment.
  • the comparator 1402 trips when the FIR filter output signal 1422 rises above the positive hysteresis threshold 1425. It resets when the FIR filter output signal 1422 drops below the negative hysteresis threshold 1426. The comparator 1402 will no longer switch states when the signal level remains within the hysteresis band 1424.
  • the resulting output signal 1428 is shown in Fig. 69A.
  • the waveforms are substantially different, as shown in Figure 69B.
  • the input signal 1431 has wider spacing between transitions.
  • the FIR filter output signal 1432 looks similar to a derivative. However, the ampli ⁇ tude of the FIR output signal 1432 is still almost the same as that of the input signal 1431.
  • the comparator 1402 trips to a high state when the filter output signal 1432 is above one hysteresis threshold 1435.
  • the derivative-like FIR filter output signal 1432 declines toward zero quite quickly, but the comparator 1402 does not trip to its other state until another edge appears due to the effect of the hysteresis band 1434. When another edge appears, the FIR filter output signal 1432 passes the negative hysteresis threshold 1436, and the comparator 1402 trips to its other state.
  • the Fig. 67 circuit is capable of effectively reshaping input signal bars from nearly infi ⁇ nite width to a width on the order of the time constant of the FIR filter 1401.
  • the rise time of the analog input signal 1410 will generally increase.
  • the Fig. 67 circuit is still able to detect and respond to bar transitions.
  • FIG. 66 is a diagram of an exemplary signal processor using a single pole/zero high pass network.
  • the Fig. 66 embodiment comprises a high pass filter 1380, a buffer 1381, and a hysteretic compara- tor 1382.
  • the high pass filter comprises a capacitor 1386 and resistor 1387
  • the buffer 1381 comprises an opera ⁇ tional amplifier 1388 in a voltage-follower configuration
  • the comparator 1382 comprises an operational amplifier 1391 with resistors 1389, 1390 and 1392 having values selected to provide a suitable amount of hysteresis.
  • a possible drawback of the Fig. 66 embodiment is that it may show increasing amounts of bar distortion as the bar widths approach the filter time constant.
  • the decoder 107 interprets the transition data by any of a variety of means such as those well known in the art.
  • the decoder 107 may determine the relative widths of bars and spaces by measuring the relative time between each transition.
  • the decoder may use the bar and space measurement data to determine which of a variety of bar codes have been read, and/or the particular characters of the bar code. Further details regarding decoding may be found, for example, in U.S. patent application Serial No.
  • the decoding step is the slow step in the procedure. It is therefore desirable to minimize the amount of time spent by the decoder 107 trying to decode invalid signals. This time minimization may be accomplished in a preferred embodiment by performing preprocessing of image data prior to presentation to the decoder 107. By so doing, voltage 82 sequences which are unlikely to correspond to a valid bar code can be rejected without tying up the decoder 107.
  • Preprocessing may be performed very rapidly by a simple edge counter.
  • a typical bar code will have on the order of 30 edges.
  • a signal sequence may be classified by the edge counter as "interesting” if it exhibits more than a predetermined minimum number of edges, 20 or 25 for example. Those signal sequences deemed “interesting” are then read out more slowly to the decoder 107 for decoding.
  • data readout from a CCD detector or other type of photosensitive array is typically destructive--i.e. , the readout destroys the data.
  • the preprocessing edge counter therefore is preferably used in conjunction with a memory storage element such as an analog shift register to enable "rescanning", that is, a second look at the same data, this time by the decoder 107.
  • FIG. 72 A system including rescanning capability is shown in Fig. 72.
  • Rescanning is accomplished through the use of an analog shift register 1702 to store the image data from a CCD detector 1703. After exposure of the CCD detector 1703 to an image, the pixel charges are transferred to from the CCD detector 1703 the analog shift register 1702.
  • a rescan multiplexer 1704 selects the input from the CCD detector 1703. After an image is transferred to the analog shift register 1702, the image data is rapidly read out of the analog shift register 1702 for preprocessing with the "rescan" function enabled.
  • the "rescan" function causes the data from the analog shift register 1702 to be circularly queued, so that after each pixel charge is read at the output end of the analog shift register 1702, it is shuttled back to the other end of the analog shift register 1702 through rescan multiplexer 1704 as the remaining charges are shifted forward by one position.
  • the image data is provided to a signal processor (e.g., edge detector) 1705, which then provides its infor ⁇ mation to an edge counter 1706 to perform preprocessing.
  • a signal processor e.g., edge detector
  • the decoder 1707 has not yet received the image data, and could be working on decoding of a previous scan. If after preprocessing by the edge counter 1706 the signal qualifies as "interesting", then the image data is read out from the analog shift register 1703 a second time at a slower speed (by altering the speed of the readout clock 1709) commensurate with the data acceptance rate of the decoder 1707, this time with the rescan function disabled.
  • Disabling the rescan function allows the analog shift register 1703 to empty in preparation for the next image, by preventing the image data from being queued in a circular fashion to the opposite end of the register 1703. Simultaneously, a new image may be read into the analog shift register 1703 as the old image is read out. If the original image data did not qualify as "interest- ing" in the first place, then the analog shift register 1703 is emptied in preparation for the next scan, either with a drain or by rapidly reading out with the rescan disabled.
  • a number of the various rescan components can be combined onto a single integrated chip, as indicated in Fig. 72.
  • the analog shift register 1703 commonly exists in CCD chips in the form of an image transfer register.
  • the preprocessing and rescan capabilities described in the immediately preceding paragraphs are particularly useful in applications involving a plurality of detector arrays.
  • Each of the arrays could have its own prepro ⁇ cessor to determine when an "interesting" signal is present, thereby allowing more efficient division of decoder time among only those detector arrays having signals likely to represent a bar code.
  • FIG. 70A-70C Further details of the preferred embodiment of a bar code reader of Fig. 1 are shown in Figs. 70A-70C, with a preferred electronic block diagram and associated wave ⁇ forms shown in Figs. 71A-71D.
  • the bar code reader 1501 depicted in Fig. 1 and Figs. 70A-70D is a handheld unit with relatively large depth of field that can operate in ambient light by employing a preferred combination of the techniques and components described previously herein.
  • the scan head 1522 and optics chassis 1506 of the Fig. 1 bar code reader are shown in more detail in Figs. 70A-70C, along with other optical system and related components.
  • Attached to the analog board 1504 and within the optical chassis 1506 is a CCD detector array 1512 of the masked variety such as depicted in Fig. 52A.
  • the CCD detector preferably has rectangular pixels having a 10:1 aspect ratio, such as shown in and described with respect to the example of Fig. 50A.
  • the CCD detector 1512 is connected to an adaptive exposure system such as shown in Fig. 43A and resident on the analog circuit board 1504 along with a microcontroller that generates CCD timing signals, implements an adaptive exposure feedback system, and provides various signals to the decoder (such as described with respect to the electronics in Fig. 71A below) .
  • an imaging lens 1509 Towards the front of the optical chassis 1506 is located an imaging lens 1509, which is shown in more detail in Figs. 73A (front view) and 73B (side view) .
  • an imaging lens 1749 (corresponding to imaging lens 1509) that comprises a multi-focus lens portion 1753 such as the Fig. 41A multi-zone lens having four concentric annular zones 1750a-d of varying focal lengths, and which form images on the CCD detector 1512.
  • the multi-focus lens portion 1753 is mounted on a cylinder lens 1755.
  • the cylinder lens 1755 and multi-focus lens portion 1753 preferably comprise a single optical element and may be formed of solid glass or plastic.
  • the cylinder lens 1755 has a concave back surface 1752 which allows correction for field of view aberrations as necessary.
  • Figs. 73A and 73B Preferred dimensions (in millimeters) are provided in Figs. 73A and 73B, and are also provided in the table shown in Fig. 73C along with other preferred character ⁇ istics.
  • the tolerance of the center thickness is no more than +0.05 millimeters
  • the tolerance of the surface irregularity is no more than 10 microns from a perfect figure
  • the tolerance of the surface power is no more than 15 microns from a perfect figure.
  • the back surface 1752 minimum clear aperture is 6 millimeters.
  • a high aspect ratio slit aperture 1508 i.e., a series of indi ⁇ vidual apertures formed into a single "continuous" aper ⁇ ture
  • a high aspect ratio slit aperture 1508 i.e., a series of indi ⁇ vidual apertures formed into a single "continuous" aper ⁇ ture
  • Figs. 22A-22C is located in front of the multi-zone lens 1509.
  • Mounted on or in the optical chassis 1506 are two pointer LEDs 1520, each illuminating a holographic diffuser 1519 and sending LED pointing beams through a pointer slit 1518 and pointer lens 1507.
  • the LED pointing system is similar to that described with respect to Fig. 65A.
  • the optical chassis 1506 is rigidly connected to the analog circuit board 1504 in order to fix the relative positions of the slit aperture 1508, multi- zone lens 1509, CCD detector 1512, and pointer LEDs 1520.
  • the preferred bar code reader 1501 further comprises a derivative based signal processor for converting a video signal from the CCD detector 1512 to a bar/space signal for the decoder, with the signal processor being imple ⁇ mented on the analog circuit board 1504 while the decoder resides on the digital circuit board 1503.
  • the bar code reader 1501 also has an optically transmissive spacer such as described with respect to Fig. 63 at the input end of the scan head 1522 with a missing top section for allowing sufficient ambient illumination to reach the bar code.
  • pointer LEDs 1520 In operation, light from the pointer LEDs 1520 passes through holographic diffuser 1519 to uniformly illuminate pointer slits 1518 such as described with respect to Fig. 65A. Pointing beams from the pointer slits 1518 are imaged by the pointer lenses 1507 in order to allow an operator to aim the unit at the bar code and to position the bar code within the depth of field of the bar code reader 1501.
  • the purpose of the LEDs is not to illuminate the bar code, which is illuminated with ambient light, but simply to assist the user in properly positioning and aiming the bar code reader 1501.
  • the LEDs 1520 are preferably turned off during the exposure interval to avoid nonuniform illumi ⁇ nation of the bar code and to conserve power in the unit.
  • Ambient light reflects off a target (e.g., an encoded symbol) located along a virtual imaging line.
  • the spacer at the front of the bar code reader 1501 prevents the unit from being brought so close to the bar code that shadowing prevents adequate ambient illumination from reaching the bar code.
  • the reflected ambient light is focused by the imaging lens 1509 onto the CCD chip 1512.
  • the CCD chip 1512 translates light intensity information into an electrical signal by any of the techniques described previously herein.
  • the electrical signal may be processed and decoded to provide recognition of the encoded symbol.
  • the pointer LEDs 1520 can consume considerable power, and preferably do not illuminate the bar code during image acquisition; therefore the LEDs 1520 are preferably turned off during image acquisition, during the powering of other high power components of the reader (beepers, indicator LEDs, etc.), and as rapidly as possible after a successful decode.
  • the electronics system 1601 comprises a CCD control ⁇ ler 1602 which may be embodied as a MC68HC05 microchip as manufactured by Motorola, Inc.
  • the CCD controller 1602 provides ROG and SHUT clock signals to a CCD chip 1603.
  • An oscillator 1607 also provides a clock signal to the CCD chip 1603.
  • the CCD chip outputs a video signal to a signal processor 1604.
  • the video signal is also provided to a peak detector and compare circuit 1606, which sends an STB signal back to the CCD controller 1602 to indicate when the signal is too big such that clipping might occur.
  • the signal processor 1604 and CCD controller 1602 are both connected to an interface 1608.
  • the signal processor 1604 and CCD controller 1602 collectively output a data valid (DV) signal, start of scan (SOS) signal, something in volume (SIV) signal, and bar/space (BAR) signal to a decoder (not shown) .
  • the decoder returns with a good read (GR) signal when it has recognized a bar code, and illumi ⁇ nates a good read LED 1610 visible to the operator of the bar code reader 1501.
  • the DV signal is generated by the CCD controller 1602 and, at the end of each exposure cycle signals the decoder to temporarily cease any decoding operations and await a new data stream from the signal processor 1604.
  • the DV signal Upon completion of the data transfer and bar width digitization to the decoder from the signal processor 1604, the DV signal is removed and decoding operations can resume. In this manner very high pixel readout rates are used to transfer the bar/ space data to the decoder in a fraction of each exposure cycle, leaving more time available for typically slower symbology decoding operations.
  • Another control signal utilized is referred to as "something in volume" or SIV signal.
  • the SIV signal is used to prevent multiple reads of the same bar code, and is particularly useful in fixed reader applications.
  • the SIV signal is output from an edge counter (internal to the CCD controller 1602) and switches state whenever it detects a number of edges substantially different (larger or smaller) than the usual number present when no object is in front of the bar code reader 1501.
  • the decoder Upon reception of the SIV signal, the decoder begins digitization and decoding operations until a successful decode occurs. After a successful decode the decoder is disabled until the SIV signal is removed and reasserted. The SIV signal is removed only when the edge count returns to its background level, and is only reasserted when a new bar code label is placed in front of the bar code reader 1501. In this way multiple reading of the same label is prevented.
  • the "mode select" (MODE) signal selects different modes of the device, each of which may, for example, be associated with a different speed requirement.
  • the "read out gate” (ROG) signal is an active high signal that drives the CCD chip's ROG input to transfer charge from the photodetectors to the transfer register.
  • the "shutter” (SHUT) signal is an active high signal that performs a shutter function by draining the charge from the photodetectors on the CCD chip 1603. The period of time from the release of this signal (i.e., going low) to the end of the ROG signal going high is the exposure time.
  • the "LED flash” (LED) signal is an active high signal that turns on the pointer LEDs 1520.
  • the LED signal is preferably not active during the exposure time.
  • the "data valid” (DV) signal is, as explained previously, an active high signal that is on during the readout of pixel information from the CCD chip 1603.
  • the DV signal being high represents the time when useful image data exists.
  • the "start of scan” (SOS) signal toggles after each read by the CCD chip 1603.
  • the DV signal is approximately centered with the SOS toggling.
  • the "something in volume” (SIV) signal is, as explained previously, an active high signal that pulses briefly when there is a transition from the background BAR pattern to something different (such as when a label has just appeared) .
  • the "bar input” (BAR) signal is an active high signal which has high signal levels corresponding to relatively darker portions of a read image and low signal levels corresponding to relatively lighter portions of a read image.
  • the "signal too big” (STB) signal is an active high signal that is on when the video output signal from the CCD chip 1602 is larger than the desired value, set by a V ref input on a comparator.
  • the exposure time is preferably reduced whenever this signal is high and increased whenever this signal is low.
  • the "speed up” (SUP) signal when made a high impedance input, will cause the pixel clock to increase to a much higher frequency. When pulled low as an output, the pixel clock will resume its normal frequency.
  • the SUP signal is used on power up to drain the CCD chip 1603 of stored charge.
  • FIG. 71B-71D Exemplary signal waveform diagrams associated with the Fig. 71A embodiment are shown in Figs. 71B-71D.
  • Figure 71B shows the relationship between the ROG signal, DV signal, video signal (i.e., CCD chip output signal), and BAR signal, and exemplary timing characteristics.
  • Figure 71C shows an enlarged timescale view of these same signals, along with the LED signal, SHUT signal, SOS signal, SIV signal, and STB signal.
  • Figure 71D shows these same signals at start up timing (such as when the analog board 1504 is energized) .
  • FIG. 46A shows a top view of the CCD detection module 1001
  • Fig. 46B shows a front view
  • Fig. 46C shows a side view
  • Fig. 46D shows an oblique view of the CCD detection module 1001.
  • the CCD detection module 1001 comprises a CCD chip 1003, one or more light emitting diodes (LEDs) 1005, an imaging lens 1006, and one or more focusing or pointer lenses 1007.
  • the pointer lenses 1007 provide focusing for the LEDs 1005.
  • the imaging lens 1006 may be a single zone lens, a high aspect ratio aperture lens, a multi-zone lens, or a combination thereof.
  • the system may comprise one or more circuit boards 1008 on which to mount the various components of the system, and may further comprise a power supply (not shown) , a CCD driver (not shown) , and signal processing electronics (not shown) .
  • the CCD driver and signal processing electronics may comprise embodiments previously described herein.
  • the CCD detection module 1001 may be configured to operate without the LEDs 1005 and pointer lenses 1007 for providing aiming beams.
  • Figures 47 and 48 show two different embodiments of systems incorporating the CCD detection module 1001 of Figs. 46A-46D.
  • Figure 47 is a diagram of an embodiment of a handheld imaging device 1020 such as may be used in point of sale applications or for inventory control.
  • the CCD detection module 1001 is located in a head portion 1022 of the hand ⁇ held imaging device 1020 and is connected to a handheld scanner controller (not shown) .
  • the CCD video output signal from the CCD detection module 1001 may be directly provided to the handheld scanner controller, or may be conditioned and/or processed first by any of the tech ⁇ niques described previously herein. If pointer LEDs are used, the pointer LEDs 1005 may be controlled by a signal from a trigger 1021 which is routed through the handheld scanner controller.
  • FIG. 48 is a diagram of an embodiment of a portable data terminal 1030 incorporating the CCD detection 1001 module of Figs. 46A-46D.
  • the portable data terminal 1030 comprises a keypad 1032 for entering data and a display screen 1031 for displaying data in a manner known in the art.
  • the CCD detection module 1001 is con ⁇ nected to a decoder (not shown) internal to the portable data terminal 1030.
  • the CCD video output signal from the CCD detection module 1001 may be directly connected to signal conditioning and/or processing electronics in the portable data terminal 1030, or may be conditioned and/or processed within the CCD detection module 1001 by tech ⁇ niques described previously herein, and thereafter pro ⁇ vided to the decoder of the portable data terminal 1030.
  • the pointer LEDs 1005 also may be controlled by the portable data terminal 1030.
  • the CCD detection module 1001 may also be mounted in a fixed fashion (e.g., on a "slim-stand” or support stand) for retail or industrial applications. Alternatively, the CCD detection module 1001 may be placed on a hand or wrist mounted imaging device for applications such as point of sale operations or inventory control.
  • FIG. 49 is a diagram of an embodiment of a multiple-CCD imaging device which may be employed to read a multi-dimensional area.
  • an imaging device 1050 comprises a plurality of CCD detectors 1051.
  • Each CCD detector 1051 has a corresponding lens 1052 for collecting light.
  • the lens 1052 may be a poly-optic lens such as described previously herein, and may be apertured or shaped in accordance with techniques also described previously herein.
  • a transparent element 1054 is placed in the top of the imaging device 1050, such that items bearing encoded symbols may be passed over the imaging device 1050.
  • Light is collected by each lens 1052 along one of a plurality of virtual imaging lines 1053 associa ⁇ ted with the lens 1052.
  • the multiple CCD detectors 1051 may be controlled by a single CCD driver configured in a manner as in Fig. 7, for example. Thus, the CCD detectors 1051 may be polled in series by a single signal processing unit.
  • the imaging device 1050 may be incorporated into a fixed scanning system and may be used for point-of-sale applications, such as retail grocery checkout applications.
  • the techniques described herein may be used in combination or, in some instances, separately so as to provide an image reader having increased capacity to operate in various lighting conditions. The described techniques may allow the image reader either to run faster where a self-generated illumination source is used, or to operate in ambient light in the absence of a self- generated illumination source.
  • CMOS active pixel sensor arrays are described in more detail in, e.g., copending provisional application Serial No. 60/003,256 entitled “Single Chip Bar Code Scanner Using CMOS Active Pixel Sensor Technology,” filed on August 25, 1995, and hereby incorporated by reference as if fully set forth herein.

Abstract

A method and apparatus for reading bar codes or other images while reducing the need for self-generated illumination. An image reader (1501) comprises an optical collection system (1509), an aperture (1508) with a high aspect ratio and/or a multi-focal lens(es). An optically transmissive spacer affixed to the front of the handheld unit, and a diffuser may also be used to reflect ambient light from the handheld unit to provide additional illumination at close range. A masked CCD detector receives collected light and is coupled to an adaptive exposure circuit. A correlated double sampling circuit coupled to the adaptive exposure circuit rejects common mode noise. The correlated double sampling circuit is connected to a signal processor which uses a derivative based technique for detecting edges of the image, to decode the encoded information from the image.

Description

DESCRIPTION
Method and Apparatus for Reading Imaσes Without Need for Self-Generated Illumination Source
Related Application Information
This application is a continuation-in-part of Application Serial No. 08/363,258 filed December 23, 1994, hereby incorporated by reference as if fully set forth herein.
Field of the Invention
The field of the present invention relates to image detection and, in particular, to a method and apparatus for reading images such as bar code labels while reducing or eliminating the need for a self-generated illumination source.
Background of the Invention
Labels bearing information in any of a number of different encoded formats are commonly affixed to products, packaging, or other items and are used in many applications. It is common, for example, to encode retail product identification information in a bar code format on a product package or a label affixed to a product package. Bar code symbols are also used on a broad range of retail packages for check-out and inventory purposes. A bar code reading device, located at the check-out station of a retail establishment, may be used by a clerk to read product identification data into an associated point-of- sale computer system. A number of existing bar codes consist of a series of parallel light and dark rectangular areas of varying widths. The light areas are often referred to as "spaces" and the dark areas as "bars". The bars and spaces are typically arranged and selected so as to define different characters of a particular bar code. A bar code label may be read by a device (e.g., a scanner) which detects reflected and/or refracted light from the bars and spaces comprising the characters. Such a device typically includes a light source for illuminat- ing the bar code label to be read. One common method of illuminating the bar code label is by use of a scanning laser beam, in which case a spot of light is swept across the bar code label and the intensity of returned light is detected by an optical detector. The optical detector generates an electrical signal having an amplitude deter¬ mined by the intensity of the detected light. Another method for illuminating the bar code label is by use of a uniform light source with an array of optical detectors connected to an analog shift register (commonly called a charge-coupled device or CCD) . In such a technique, as with a scanning laser, an electrical signal is generated having an amplitude determined by the intensity of the collected light. In either the laser or CCD technique, the amplitude of the electrical input signal has one population of levels for dark bars and a second population of levels for light spaces. As a label is scanned, positive-going transitions and negative-going transitions in the electrical signal occur, signifying transitions between bars and spaces. The electrical signal may be converted to a binary scan signal which is analyzed to determine the arrangement of bars and spaces of the scanned label. The bar and space information is provided to a decoding unit to determine whether the bar code is recognized and, if so, to decode the information contained in the bar code.
An essential element of virtually every existing bar code reader system is a means for illuminating the bar code. Laser based scanners, for example, typically use a flying laser spot to illuminate the bar code and detect the returned light. Wand devices generally use a light emitting diode (LED) to illuminate a small portion of the bar code and a detector to image the bar code while the operator physically sweeps the illuminated beam across the bar code. Video based readers, such as readers utilizing charge-coupled devices to receive input, typically use a plurality of LED's to illuminate the entire bar code with uniform intensity, and then image the entire bar code with a CCD sensor.
Each of the above systems relies on self-generated illumination for the purpose of distinguishing the bars and spaces of the bar code or for increasing the amount of collected light to enable reasonable performance. The requirement of self-generated illumination may, however, lead to additional complexity, a larger or bulkier device, and/or higher power consumption. For example, a scanning device using a flying laser spot may require a patterned mirror structure, bifocal lens, laser light source, and motor to sweep the laser spot across a target. Each of these components adds complexity and bulk to the overall system. Furthermore, the motor used for sweeping the laser spot tends to be one of the least reliable compo- nents of a scanner, followed by the laser illumination source. The use of a motor and laser illumination source tends to make a system less reliable and more costly. In short, the requirement of a self-generated illumination generally makes an image reading system more complex, costly and unreliable. These drawbacks may be significant in certain applications, particularly in low-cost, hand¬ held reading applications.
In addition, CCD readers often have significant depth of field limitations. For example, CCD readers using red LED's for illumination typically have a depth of field of less than one inch. This limits the types of applications in which CCD readers may be used, and limits the versatil¬ ity of such CCD readers in general. Moreover, due to low sweep speeds, use of CCD readers has also generally been limited to handheld devices.
It would therefore be advantageous to provide a bar code reader capable of operation without self-generated illumination. It would further be advantageous to provide such a bar code reader having an increased depth of field, that may be utilized in a broad base of applications including handheld devices as well as other applications. It would further be advantageous to provide a bar code reader having a reduced number of moving parts and exhibiting increased reliability.
Summary of the Invention
The invention provides ir\ various aspects a method and apparatus for reading bar codes or other images while reducing or eliminating the need for self-generated illumination.
A preferred embodiment of the invention utilizes a CCD based imaging system. The imaging system comprises an adaptive exposure circuit which is used to prevent satura¬ tion of a CCD detector at high illumination levels, thereby increasing the system's dynamic range. The imag¬ ing system further comprises an efficient optical collec¬ tion system, which may include an aperture with a high aspect ratio and/or a multi-plane imaging technique. The imaging system thereby may have an increased depth of field at a particular light level, or an increased read rate, or both. A correlated double sampling circuit is preferably used in conjunction with the CCD detector to allow lower light levels to be detected by the system. The correlated double sampling circuit may also provide an increased depth of field, an increased read rate, or both. The imaging system further takes advantage of derivative based signal processing for enhanced edge detection. Various other embodiments may utilize some but not all of the above elements, or may include additional refinements, while obtaining the benefit of reducing or eliminating the need for self-generated illumination. Brief Description of the Drawings
The various objects, features and advantages of the present invention may be better understood by examining the Detailed Description of the Preferred Embodiments found below, together with the appended figures, wherein:
Fig. 1 is diagram showing a perspective partial cut¬ away view of a preferred handheld bar code reader using ambient light.
Fig. 1A is a block diagram of an optical reading system in accordance with one or more aspects of the present invention.
Fig. 2 is a diagram of a CCD detector receiving a focused beam of light.
Fig. 3A is a potential diagram illustrating the processes of charge accumulation and charge transfer from a photodiode to a CCD shift register location.
Fig. 3B is a physical illustration of a charge coupled device having five potential wells.
Fig. 3C is a potential diagram illustrating a transfer of charge along a CCD shift register.
Fig. 4 is a diagram of a simplified output stage of a CCD detector.
Fig. 5 is a diagram of a bar code and a corresponding waveform after filtering. Fig. 6 is a circuit diagram of a driver circuit for driving a CCD chip.
Fig. 7 is a block diagram of a multiple-CCD detector system.
Fig. 8A is a block diagram of a correlated double sampling circuit, and Fig. 8B is a graph of waveform diagrams associated with the sampling circuit of Fig. 8A.
Fig. 9 is a circuit diagram of a CCD output stage.
Fig. 10A is a block diagram of a dual sampler processor, and Fig. 10B is a graph showing timing diagrams for the dual sampler processor of Fig. 10A. Fig. 11A is a circuit block diagram of a dual slope processor, and Fig. 11B is a graph showing timing diagrams for the dual slope processor of Fig. 11A.
Fig. 12 is a graph of transfer functions for the dual sampler processor of Fig. 10A and the dual slope processor of Fig. 11A.
Fig. 13A is a diagram of an optical system based on a single lens model, and Fig. 13B is a diagram of a single lens model showing creation of a blur spot when the target image is moved away from the focal point.
Figs. 14A-14D are diagrams of various types of aberrations and distortions that may be associated with an optical system.
Fig. 15 is a graph illustrating a cosine law effect. Fig. 16 is a diagram illustrating light collection.
Fig. 17 is a diagram comparing the point spread function shapes through focus for a round aperture and an oblong/elliptical aperture.
Fig. 18 is a diagram illustrating change in elliptical spot size with respect to tilt angle.
Fig. 19 is a graph showing illustrative depth of field plots for a round aperture and for an elliptical aperture collecting the same amount of light.
Fig. 20A is a diagram of an unapertured lens interposed between a target and an imager, and Fig. 2OB is a diagram of an apertured lens in a similar relative position.
Figs. 21A and 21B are diagrams showing an oblique view and a front view, respectively, of a symmetric ulti- focal lens.
Figs. 22A-22C are diagrams showing various views of a multi-focal lens system combined with an elliptical aperture.
Figs. 23A, 23B and 23C are graphs showing a minimum renderable bar width as a function of distance from a single-lens, a poly-optic lens with no interference, and a poly-optic lens with inter-zonal interference, respectively.
Figs. 24A and 24B are diagrams of point spread functions for a center zone of the Fig. 21A-21B lens through focus and away from focus, respectively.
Figs. 25A and 25B are diagrams of point spread functions for a ring zone of the Fig. 21A-21B lens through focus and away from focus, respectively.
Figs. 26A through 30C are approximated point spread function plots of various combinations of center and ring zones of various focal lengths.
Figs. 31A-31C are diagrams of a symmetric multi-focus lens shown from several viewpoints with a plurality of apertures. Fig. 32 is a diagram of a cut-away symmetric lens.
Fig. 33 is a diagram of an asymmetric lens having its optical axis in the center of the lens.
Figs. 34A is a diagram showing off-axis tilt when the center zone focuses to the closest depth and the outside ring zone focuses to the farthest depth, and Fig. 34B is a diagram showing the reverse situation.
Figs. 35A and 35B are diagrams showing different embodiments of a lens system having an array of lenses arranged in a pattern around an imaging device. Figs. 36A-36C are diagrams of various embodiments of a lens system for planar coverage having an array of lenses arranged in a pattern around an imaging device.
Figs. 37A-37C are diagrams showing other embodiments of a lens system for planar coverage having an array of lenses arranged in a pattern around an imaging device and including a collimating lens.
Figs. 38A-38B are diagrams of alternate embodiments of a poly-optic lens system.
Figs. 39A-39B are diagrams of alternate embodiments of an optical system using curved mirrors to collect light. 8
Figs. 40A-40D are diagrams of a particular poly-optic lens from different angles illustrating light collection.
Figs. 41A-41C are diagrams of a preferred poly-optic lens. Figs. 42A-42C are diagrams of other embodiments of optical systems using shutters to reduce inter-zonal interference.
Fig. 43A is a circuit diagram of a preferred adaptive exposure circuit, and Fig. 43B is a graph of waveform patterns for the circuit of Fig. 43A.
Figs. 44A and 44B are block diagrams showing alternative configurations of an optical system employing an adaptive exposure circuit.
Fig. 45A is a block diagram of a signal processing circuit for determining transitions in a video signal, and Fig. 45B is a graph of waveform patterns for the circuit of Fig. 45A.
Figs. 46A-46D are diagrams showing various views of a CCD detection module incorporating various aspects of the present invention.
Fig. 47 is a diagram of an embodiment of a handheld imager incorporating the CCD detection module of Figs. 46A-46D.
Fig. 48 is a diagram of an embodiment of a portable data terminal incorporating the CCD detection module of Figs. 46A-46D.
Fig. 49 is a diagram of a multiple-CCD omnidirec¬ tional imaging device for reading over a multi-dimensional area. Figs. 50A and SOB are diagrams illustrating the effect on an output signal of rotating a bar code with respect to a CCD array.
Fig. 51 is a graph comparing pixel width versus roll angle for a CCD array with square pixels and one with rectangular pixels (having a 4:1 aspect ratio) .
Figs. 52A and 52B are diagrams showing an oblique view and a side view, respectively, of a masked CCD array. Fig. 52C is a diagram of a masked CCD array embodi¬ ment including an additional lens for reducing the effect of vignetting.
Fig. 53 is a diagram of an imaging system including an anamorphic lens system.
Figs. 54A and 54B are diagrams of apodized poly-optic lenses for reducing inter-zonal interference.
Figs. 55A and 55B are line spread functions associated with the apodized lenses of Figs. 54A and 54B. Figs. 56A and 56B are diagrams illustrating a multi¬ element lens system for enhancing depth of field while minimizing interference among lens elements.
Fig. 57 is a diagram illustrating a multiple-focus optical system with a poly-optic lens having some of the aberrations of the zone uncorrected so as to narrow the angular field of view for the longer focal length zones.
Fig. 58 is a diagram of a multi-directional optical system using cylindrical optics to allow variously oriented images (e.g., bar codes) to be read. Fig. 59 is a diagram of a one-axis filter for determining the orientation of a bar code.
Figs. 60A-60C are diagram of a multi-directional optical system using a beam splitter and mirrors, with various alternative area and linear sensor configurations depicted in Figs. 60D-60F.
Fig. 61A is a diagram of another embodiment of a multi-directional optical system employing a mirror basket to divide a single projected image and independently rotate the different segmented images, and Fig. 61B is a particular application of a mirror basket with a handheld bar code reader.
Figs. 62A-62F are diagrams associated with another embodiment of a multi-direction reader employing a kaleidoscope mirror for projecting multiple rotated images of a bar code onto a detector array.
Fig. 63 is a diagram of an optically transmissive spacer extending forward from a bar code reader housing to allow ambient light to reach a bar code or other image to be read.
Fig. 64A is a diagram of an optical diffuser, while
Fig. 64B is a diagram illustrating the effect of a system employing the optical diffuser of Fig. 64A, and Fig. 64C is a diagram comparing the operation of a system without an optical diffuser.
Fig. 65A and 65B are side view and front view diagrams, respectively, of an LED pointer system, and Fig. 65C is a diagram illustrating various pointer shapes at different distances.
Fig. 66 is a diagram of a signal processor circuit including an infinite impulse response high pass filter.
Fig. 67 is a diagram of a signal processor circuit including a finite impulse response high pass filter.
Figs. 68A and 68B are graphs of exemplary impulse response coefficients for the signal processor of Fig. 67, and Figs. 69A and 69B are exemplary waveform diagrams also associated with the signal processor of Fig. 67. Figs. 70A and 70B are different depth cross-sectional views of the scan head of the bar code reader of Fig. 1.
Fig. 70C is a top sectional view of the optical chas¬ sis mounted in the scan head shown in Figs. 70A and 70B.
Fig. 71A is an electrical block diagram of a pre- ferred electronics and control system for the Fig. 1 bar code reader, and Figs. 71B through 71D are exemplary waveform diagrams associated with the Fig. 71A electrical block diagram.
Fig. 72 is a block diagram of a bar code reader having rescan capabilities.
Figs. 73A and 73B are diagrams of a mounted multi- focal lens showing preferred dimensions thereof, and Fig. 73C is a table showing preferred characteristics of the multi-focal lens of Figs. 73A-73B. Detailed Description of the Preferred Embodiments
Figure 1 is a cut-away view of a preferred embodiment of a bar code reader. The bar code reader 1501 depicted in Fig. 1 is a handheld unit with relatively large depth of field that can operate in ambient light by employing a preferred combination of the techniques and components as will be described in detail herein. The handheld bar code reader 1501 comprises a housing 1502 encasing most or all of the optical system and electronics. The bar code reader 1501 further comprises a digital board 1503 on which is mounted an optics chassis 1506, and an analog board 1504 connected to the back of the optics chassis 1506 and also connected to the digital board 1503. A handle board 1505 is also encased in a portion of the housing 1502 forming a handle below the scan head 1522.
The scan head 1522 and optics chassis 1506 are shown in more detail in Figs. 70A-70C, along with other optical system and related components, and described in more detail later herein. On the optical chassis 1506 is mounted an imaging lens 1509 such as imaging lens 1749 shown in Figs. 73A-73B. The imaging lens 1509 is prefer¬ ably a multi-focus lens providing an enlarged depth of field. Details regarding a variety of suitable multi- focus lenses and other optical schemes, as well as preferred electronics for facilitating operation in ambient light, are described in more detail later herein.
Figure 1A is a block diagram of an image reading system 100 in accordance with one or more aspects of the present invention. The image reading system 100 of Fig. 1A comprises an optical collection system 103, a CCD detector 104, a pre- amplification block 105, a signal processing block 106, and a decoding block 107. The optical collection system 103 operates in the presence of ambient illumination 101, gathering ambient light reflecting off a target (such as a bar code label 102) and focusing the gathered light on the CCD detector 104. The CCD detector 104 interacts with the pre-amplification block 105 to obtain a CCD input signal comprised of imaging data. The CCD input signal may be processed by the signal processing block 106 and provided to the decoding block 107 for decoding, as further explained herein.
Figure 2 is a diagram of a CCD detector 110 receiving a focused beam of light 115. Light is reflected from a target located at a target line 114. The reflected light is passed through a lens 113 and focused onto the CCD detector 110. The CCD detector 110 may be embodied as a linear integrated chip comprising an array of photodiodes 112. In operation, the CCD detector 110 acts in a manner similar to a video camera, with the array of photodiodes
112 corresponding to pixel locations. The pixel size and spacing along with the focusing limitations of the lens
113 generally define the depth of field and the minimum bar code size for an imager incorporating such a CCD detector 110.
Linear CCD detectors have four basic elements. These four basic elements relate to photon to electron conver¬ sion, charge storage, charge transfer, and charge to voltage conversion. Photon to electron conversion is per¬ formed by multiple photodiodes (e.g., pixels 112). Typical photodiodes have a quantum efficiency of 50%, which means that one electron charge is generated for every two photons that hit the photodiode. The electron charges are stored in a potential well, which acts like a capacitor. When an exposure period is complete, a read out gate pulse received from a controller allows the accumulated charges to be dumped into a CCD shift regis¬ ter. A potential diagram illustrating the processes of charge accumulation and charge transfer from a photodiode to a CCD shift register location is shown in Fig. 3A.
Charge transfer along the shift register occurs while the photodiodes collect charge during the next exposure period. The charge coupled device may be embodied as a distributed gate metal-oxide silicon field-effect transistor (MOSFET) 130, a physical illustration of which is shown in Fig. 3B. The distributed gate MOSFET 130 has a long source-drain channel 131 and contains a plurality of gates 132 (one gate 132 per pixel 112) . Charge is injected at each gate location (e.g., shift register loca¬ tion) by a photodiode, by the process depicted in Fig. 3A. Charge is transferred along a chain of shift register locations by controlling the gate voltages. For the MOSFET 130 of Fig. 3B, three clocks are needed to transfer the charge along the shift register. Photodiodes can thus inject charge every three gates. Alternatively, two clocks may be used so as to allow the photodiodes to inject charge every two gates. A potential diagram illus¬ trating the described charge transfer process is shown in Fig. 3C.
At the end of the analog shift register, an amplifier converts the charge to voltage. This conversion is typic¬ ally accomplished by allowing a capacitor to integrate the charge and then buffering the voltage to an output pin. A MOSFET is used to clear the charge from the capacitor between transfers of each pixel . The conversion process may also involve amplification of the converted voltage level. The output voltage generally has a DC bias level equal to about half of the power supply voltage. The higher the input light intensity, the lower the DC bias voltage, as the photoelectric charge is in the form of injected electrons.
Figure 4 is a diagram of a simplified output stage of a CCD detector 110 such as shown in Fig. 2. In Fig. 4, a flow of electrons 150 is accumulated by an integration capacitor 152 (also referred to a sense capacitor) . A transistor 154, such as a MOSFET, clears the stored charge between each pixel. The MOSFET 154 may be referred to as a reset transistor or a reset MOSFET. The MOSFET 154 is connected to a reference voltage source 156. The voltage level of the sense capacitor 152 may be read through a buffer stage 155, which also may comprise a MOSFET. The buffer MOSFET 155 is connected at one end to a positive bias voltage 160, at the other end through a current sink 158 to a negative bias voltage 159. In one embodiment, the current sink 158 comprises a resistor, and the negative bias voltage 159 is set to ground level.
Further information pertaining to the operation of charge-coupled devices may be found in J. Millman, Microelectronics. Digital and Analog Circuits and Systems, (McGraw-Hill 1979), pp. 298-312, J.D. Mosley, "CCD Imaging Arrays," EDN Magazine, (Aug. 20, 1990), pp. 116-126, and in the 1993 Linear Sensor IC Databook published by Sony Corporation.
The CCD detector 110 provides data in the form of a discrete time analog signal. The analog signal appears similar to "stair-steps", with data from each pixel contributing to one of the steps. Most CCD chips have an output signal that is in a low state when the image is bright (white) and a high state when the image is dark
(black) . The signal output from the CCD detector 110 may be low pass filtered to arrive at a smooth signal, such as shown in Fig. 5. In Fig. 5, an exemplary portion of a bar code is shown having darker bars 170 and lighter spaces 171. A corresponding CCD output signal 172 is shown, after low pass filtering, above the bar code portion. The filtered CCD output signal 172 is shown in a high state when the darker bars 170 are read, and a low state when the lighter spaces 171 are read. The CCD output signal 172 may have a sizable DC component (e.g., 4.5 volts).
The amplitude of the CCD output signal 172 depends mainly on the sensitivity of the photodiode array and the exposure time. During the exposure time, the photo- detectors in the CCD detector 110 collect charge. Periodically, the accumulated charge is dumped into an analog shift register. The information in the analog shift register is serially read out to form an analog CCD signal which, when filtered, corresponds to the CCD output signal 172. All of the accumulated data must be transferred from the photodetectors and read out between exposures, so that residual charge does not distort the data from the next exposure period. If it is required to have an exposure time faster than the data read out time, a shuttered CCD device may be used, thereby allowing the exposure time to fill some fraction of the read out time.
The CCD detector 110 may particularly be embodied as a CCD chip of the ILX series as manufactured by Sony, such as an ILX-503A chip (having 2048 pixels) or ILX-505A chip
(having 2592 pixels) . The Sony ILX-703A may be used in applications requiring an electronic shutter.
Descriptions of these devices may be found in the Sony
1993 CCD Linear Sensor IC Databook. Figure 6 is a circuit diagram of a driver circuit suitable for driving an ILX-503 or ILX-703 CCD chip 204. The Fig. 6 driver circuit generates a pixel clock signal (øCLK) 201 and a read out gate clock signal (φROG) 202 to facilitate transfer of charge from the photodetectors to the analog shift register. The pixel clock signal 201 is generated by a resonator oscillator circuit 203 as depicted in Fig. 6. The read out gate clock signal 202 is generated by a read out gate clock generator circuit 205 as also depicted in Fig. 6. The read out gate clock generator circuit 205 comprises an oscillator circuit 207 followed by a pair of one-shot circuits. Another one-shot circuit 208 acts to temporarily disable the pixel clock signal 201 when the read out gate clock signal 202 is generated. The CCD chip 204 generates a CCD video output signal 206.
The CCD output signal 206 may, as noted, be provided to a low-pass reconstruction filter and thereafter to other circuitry for signal processing, decoding and other functions as further described herein. The low-pass reconstruction filter may be embodied as part of the pre- amplification block 105 or the signal processing block 106. The low pass filter preferably is constructed so as to block out signals having a frequency higher than one- half the frequency of the pixel clock signal 201. The low pass filter preferably has a steep filtering character¬ istic, since a gradual filter may increase the number of pixels per bar required for reading a bar code. The low pass filter may also provide gain, if necessary, depending on the light level and the exposure time.
To achieve necessary coverage, multiple linear CCD arrays may be used for imaging. Sensitivity tends to decrease with larger imagers, making the use of multiple linear CCD arrays preferable in certain applications. Moreover, cost tends to increase geometrically with larger CCD arrays. Although a number of embodiments are described herein in terms of linear CCD arrays, it should be noted that many principles of the embodied devices described are applicable to area CCD sensors as well.
Figure 7 is a block diagram of a multiple-CCD detector system, which may be configured, for example, to implement a low-cost multiple scan line system. In Fig. 7, a plurality of CCD detectors 230 are clocked by a single CCD driver circuit 231, such as shown previously in Fig. 6. The CCD driver circuit 231 provides a pixel clock signal and a read out gate clock signal 233 to each of the CCD detectors 230. The CCD detectors 230 are coupled to a signal processor 232 via a high speed analog multiplexer 235, which selects the CCD detectors 230 in sequence for processing. A counter 237 may be coupled to the CCD driver circuit 231 so as to count clock pulses and switch the high speed analog multiplexer 235 at appropriate times. The signal processor 232 receives a CCD output signal 238 from each CCD detector 230 in sequence, as controlled by the analog multiplexer 235, and thereby processes the CCD output signals 238 repetitively, in a sequential fashion. The CCD detectors 230 may be physically placed in a variety of different patterns so as to generate a desired imaging pattern. The CCD detectors 230 may be completely separate, mounted on a common circuit board, or constructed on the same chip substrate.
There are a variety of sources of noise in CCD chips which may need to be accounted for in designing a CCD- based optical system. Four sources of noise in particular are noted here. The first is photodetector shot noise, which may be described by the following equation:
I» - VUelβB) where Id is the dark current (in amps) of the photo- detector, B is the bandwidth of the system (in hertz) , and e is the charge on an electron (i.e., 1.6xl0-19C) . Dark current Id is a strong function of temperature, doubling every 25°C (in the case of the ILX-503A chip) . The effect of photodetector shot noise becomes more pronounced the longer the photodetector integrates. Thus for longer exposure times, the photodetector shot noise will increase. The effect of this shot noise, however, tends to be offset by the fact that longer exposure leads to more signal. The signal-to-noise ratio due to shot noise is proportional to the square root of the exposure time. A second source of noise is associated with the process of charge transfer. This source of noise results from surface states (i.e., 1/f) and thermal noise gener¬ ated by the charge transfer MOSFET. Charge transfer noise manifests as a linear growth of noise from the starting pixel to the ending pixel. While the first pixel to be transferred stays in the charge transfer register for only a short time, the last pixel remains much longer, there¬ fore accumulating much more noise than the first pixel. Because charge transfer efficiencies are generally high, this noise is usually not significant enough to degrade performance.
A third source of noise is known as kTC noise. This type of noise is caused by a switched capacitor effect related to the reset transistor at the end of the CCD chain. A similar effect may also occur in switched capacitor filters. This source of noise is typically the dominant noise source in the CCD chip, but can be miti¬ gated or removed by other techniques discussed herein, such as correlated double sampling.
A graph of kTC noise power as a function of frequency is shown in Fig. 9B for a CCD output stage shown in Fig.
9A. The CCD output stage in Fig. 9A corresponds to that of Fig. 4, except that the reset MOSFET 154 in Fig. 4 is modelled by its internal resistances in an OFF state
(i.e., Roff) and in an ON state (i.e., Ron) . In Fig. 9B, one graph 260 shows kTC noise power when the reset MOSFET
154 is turned off and thus has a resistance of Ro££, while another graph 261 shows kTC noise power when the reset
MOSFET 154 is turned on and thus has a resistance of Ron.
A fourth source of noise results from the charge-to- voltage conversion process. The charge-to-voltage ampli¬ fier is subject to surface states (i.e., l/f) and thermal noise of the FET, as well as thermal noise due to other resistances. This noise source dominates after the kTC noise and is not affected by exposure time. This noise source has only a weak temperature dependance.
Figure 8A is a block diagram of a particular type of correlated double sampling circuit such as may be employed to reduce kTC noise. Figure 8B is a diagram of waveforms illustrating some of the noise reduction properties of the circuit of Fig. 8A.
A correlated double sampling circuit such as shown in Fig. 8A may advantageously be employed to remove reset noise generated by the periodic resetting of the sense node by the reset MOSFET in the CCD output stage (shown, e.g., in Fig. 9A) , thereby improving signal-to-noise ratio. In general, reset noise may be expressed in terms of the following relationship:
Nr(V) = V/(4kTBR) where k = Boltzmann's constant = 1.38 x 10"23, T = absolute temperature (Kelvin) , B = noise power bandwidth (Hertz) , and R = resistance (Ohms) . The correlated double sampling circuit generally operates by sampling the CCD output voltage once after the reset pulse has been switched off and once again after signal charge has been read out, thereby providing a direct measurement of charge. This charge measurement is generally independent of reset noise, since the reset noise will be "correlated" over the duration of a pixel period. In other words, the noise voltage due to the channel resistance will not change over the period when the two samples are taken, allowing the reset noise to be eliminated externally from the CCD chip. As shown in Fig. 8A, an input signal 240 (such as output from CCD chip 204 of Fig. 6) is buffered by an amplifier 241. The output of the amplifier 241 is con¬ nected to a DC restoration circuit 239 comprising an analog switch 243, a capacitor 242, and a bias voltage 244 (which may be ground) , configured as shown. A clamp sig¬ nal 255 controls the analog switch 243. The DC restora¬ tion circuit 239 is connected to another buffer 245. Buffer 245 is connected to another analog switch 246. The analog switch 246 is controlled by a sample signal 259. The other end of the analog switch 246 is connected to a capacitor 247 and to another buffer 248. Buffer 248 outputs an output signal 249.
The input signal 240 may appear, for example, like the illustrative input waveform 250 of Fig. 8B. The wave- forms of Fig. 8B have been idealized somewhat for purposes of explanation. The heavier shaded lines in the sample input waveform 250 illustrate relatively low frequency noise 257 appearing in the input signal 240. A feed- through pulse 254 occurs in the input waveform 250 when the reset MOSFET is active. During the interval of the feedthrough pulse 254, relatively high bandwidth noise may occur. However, once the feedthrough pulse 254 is ended and the reset MOSFET is inactive, a generally low fre¬ quency, constant noise pattern will be expected. This low frequency noise 257 occurring in the interval 255 between feedthrough pulses 254 may be removed or substantially reduced by the correlation operation of the circuit of Fig. 8A.
In operation, the input signal 240 is applied to the input buffer 241. When the clamp signal 255 (shown as waveform 251 in Fig. 8B) is active, the capacitor 242 is coupled to the bias voltage 244 through the analog switch 243. The capacitor 242 is thereby charged to the reset voltage level 256, as shown in relation to waveform 250 of Fig. 8B. The clamp signal 255 is then deactivitated, releasing the clamp (i.e., opening the analog switch 243) but allowing the capacitor 242 to remain charged at the reset voltage level 256.
The input waveform 250 then switches to a level 265 determined by the amount of charge collected by a particu- lar photodiode in the CCD detector. The output of the buffer 241 follows the input waveform 250, which forces the opposite end of the capacitor 242 to a new voltage level, since the capacitor 242 maintains the same voltage across it that it had prior to the changing levels in the input waveform 250. The voltage at the input of the buffer 245 therefore comprises the difference between the voltage level 265 of the input waveform and the reset voltage level 256. Because the reset voltage level 256 of the input waveform 250 contains noise 257 that is cor- related over a pixel period, common mode noise is thereby removed from the output of the DC restoration circuit 239.
The voltage level of the capacitor 242 is sampled when the sample signal 259 (shown as waveform 252 in Fig.
8B) is activated. When the sample signal 259 is activ- ated, capacitor 247 charges to a voltage level determined by the voltage of the capacitor 242. The sample signal 259 is then deactivated, causing switch 246 to open, while capacitor 247 holds the voltage level until the next sample period of sample signal 259. An output waveform 253 associated with input waveform 250 is also shown in Fig. 8B. The output waveform 253, such as may appear at the output signal 249 of the correlated double sampling circuit of Fig. 8A, comprises a sequence of sampled voltage levels of the input waveform 250, and is offset by the bias voltage 244. As noted, the output waveform 253 generally has a significant reduction in common mode noise.
A more detailed diagram of a particular type of correlated double sampling circuit, known as a dual sampler processor, is shown in Fig. 10A. Timing diagrams for the correlated double sampling circuit of Fig. 10A are shown in Fig. 10B. In Fig. 10A, a CCD signal 270 is connected to a pre-amplification stage 273. A resistance RL represents the output resistance of the CCD detector. The pre-amplification stage 273 is connected to a low pass filter stage 274 comprising an amplifier 275, two resist- ors 276 and 277, and a capacitor 278, configured as shown. The transfer function of the low pass filter stage is:
|HRC(f) j2 = 1/(1 + (2τrTD)2) where TD = the RC time constant of resistor 276 and capacitor 278. The low pass filter stage 274 is connected to a capacitor 280. Capacitor 280 is connected to a buffer stage 282 and one end of a switch SI. The other end of the switch SI is connected to ground 286. The buffer stage 282 is connected to a second switch S2. The other end of the switch S2 is connected to a capacitor 283 and is buffered by another buffer stage 284. Buffer stage 284 outputs a correlated signal 288.
Operation of the correlated double sampling circuit shown in Fig. 10A may be explained with reference to the timing diagrams of Fig. 10B. The CCD signal 270 may be divided into a plurality of pixel periods, each of which comprises input from a particular pixel, as noted. Timing of the dual sampling processor may be divided into five non-exclusive time intervals within a pixel period. These time periods may be referred to as (1) reset, (2) reference, (3) clamp, (4) video dump, and (5) sample. Each of these time periods are explained more fully below. The reset period 310 shown with respect to waveform 300 in Fig. 10B is commenced when the reset switch (e.g., reset MOSFET as shown in Fig. 4) is turned on, which starts a pixel readout sequence. Waveform 300 in Fig. 10B is an example of a CCD video output signal starting from the time the reset switch is turned on. During the reset period 310, the sense capacitor 152 in the CCD output stage (e.g., as shown in Fig. 4) is set to the voltage level of the reference voltage source 156, thereby setting a positive charge packet in sense capacitor 152 to be eroded by signal electrons. The time constant associated with this operation is preferably very small. For example, the time constant may be 800 picoseconds, where the sense capacitor 152 has a value of 0.4 pF and an "on" resistance of 2 kΩ. As a consequence of the small time constant, the reset pulse width applied to the reset MOSFET 154 to control clearing the charge on the sense capacitor 152 can be made quite small if desired (e.g., a few nanoseconds) . Application of a reset pulse to the reset MOSFET 154 generates a feed-through pulse seen in the CCD signal 270. This feed-through pulse is caused by an effective capaci¬ tor divider network comprising the sense capacitor 152 and a gate-to-source capacitance Cgs of the reset MOSFET 154. After the reset period 310, a reference period 312 occurs after the reset switch (i.e., MOSFET 154) is turned off. The voltage level at the sense node (e.g., at the sense capacitor 152) quickly decays to the reference potential 156 during this time. The sense node retains the reference potential until signal charge is dumped onto the sense capacitor 152.
A clamp period 314 occurs during the reset period 312. During the clamp period 314, the clamp switch SI is turned on, thereby forcing the voltage level at the input to the first buffer stage 282 to ground level or, alterna¬ tively, a non-zero reference level if desired. This action establishes a precise reference level for each pixel before video dump occurs. It also eliminates reset noise introduced during the time the reset switch 154 was on. The clamped level is considered the first "sample" taken by the dual sampler circuit. The clamp is released (i.e., switch SI turned off) at the start of a following video dump period 316.
After the clamp period 314 follows the video dump period 316. During the video dump period 316, charge from the summing well is transferred onto the sense node (i.e., at the sense capacitor 156) when the output transfer gate becomes active. This causes the sense node to be dis¬ charged by an amount proportional to the signal level in that summing well of the CCD.
During the video-dump period 316 occurs a sample period 318. The video si nal 300, after it has settled, is sampled and held by the action of switch S2. Switch S2 is turned on before video dump to assure that sampling begins when the clamp switch is "on" at ground potential. If sampling were not to start until after video dump, the response would exhibit a slight non-linearity resulting from the finite bandwidth and response time of the sample- and-hold circuit. Because switch S2 is turned on before video dump, sampling begins at the same potential for each pixel. The time period when the clamp is released to when the video is held is referred to as the clamp-to-sample time or sample-to-sample time ts.
A preferred timing guideline is to position the reset clock far from the video dump edge due to its large ampli¬ tude. This positioning allows sufficient settling time of the reference level before the video signal is sampled. The reset clock feed-through pulse is eliminated as a result of the clamp switch, as the video signal is clamped over the entire pixel period except during the sample-to- sample time ts. The sample-to-sample time ts and electrical bandwidth TD are preferably selected to achieve low noise perform¬ ance. The correlated double sampling processor acts as a band-pass filter in that low frequency noise is rejected when ta is made smaller (because these low frequencies become correlated) , while high frequency noise is rejected when the bandwidth is made smaller. The center frequency of the effective band-pass filter is tuned by adjusting t s and TD while maintaining a high signal amplification. The optimum center frequency is selected when the noise is minimized for a given CCD noise input spectrum.
Preferably, the correlated double sampling circuit is configured so as to completely eliminate kTC noise and to optimally filter white and 1/f noise generated by the on- chip CCD amplifier, thereby improving overall signal-to- noise ratio.
Another type of correlated double sampling circuit that may be used is a dual slope processor. The perform¬ ance of the dual slope processor when operating at a slow data rate is very similar to that of the dual sampler processor, but the dual sampler processor is preferred at high pixel rates. A diagram of a dual slope processor is shown in Fig. 11A. Timing diagrams for the dual slope processor of Fig. 11A are shown in Fig. 11B.
In the dual slope processor of Fig. 11A, a resistor 352 represents the output resistance of a CCD detector. A CCD output signal 350 is connected to a pre-amplifier 353. The output of the pre-amplifier 353 is connected to an inverting amplifier 355 and a non-inverting amplifier 354, each of which have a gain of unity. The output of the inverting amplifier 355 and the output of the non- inverting amplifier 354 are each connected to a three- position switch 360. The output of the non-inverting amplifier 354 is connected through a resistor 356 to one position 361 of the three-position switch, while the output of the inverting amplifier 355 is connected through a resistor 357 to another position 363 of the three- position switch 360. A third position 362 of the three- position switch 360 is left open. A common end 364 of the three-position switch 360 is connected to an integrator 370. The integrator 370 comprises an operational ampli¬ fier 366 and a capacitor 367. The input and the output of the operational amplifier 366 are connected to opposite ends of the capacitor 367 as shown in Fig. 11A. Another switch 365 is connected from the input of the integrating amplifier 366 to ground 369.
In operation, the CCD video signal 350 is amplified by the pre-amplifier 353. The pre-amplifier 353 may have a gain of, for example, 2 V/V. The output of the pre- amplifier 353 is fed into the integrator 370 by way of the three-position switch 360. Processing of the CCD video signal 350 begins by first resetting the integrator 370 to ground potential by turning on switch 365 and setting switch 364 to position 362. After resetting the inte- grator 370, switch 365 opens, and the three-way switch 360 switches to the first position 361 so as to allow the reference voltage level to be integrated onto the integrating capacitor 367 for a sample time ts.
The effect of integrating is to subtract the reset potential and thereby eliminate common mode reset noise. After integrating for ts seconds, charge is dumped onto the sense node (e.g., the sense capacitor 152) of the CCD. Simultaneously the three-position switch 360 is switched to the second position 363 to connect the integrator 370 to the output of the inverting amplifier 355. The refer¬ ence level plus the video level are integrated for ts seconds when the switch 360 opens to the third position 362. The signal is then held by the integrator 370.
Output signals from either the dual sampler or dual slope processors may be digitized by suitable means, such as and analog-to-digital converter, and processed digit¬ ally, or may be processed using some or all of the analog techniques described further herein.
Transfer functions for the dual sampler and dual slope processors as a function of normalized frequency (27rfts) are compared in Fig. 12, assuming ts = 2TD for the dual sampler processor. As shown in Fig. 12, the filter- ing action is very close between the two different types of processors, although the correlated dual sampler processor is preferred for higher data rates.
The particular characteristics of CCD detectors place special requirements on the optical system. For example, a CCD-based bar code reader designed to read fast moving bar code labels must have a relatively short exposure time. In order to read in ambient light conditions, with little or no self generated light, the optical system of such a bar code reader is thus preferably configured so as to gather the maximum light available.
The invention contemplates a number of different optical structures that are suitable for use in the pre¬ ferred embodiments. These optical structures include, for example, poly-optic and multi-focus lenses, graded index elements, optical flats, prisms, and others. These struc¬ tures may be used alone or in combination to increase the light gathering power and depth of field of the various embodiments described herein. A simple optical system for a linear CCD is a single lens system. A single lens model is diagrammed in Fig. 13A, which shows some of the basic relationships between reading distance and length of scan line. The single lens model of Fig. 13A shows the gathering of light from a target 400 focused through a lens 401 and projected onto an imager (e.g., CCD detector) 402. As shown in Fig. 13A, the lens 401 can be thought of as projecting from the imager 402 a virtual scan line of length WF at a distance XF where the lens 401 is focused. The ratio of the virtual scan line length WF to the imager length WA is the ratio of the focal distance XF to the imager distance I:
WF = Wi XF/I The imager distance I can similarly be calculated from the lens equation: I = XPf/(XF - f) Each pixel is "projected" along the length of the virtual scan line WF. The image of the pixels by virtue of such projection is magnified by the XF/I ratio.
The width WF corresponds to the field of view of the imager 402 at a distance XF. The imager 402 will have an angle of view θ as follows: θ = 2 Arctan (Wi/2XF) Design of an optical system may involve considera¬ tions of focus blur, diffraction, aberrations, variable light intensity, light scattering, label reflectivity, and label print contrast ratio. These topics are discussed briefly below before a more complete description is pro¬ vided of the preferred embodiments of optical structures in accordance with various aspects of the present invention.
The phenomenon of focus blur relates to the distance of the target image from the focal spot of the lens. If a bar code is located at a distance XF from the lens and within the field of view width WF, then the bar code will be in focus on the imager 402. If, as shown in Fig. 13B, the bar code is moved closer to or farther away from the lens 401, the image will become blurry. This blurriness results from the bar code being imaged by a "virtual spot" 405 or blur spot that is larger than the focal spot at the optimum focal point 406. The size of the blur spot 405 is a linear function of distance from the focal point 406, as can be seen in Fig. 13B. The blur spot 405 grows more rapidly with distance for a wider lens aperture (opening) than for a narrower lens aperture. The ratio of the aperture size to the focal length is called the f-number or f-stop and is written as f/#. Thus, for example, an f/4 lens will have half the depth of field of a f/8 lens, because the diameter of the aperture of the f/8 lens is half the aperture diameter of the f/4 lens. An analogy can be drawn between a CCD imaging system and a laser optic system, where the waist of the laser beam corresponds to the focal point of the imaging system, and the size of the laser spot corresponds to the image of the point spread function at the target (e.g., bar code) . In practice, the spot at the focal point 406 (or "waist") may not be of negligibly small size, but rather may be somewhat larger due to the effects of diffraction and lens aberrations. The size of the lens aperture affects the spot size resulting from diffraction, which may occur when only a portion of a wavefront of light is allowed through the aperture. Larger apertures generally result in a smaller spot size, while smaller apertures result in a larger spot size. The effects of diffraction also depend on the wavelength of the light, whether ambient or self- generated. For example, blue light tends to create a smaller spot size than red light. In an optical model for CCD-based image reading systems, the effects of focus blur and diffraction spot size may estimated as a root mean square (i.e., the square root of the sum of the squares of both spot sizes) .
In addition to diffraction, lens aberrations (i.e., deviations from ideal lens shape) may lead to larger waist spot sizes. The resulting waist spot size may be expressed as D(l+K), where D is the diffraction limited spot size and K is an aberration factor. For an aberra¬ tion free lens, K=0. There are several types of aberrations, including spherical aberration and chromatic aberration. Spherical aberration is the primary type of aberration with large aperture lenses (larger than f/3) . Spherical aberration arises from the difference between a perfect shaped lens and a spherical shaped lens. To form a perfect (diffrac¬ tion limited) image, a spherical wavefront is required. However, a spherical lens shape is not the correct shape to create a spherical wavefront. The correct lens shape depends on the object-to-image distance ratio (referred to as the magnification ratio or conjugate ratio) . The difference between this perfect shaped lens and a spherical shaped lens causes spherical aberration. The effect of spherical aberration, as shown in Fig. 14A, is an effective shift in focal length depending on where radially on the lens the light passes (i.e., ray height) . The region of the lens near the center has one focus, while concentric rings of the lens radiating outward from the center will have differing focal lengths depending on their distance from the center. Spherical aberration can be minimized or eliminated by using a properly shaped spherical lens (e.g., plano-convex, double-convex, meniscus) , shaping the lens correctly (such as with aspheric lenses) , or by using multiple lenses (e.g., doublets or triplets).
Chromatic aberration relates to the variation of focal length as a function of wavelength. The focal length variation occurs because the index of refraction of an optical substance varies with wavelength (an effect called dispersion) . Chromatic aberration is the primary type of aberration with apertures smaller than f/2, when astigmatism is not present. Chromatic aberration can be corrected by using two different lenses with different dispersion properties.
Coma relates to a variation in magnification as a function of ray height when imaging objects not on the optical axis. The effect of coma, illustrated in Fig. 14B, for example, is typically overshadowed by the effect of astigmatism, but can nevertheless be corrected by proper placement of the aperture.
Astigmatism is the difference in focus between the horizontal and vertical axis. The effect of astigmatism, illustrated in Fig. 14C, for example, is to cause the image of tangential lines (i.e., vertical axis) to appear at one distance and radial lines (i.e., horizontal axis) to appear at a different distance. Lenses with or without astigmatism will generally exhibit field curvature, which is the bending of the image plane. Field curvature is characterized by different tangential or sagittal lines, as shown in the comparative diagram of Fig. 14D. In a non-astigmatic system, the tangential field curvature and sagittal field curvature are the same and lie on the Petzval surface, while in an astigmatic system they are different. Another consideration is cosine law effect, a term which describes the variation of light intensity across an imager. Cosine law effect is most pronounced with a wide angle field of view. Several factors each having a "cosine" component combine to form an intensity distribu- tion that varies as the cosine to the fourth power as a function of field angle. Wider field angles thus produce larger variations in signal intensity over the imager, while narrower field angles produce smaller signal intens¬ ity variations. Cosine law effect should be taken into account so as to collect enough light (i.e., obtain a suf¬ ficient signal level) at the edges of a virtual scan line. A graph of the cosine law effect is shown in Fig. 15.
The amount of light collected by the imager 402 depends in part on the intensity of the light (whether ambient or not) and the reflectivity of the bar code label. A light collection model is illustrated in Fig. 16. If a target label 410 with reflectivity r is illumi¬ nated with light 411 having an irradiance Ii lumens/m2, then Ii*r lumens/m2 will generally be re-radiated from the label. A Lambertian (perfectly diffuse) source that emits N lumens/m2 into a hemisphere emits N/π lumens/steradian-m2 at normal incidence to the source. Therefore, the radi¬ ance of the target label (at normal incidence) having irradiance Ii-r may be expressed as
Figure imgf000032_0001
lumens/ steradian-m2. An imaging system will receive an,irradiance at the imager 402 equal to ιτTN'sin2θ, where N' is the object radiance in lumens/steradian-m2, T is the trans¬ mission of the system, and θ is the half angle from the center of the imager to the edge of the lens. Recalling that the imager distance I can be expressed as I=XFf/ (XF - f) , where f is the local length and XF is the distance to focus, the amount of collected light Ic in lumens/m2 is therefore given as follows:
Ic = I1-rTsin2θ and θ = Arctanf (XF - f)/2AXF}
For small aperture size:
IC=IX- (rT/4A2) • (XF - f)2/XF 2 Where lt = Incident Light (lumens/m2) Ic . Collected Light (lumens/m2) r Label Reflectance (percentage)
T = Lens Transmission (percentage) θ = Aperture Half Angle (degrees) A = f/# of Aperture (unitless) f = Focal Length of Lens (meters) XF = Distance to Focus (meters)
Thus, the amount of light collected at the imager 402 is independent of distance of the target label from the lens 401. The collected light is, however, proportionally affected by the intensity of the light (ambient or other- wise) illuminating the label, the reflectance of the label, and the lens transmission. The collected light is inversely proportional to the f-number (f/#) squared, and depends to some degree on the position of focus (XF) relative to the focal length. To determine the modulation level of the CCD video output signal, the collected light Ic may be multiplied by the print contrast ratio of the target label. It should be noted that the above calculations do not take into account any signal modulation reduction resulting from modulation transfer function (MTF) effects or cosine law effects.
Further information relating to optics and optical systems may be found in W.J. Smith, Modern Optical Engineering -- The Design of Optical Systems (McGraw-Hill, 2d ed. 1990) . A convenient model for estimating optical system performance and determining parameters for optimum lens design is given as follows:
Focus Blur (mils) = V [(Foe)2+(Diff)2] where Foe is geometric focus blur, and Diff is diffraction limited minimum spot size;
Foe (mils) = 1000- [f/(f/#)] • [(R - R0)/R0] where R is distance from lens (inches) , R0 is focal point (inches) , f is focal length of lens (inches) , and f/# is relative aperture (f-stop) ;
Diff (mils) = 9.61xl0"5λR- (f/#)/f where R is distance from the lens (inches) , λ is nominal wavelength (nanometers) ; f is focal length of the lens (inches) , and f/# is relative aperture (f-stop) ; Resolution (mils) = 1000WR/ [I (f,R0)N] - 1000WR/ (fN) where R is distance from the lens (inches) , W is width of the CCD array (inches), I(f,R0) is CCD image plane (inches) , and N is the number of pixels in CCD array,- Field of View (mils) = 8.333 WR/[I(f,R0)] - 8.333 WR/f where R is distance from the lens (inches) , W is width of the CCD array (inches), and I(f,R0) is CCD image plane (inches) . Based on the above model, other optical parameters may be determined as follows:
Diffraction = 2.44λAX/i Minimum Spherical = 0.067X/A3
Chromatic = X/(2-V-A)
Coma = Θ-X/(56A2)
Astigmatism = Θ2-X/(2A) where λ is the wavelength in meters, A is the relative aperture (f/# or f-stop) , V is the Abbe number, and θ is the half field angle in radians.
Typically, at least 2.5 pixels per minimum element width are needed to resolve bar codes. Focus blur can frequently be 1.9 times bigger than the minimum element width. Spots as big as 2.4 times the minimum element width may still provide workable results in low noise systems. Certain of the preferred embodiments make use of shaped apertures and specialized lenses in order to increase light gathering capability and allow operation in reduced light, such as in ambient light conditions. Since in many applications a bar code reader images only bars/ spaces and not necessarily true two dimensional images, some optimization of the lens aperture is possible. Increasing the size of the aperture in the direction of the bars generally provides greater collection with minimal effect on the depth of field characteristics. While this is strictly true when the bars are parallel with respect to the aperture, as the bars are rotated when reading the bar code at an angle the elliptical point spread "spot" shape grows with respect to the bar code. This phenomenon is illustrated in Fig. 17. Figure 17 is a diagram comparing the point spread function shapes through focus for a round aperture 430 and an oblong or elliptical aperture 431, both having the same light collection. As shown in Fig. 17, at a focal point 435, the round aperture through focus point spread spot 436 and the elliptical aperture through focus point spread spot 437 are each at a minimum. However, the elliptical aper¬ ture through focus point spread spot 437 is larger in the lateral direction than the round aperture through focus point spread spot 436.
The elliptical spot size changes with respect to the tilt angle, as illustrated in Fig. 18. In Fig. 18, wc represents the narrowest width of the spot, θ represents the tilt angle (e.g., with respect to the vertical bars of a bar code) , and w represents the projected width. The projected width may be characterized by the following equation: w = { (e2-w0 2)/[(e2 - 1) • (cos θ)2 + 1] }12 where e represents the ratio of height to width. The area of the elliptical spot is given by the following equation:
Area = π•e•w 2 The above equation indicates that the increase in area of the elliptical spot is directly proportional to the height to width ratio.
A narrower aperture, such as the elliptical aperture 431 (or other oblong aperture, such as a rectangular aper¬ ture) shown in Fig. 17, generally tends to increase the depth of field. The height of the aperture 431 in the direction parallel to the bars is enlarged to equalize the collection. Figure 19 is a graph showing illustrative depth of field plots for a round aperture 430 and for an elliptical aperture 431 collecting the same amount of light. The graph of Fig. 19 plots the minimum resolvable bar width on one axis 440 against the distance from the lens on the other axis 441. As shown in Fig. 19, the narrower aperture creates a larger depth of field than the wider aperture, but has a larger minimum spot size due to diffraction.
An optical system employing a small, round aperture (e.g., aperture 430) generally has a relatively good depth of field but suffers from the need for a long exposure time. Such a system is shown, for example, in Fig. 20A. In Fig. 20A, light is reflected from a target 444 (e.g., a bar code, shown from a back view) and focused through a round lens 443 (or lens having a round aperture) onto a CCD detector 442. In the optical system of Fig. 20A, as little as 0.02% of the reflected light may reach the CCD detector 442.
An oblong aperture (e.g., aperture 431) speeds up exposure time by allowing more light collection. A system comprising and oblong aperture is shown, for example, in Fig. 20B. In the optical system of Fig. 20B, light is reflected from a target 458 (e.g., a bar code, shown from a back view) through an apertured lens 457 (or an oblong or similarly shaped lens) onto a CCD detector 456. In one embodiment using a rectangular-shaped aperture 445, a narrow axis 446 of the aperture 445 corresponds to an f- stop for the desired depth of field (e.g., f/20) , while a long axis 447 of the aperture 445 has a larger f-stop (e.g., f/5) and therefore gathers more light. Although the range of angles of bar codes that the optical system of Fig. 20B will detect may be reduced in comparison to the Fig. 20A system by use of a oblong aperture, the collected light can be substantially increased (e.g., by as much as ten times or more) .
A preferred embodiment of a CCD imaging device employs a CCD detector 1079 as shown in Fig. 50A whose pixels 1080 are rectangular in shape, with their long axes perpendicular to the line formed by the linear array of pixels 1080. Like the oblong-shaped aperture 431 described above in Fig. 17, the Fig. 50A rectangular pixel structure enhances the detection efficiency of the CCD detector 1079 without significantly degrading spatial resolution, so long as the CCD detector 1079 is not rotated at too much of an angle with respect to the barcode (or other image) being read. Figures 50A and 50B are diagrams illustrating the effect on an output signal 1082A, 1082B of rotating a bar code with respect to the CCD detector 1079. As shown in those figures, the unrotated bars 1081A projected onto the CCD detector 1079 in Fig. 50A result in a relatively clean square-wave output signal 1082A. In contrast, the partially rotated bars 1082A projected onto the CCD detector 1079 in Fig. 50B result in an output signal 1082B with sloped rising and falling edges due in part to the spreading of the bars 1081A across a greater number of pixels 1080.
To reduce the effect of image rotation in the Fig. 50A embodiment having rectangularly shaped pixels 1080, the pixel aspect ratio should be selected so that the resolution of the bar code (or other image being read) is not impaired with labels that have a tolerable amount of roll with respect to the imaging device. Moreover, if the pixels are too tall, the projected image of the pixels will provide too little resolution to properly image the barcode. Figure 51 is a graph comparing pixel width versus roll angle for a CCD array with square pixels and with rectangular pixels (having a 4:1 aspect ratio). As shown in Fig. 51, the angle of bar code label rotation does not significantly affect the normalized projected pixel width for a CCD array with square pixels; however, it does affect the normalized projected pixel width for a CCD array with rectangular pixels having a 4:1 aspect ratio. Thus, it is possible that more rectangular pixels would be needed to obtain the same performance as square pixels for handling rotated bar code labels.
The maximum acceptable aspect ratio K^* of pixels 1080 for a given maximum roll (0) and minimum number of pixels per bar (N) can be determined mathematically, according to relationships expressed in the following equations: Wβff(β, K) = K-sin(0) + cos(0) (51.0)
N = Wef£(0, K-f)-cos(0) (51.1) where We£f is the projected or effective width (i.e., how big the pixel width looks with a certain rotation) , and K is the aspect ratio. The value f is a factor reflecting the tolerance of the signal processing electronics to a distorted output signals. For a derivative based signal processor (described elsewhere herein) , f is generally equal to about 0.8. The minimum number of pixels per bar
(N) is typically at least 2.5. According to equation 51.0, the effective pixel width We££ changes according to the rotational angle θ . For θ = 0°, Weff = 1, while for 0 = 90°, Weff = K. For 0° < θ < 90°, Weff varies according to the graph of Fig. 51. According to equation 51.1, as the bar code is rotated it looks bigger by a factor of cos 0, in partial dependence on the effect of the signal processing (as indicated by the factor f) . The more pixels 1080 per bar are allowable (i.e., the larger N is), the taller pixels 1080 can be, thus increasing sensitivity to incoming light. An example wherein the maximum tolerable roll (0) is ±23° with respect to the imager, and wherein f = 0.8 and the number of pixels per bar N is 3.75, the maximum acceptable pixel aspect ratio is found to be:
Kmax(N,0,f) = Kmx (3 .15 , 23°, 0.8) = 10 (51.2) Thus, in the above example, the maximum acceptable pixel aspect ratio is about 10 to 1.
Use of rectangular pixels such as described above with respect to Figs. 50A and 50B is not limited to CCD devices, but may be applied to any pixel-based photodetection device. If an existing CCD detector 1079 has an aspect ratio that is considered larger than optimum to handle rotated bar code labels at all tolerable angles, there are several ways to modify the system to provide a more acceptable aspect ratio without requiring the CCD detector 1079 itself be modified. For example, as shown in Figs. 52A and 52B, the CCD detector 1079 may be transformed into a CCD detector 1090 having a more acceptable aspect ratio by use of a metal mask 1091 over the pixels. The metal mask 1091 has a slit 1093 which allows light to reach only the portions of the CCD pixels conforming to the height of the mask, thereby allowing relatively simple mechanical adjustment of the effective height of the CCD pixels.
Use of a metal mask 1091 may prove to be more economical than use of a different CCD detector having the proper pixel aspect ratio. The mask 1091 can be applied externally in front of the CCD detector 1090, as shown in Figs. 52A and 52B. The mask 1091 is preferably physically close (e.g., touching) to the imager plane so as to reduce the effect of vignetting, which would cause light rays to spill over into regions of the CCD photodetectors not intended for light collection. Figure 52B shows a flux density graph 1092 adjacent to the CCD detector 1090, which illustrates a gradual roll off of flux density caused by small amounts of vignetting even with a properly positioned mask 1091. If, however, the mask 1091 is too far away, it ceases to be useful for reducing the pixel aspect ratio and simply reduces light collection efficiency.
In a preferred embodiment, the top layer metal mask of the CCD chip itself is modified so as to provide an appropriate sized slit 1093 in front of the CCD photo¬ detectors. For example, with a Sony ILX-526 CCD chip, a preferred mask slit on the top layer metal mask of the chip has a height dimension and a width dimension so as to provide a 10:1 pixel aspect ratio. An alternative modified CCD imaging system with an external mask is shown in Fig. 52C, wherein extra optics are used to re-image the pixels to a new imaging plane. In the Fig. 52C imaging system, an imaging lens 1097 is positioned in front of a mask 1096 (opposite the CCD array 1098) having a slit 1099. The mask 1096 is positioned further away from a CCD detector 1098 than the Fig. 52A- 52B system. In between the mask 1096 and the CCD detector 1098 is positioned a second lens 1095. The second lens 1095 reduces or eliminates vignetting by re-directing light focused by the imaging lens 1097 onto the CCD detector 1098. Hence, the corrective mask 1096 is more effective in this system than in the system of Figs. 52A- 52B, and the problems associated with vignetting are less of an issue. The Fig. 52C system does, however, have the additional complexity of a second lens and a longer path length for the optical system.
Another alternative for modifying pixels of a CCD detector with too large an aspect ratio is an anamorphic lens system, such as shown in Fig. 53. The Fig. 53 system includes a CCD detector 1100 in front of which is posi¬ tioned a first aspheric cylinder lens 1102 in a vertical orientation with respect to the CCD detector 1100 and a second aspheric cylinder lens 1104 in front of the first aspheric cylinder 1102 and in a horizontal orientation with respect to the CCD detector 1100. Each aspheric cylinder lens 1102, 1104 provides significant modulation along one axis only. The horizontally oriented aspheric cylinder lens 1104 is provided with a lower magnification power than the vertically oriented aspheric cylinder lens 1102, thereby reducing the effective aspect ratio of the pixels 1106 of CCD detector 1100. An element of height HI and width Wl (i.e., aspect ratio of H1:W1) is therefore, by the operation of aspheric cylinder lenses 1102 and 1104, altered in apparent dimension at the CCD detector 1100 to have an aspect ratio of H2:W2. Alternatively, the "projection" of the pixels 1106 of the CCD detector 1100 onto the bar code plane are reduced in height from H2:W2 to H1:W1, where H1/H2 < W1/W2.
The distance from the CCD detector 1100 to the vertically oriented cylinder lens 1102 is given as Il r and the distance from the CCD detector 1100 to the horizont- ally oriented cylinder lens is given as I2. Likewise, the distance from the bar code 1105 to the vertically oriented cylinder lens 1102 is given as 01# and the distance from the bar code 1105 to the horizontally oriented cylinder lens is given as 02. The height reduction of the system is Mj/ i, where x = O-Vlj and M2 = 02/I2. Exemplary distance relationships between the various components are given in Fig. 53 for a height reduction of 1/3.
A possible drawback with the Fig. 53 system is that aspheric cylinders may be hard to manufacture. A preferred embodiment of the present invention also makes use of a multi-focus lens system (e.g., a poly-optic system) . Generally, use of a multi-focus lens allows an increase in depth of field without sacrificing the ability to collect light. Each lens element (called a zone) of a multi-focal lens is focused at a different distance. In one embodiment, further described herein, the field from each zone is simultaneously focused on the CCD detector. In another embodiment, also described further herein, each zone is multiplexed (with an LCD shutter, for example) on the CCD detector instead of simultaneously superimposing all zones. In yet another embodiment, also described further herein, each zone is imaged onto one of a plurality of separate CCD detectors.
An exemplary embodiment of a multi-focus lens system combined with an elliptical aperture is shown in various views in Figs. 22A-22C. The symmetric multi-focal lens 450 has concentric zones and is depicted by itself in Figs. 21A and 21B. Figure 21A shows the multi-focus lens 450 from an oblique view, while Fig. 21B shows the same lens 450 from front view. As shown, the lens 450 is divided into a plurality of concentric zones 451. Such a lens 450 has a symmetric configuration and may be suitable for machining with a diamond turned lathe.
The lens and aperture system of Figs. 22A-2C are shown from three different views. Figure 22A shows a side view, Fig. 22B shows a top view, and Fig. 22C shows an oblique view. The lens and aperture system of Figs. 22A- 22C comprises an elliptical (or rectangular) aperture 461 located in front of a multi-focus lens 460. The ellipti¬ cal aperture 461 preferably has a high aspect ratio. The elliptical aperture 461 and multi-focus lens 460 are placed before an imager 462, such as the CCD detector described earlier herein. In operation, light reflected from a bar code 463 passes through the elliptical aperture 462 and is focused by the multi-focus lens 460. Each zone 465 of the multi-focus lens 460 focuses the reflected light 466 in a different manner on the imager 462, because each zone 465 focuses at a different distance (i.e., has a different focal length in front of the lens 460) . The depth of field is thereby expanded over that of a single- focus lens.
The response of a multi-focus lens (such as lens 450) may be determined under principles of superposition by summing the responses of each lens zone 451 by itself. If a shutter is used to screen out zones 451 of the ulti- focus lens 450 (e.g. by use of an LCD) , only one zone 451 (or however many zones are unscreened) at a time need be analyzed, and the result of each individual zone super¬ imposed to obtain a composite response for the lens 450.
Figure 23A is a graph showing the minimum renderable (i.e., resolvable) element width as a function of distance from the lens in a single-lens system. Figure 23B is a graph showing minimum renderable bar width as a function of distance from the lens for a poly-optic (i.e., multi- focus) lens system, assuming no interference. Each parabolic-shaped plot 471 represents a particular zone 451 of the poly-optic lens 450. A focal point 470 corresponds to the minimum point of the parabola. The plots 471 are superimposed on one another to graphically show the crea¬ tion of an enlarged depth of field for the poly-optic system. However, it is also possible that the different zones 451 may interfere with one another, causing "dead zones" in the depth of field plot. These dead zones 472 are depicted graphically in Fig. 23C, which is a similar plot to Fig. 23B but assumes that there is intra-lens interference from the poly-optic lens 450. Interference between zones 451 may be caused by lobes in the line spread function of the lens 450, by mis¬ alignment of the optical centers of the zones 451, or a difference in the field of view of each zone 451. Different lens configurations generate different lobe patterns. Some point spread functions for a few common poly-optic lenses are shown in the Figs. 24A through 30C and described below.
A simple multi-focus lens 450 configuration comprises a set of concentric zones 451, each with its own focal length, as shown previously in Figs. 21A-21B. The point spread function through focus for the center zone 452 of the Fig. 21A lens is shown in Fig. 24A at focus and Fig. 24B away from focus. The point spread function through focus for a particular ring zone 454 (e.g., any other zone 451 besides the center zone 452) is shown in Fig. 25A at focus and Fig. 25B away from focus. The diagrams of Figs. 24A and 25A are full scale, while the diagrams of Figs. 42
24B and 25B are reduced relative thereto by a factor of 10:1.
Figures 26A through 30C depict point spread and line spread function plots of various combinations of center and ring zones of various focal lengths. The optical point spread function images appearing in Figs. 26A through 30C have been approximated for simplicity. In general, the point spread function far from focus can be well approximated by geometric optics. For a circular aperture, the shape of the point spread function far from focus is similar to a cylinder. At or near focus, the point spread function's shape is determined by aberrations
(if they are large) or by diffraction (if the aberrations are small) . The point spread functions depicted in Figs. 26A through 30C are primarily diffraction limited (i.e., they assume only small aberrations) at or near focus. For a circular aperture, the shape of the point spread func¬ tion at or near focus should be close to the shape of an Airy disk. Figure 26B is a diagram of one image 480 created from a center zone 452, while Fig. 26A is a diagram of another image 481 created from a ring zone 454, where the center zone 452 and ring zone 454 have differing focal lengths. Figure 26C is a diagram showing a combined image 482 created from the combination of the center zone 452 with the ring zone 454 with only the center zone 452 in focus. In Fig. 26C, the image from the ring zone 454 is not readily visible mainly due to its low flux density, since it is out of focus. However, the plots of the line spread functions 483 and 484 clearly show the effect of the ring zone 454. The line spread functions 483, 484 each have a main peak 488 and 489, respectively, corresponding to the center zone 452 and a pair of much lower peaks 490 and 491, respectively, corresponding to the ring zone 454. As the center zone 452 goes out of focus, the effect of the ring zone 454 will become considerably more pronounced. Figures 27A-27C illustrate point spread functions in between the focus of both zones. Figure 27B is a diagram of an image 500 created by the center zone 452, while Fig. 27A is a diagram of an image 501 created by a ring zone 454, assuming differing focal lengths. Figure 27C is a diagram of a combined image 502 created from both the two zones 452, 454, while in between the focus of both. It may be noted that, while there is a "dip" 503 or dead spot in the center of the combined image 502, the line spread functions 505, 506 are relatively free of the effect of this dip 503.
Figures 28A-28C illustrate point spread functions when the ring zone 454 is in focus. Figure 28B is a diagram of an image 510 created by the center zone 452, while Fig. 28A is a diagram of an image 511 created by a ring zone 454, assuming differing focal lengths. Figure 28C is a diagram of a combined image 512 created from both the two zones 452, 454. The combined image 512 of Fig. 28C is very similar to that of Fig. 26C, wherein only the center zone 452 was in focus. The line spread functions 515, 516 in Fig. 28C, however, show the effect of the out- of-focus center zone 452.
Figures 29A-29C illustrate a more severe effect, wherein a ring zone 454 is not blurred enough with respect to the center zone 452. The result is an image 522 having a ring structure. Figure 29B is a diagram of an image 520 created by the center zone 452, while Fig. 29A is a dia¬ gram of an image 521 created by the ring zone 454. Figure 29C is a diagram of a combined image 522 created from both the two zones 452, 454 contains the ring structure, as mentioned, which appears as diffuse light in a ring pat¬ tern around the center focal point. The line spread func¬ tions 525, 526 indicate the effect of this ring structure. The ring structure may cause excessive inter-symbol interference (ISI) when attempting to read bar codes.
Because of the ISI that may be caused by a ring zone such as shown in Fig. 29C, an apertured ring zone (a sector) is preferred for reading bar codes. Figures 30A- 30C are diagrams of images created under the same condi¬ tions as Figs. 29A-29C, but for a system where the ringed zone is apertured with two circular apertures. Figure 3OB is a diagram of an image 530 created by the center zone 452, while Fig. 30A is a diagram of an image 531 created by the apertured ring zone 454. Figure 30C is a diagram of a combined image 532 created from both the two zones 452, 454. The use of two circular apertures as shown in the horizontal line spread function 536 of Fig. 30C successfully removes the side lobes in the horizontal pattern. This form of aperturing can thus be used to read bar codes whose bars are vertically oriented without creating an undue amount of inter-symbol interference from the multi-focus lens.
Figure 31A-31C are diagrams illustrating a particular apertured multi-zone lens configuration providing a line spread function similar to that in Fig. 30C. In Figs. 31A-31C, a symmetric multi-focus lens 601 is shown from several viewpoints with a plurality of apertures 602, one aperture 602 located in front of each one of a plurality of zones 603 the lens 601. While Fig. 30C shows an image resulting from two circular apertures, the apertured multi-focus lens 601 of Figs. 31A-31C provides a plurality of images from the multiple ring zones and a corresponding plurality of peaks in the vertical line spread function 535. However, the apertured multi-focus lens 601 of Figs. 31A-31C would, like the configuration of Fig. 30C, also have a horizontal line spread function 536 relatively free from intersymbol interference.
The apertures 602 of Fig. 31A or 3IB need not be circular, but may be square, rectangular, or any other suitable shape. In a preferred embodiment, each of the apertures 602 are roughly square in shape. When physic- ally manufactured the square apertures can be melded into a continuous opening similar to that of the aperture 461 in Figs. 22A-22C. While such a continuous opening may be "rectangular" in overall shape, it functions as a series of discrete, individual apertures, one for each lens zone. The selection of the height of the apertures 602 in front of each zone may be a compromise balancing light collection requirements and roll response characteristics. The width of each aperture 602 controls the depth of field for its lens zone, and is preferably the same for each aperture 602 so as to minimize interzonal interference.
It is preferable for each of the apertured zones 603 to have a different focal length. This may be achieved in at least two different ways. In a first technique, a symmetric lens 610 is physically cut immediately below the center zone 611, as shown in Fig. 32. The cut lens 612 will have the characteristics of an asymmetric multi-focus lens. In a second technique, the lens 620 may be molded in an asymmetric fashion, as shown in Fig. 33, such that the lower portions 621 of each ring zone have a different focal length than the upper portions 622 of each ring zone. An advantage of the latter technique is that the lens could be constructed so as to have its optical axis 625 in the center of the lens 620, as shown in Fig. 33.
For the above asymmetric lens configurations, the off-axis tilt is typically smallest when the center zone 627 focuses to the closest depth and the outside ring zone 628 focuses to the farthest depth, as shown in Fig. 34A. Figure 34B illustrates the reverse situation, which may lead to greater pointing problems.
An alternative optical collection system employs a "masked" or apodized multiple focus lens, such as shown in either of Fig. 54A or 54B. In Fig. 54A, a multiple-zone lens 1120 is shown having concentric ringed zones 1121a through 1121d, and further having darkened portions 1122 which are masked or apodized. The apodized portions 1122 can be created by depositing silver or other reflective material on the lens 1120. Likewise, in Fig. 54B, a multiple-zone lens 1130 is shown having concentric ringed zones 1131a through 1131d, and also having apodized portions 1132. The apodized lenses 1120, 1130 of Figs. 54A and 54B provide a multiple focal plane optical col¬ lection system which maintains high collection efficiency while reducing interference between light rays from the different focal planes that are simultaneously incident on a CCD detector. The masking can therefore replace the aperture(s) of the embodiments shown in Figs. 22A-22C and 31A-31C, inasmuch as the masking has a similar effect as the aperture(s) in reducing inter-zonal interference. The Fig. 54A and 54B lenses 1120, 1130 preferably each comprise a set of concentric annular zones 1121a-d and 1131a-d, respectively, having varying focal lengths. The area of each zone 1121a-d and 1131a-d is preferably substantially equal, so that substantially equal amounts of light are collected from each zone. This condition requires that rN = rcβnterVN for an unmasked lens 1120 or 1130, where rN is the outer radius for zone N, and rcenter is the radius of the center zone 1121d or 1131d. The radii required to maintain constant collection area from zone to zone will change in a straightforward fashion from this equation to account for the reduction in area due to the masking.
Figure 55A is a graph of individual line spread functions 1141 through 1144 for each lens zone 1121a-d or 1131a-d of the multi-focal lenses of either Fig. 54A or 54B. Lobes that would otherwise exist in the horizontal line spread functions are reduced or eliminated by use of a mask (e.g., apodization on the lens itself) to maintain a constant vertical height for each zone 1121a-d or 1131a- d of the lens 1120 or 1130. The resulting composite or vertically-integrated horizontal line spread function 1146 is shown in Fig. 55B. As shown in Fig. 55B, the resulting vertically-integrated horizontal line spread function 1146 is relatively flat, without lobes. The line spread function varies depending on the focal distance and the effect of aberrations and diffrac¬ tion. For the regions of depth of field (i.e., non- intermediate regions) where geometric optics prevails over aberrations and diffraction, the Fig. 54A and 54B masking will provide a composite line spread function as shown in Fig. 55B. Near a focal plane, however, aberrations and diffraction may prevail over geometric optics causing reduced performance. Many of the families of masks that work well in near and far field will not work well at intermediate distances. Preferably, an optimal mask pattern is selected which minimizes sidelobes over a maximum range of distances. The optimal mask pattern selection depends upon a variety of factors, including the number and position of focal planes present in a particu¬ lar optical collection system, the degree of aberration correction required, the angular field of view required, and/or the size of the barcodes to be read.
In addition to the poly-optic lenses of Figs. 21A-21B and the apodized poly-optic lenses of Figs. 54A-54B, another method for constructing multi-focus lenses having the desired properties is to combine a plurality of indi- vidual lenses, one for each zone. Each individual lens has an off-center circular aperture and is optimized for the desired focus condition of the corresponding zone. Each lens zone in turn is optimized. The final lens essentially comprises a piecewise continuous combination of each individual lens zone.
It may, on the other hand, be possible to make a multi-focus lens with just one continuous lens curvature. In such a case, the lens can be optimized in at least two ways. First, a global merit function can be made that takes into account the focal properties of each zone. The entire aspheric deformation coefficients can then be solved at once. The second method is to allow each zone a few deformation coefficients, giving the center zone the lowest order coefficients and the edge zone the highest order coefficients. The system is optimized in order, from the center zone to the edge. This method involves an iterative process, as the edge zones may need to cause slight modifications to the lower order coefficients so as to minimize aberrations.
Several particular optical structures are discussed below, many or all of which exemplify the advantages of poly-optic design thus far described.
One particular optical lens structure is shown in Fig. 35A. In Fig. 35A, which is drawn from a top view, an array 640 of single-focus lenses 641 are arranged in an arc-like pattern around an imager (e.g., a CCD detector 642) . Each lens 641 is a slightly different distance away from the CCD detector 642, so the pattern formed by the lens array is not a perfect arc. Alternatively, a set of different focal length lenses could be aligned in an arc to achieve the same result. The centers of the scan lines 643 associated with each of the lenses 641 fan out in a radial fashion from the CCD detector 642. The focal points 644 of the lenses 641 are staggered, so that the top lens 641a has the closest field depth 645a (because it has the nearest focal distance) , the next lens 641b has the next closest field depth 645b, and so on, until the last lens 641e which has the farthest depth of field 645e.
An advantage of the optical structure shown in Fig.
35A is that the structure may provide for a seamless depth of field without using shutters, so long as the lenses 641 are arranged so as not to cause interference with one another. Another advantage is that the structure mitigates the effect of coma, due to the radial fanning of the centers of the scan lines 643.
A related embodiment of an optical structure is shown in Fig. 35B. The Fig. 35B structure is similar to the Fig. 35A structure, except that the Fig. 35B embodiment is shown from a side view instead of a top view, so that the array 650 of lenses 651 expands the depth of field vertically rather than horizontally. The Fig. 35B structure has similar properties of an arc-like pattern of lenses 651, radially fanning scan lines 653, staggered depth of fields 655, and a corresponding mitigation of the effect of coma and aggregate expansion of depth of field for the system.
Another embodiment of an optical structure is shown in Fig. 36A, which is a modification of the system of Fig. 35B. Like Fig. 35B, the Fig. 36A embodiment depicts an array 660 of lenses 661 arranged in an arc-like pattern around a CCD detector 662. Scan lines 663 fan out radially from the CCD detector 662. The Fig. 36A embodiment further adds a set of mirrors 667 arranged to form the radial scan lines 663 into a collimated set of scan lines 668. The individual depths of field 665 are staggered for each lens 661 along the collimated scan lines 668. A composite depth of field 669 is shown adjacent to the collimated scan lines 668. An advantage of the Fig. 36A embodiment is that it provides a more dense coverage of a plane. In place of mirrors 667, prisms or light-directing lenses may be used instead, or any combination of mirrors, prisms and/or light directing lenses. Figure 36B shows the Fig. 36A embodiment implemented using prisms 677 instead of mirrors. Figure 36C shows the Fig. 36A embodiment implemented using a single collimating lens 687 to perform a similar function.
Another embodiment of an optical structure is shown in Fig. 37A. The Fig. 37A structure is similar to that of Fig. 36C, but reverses positions of the large single lens 707 and the array of lenses 701. A large single collimat¬ ing lens 707 is located in front of an imager (e.g., a CCD detector 702) . A plurality of smaller lenses 701 are positioned in front of the larger lens 707, and are staggered slightly in distance from the larger lens 707. Scan lines 708 from the smaller lenses 701 are directed in a collimated pattern to provide dense coverage of a plane. The scan lines 708 are directed by the larger lens 707 to the CCD detector 702. Each of the smaller lenses 701 has a different depth of field 705, so that a larger composite depth of field is created. Another embodiment of an optical structure is shown in Fig. 37B. The structure of Fig. 37B is conceptually similar to that of Fig. 37A. Figure 37B, like the structure of Fig. 37A, comprises a large collimating lens 717, convex in shape, positioned in front of a CCD detector 712. The Fig. 37B structure also comprises an array of smaller lenses 711. Rather than staggering the distance of the smaller lenses 711 from the larger lens 717, however, in Fig. 37B is added a set of staggered optical flats 719 positioned between the larger lens 717 and the array of smaller lenses 711. The staggered optical flats 719 shift the focal point of the smaller lenses 711 and provide staggered of depths of field 715 along the collimated scan lines 718 associated with the array of lenses 711.
An advantage of the Fig. 37B system is that the smaller lenses 711 need not be staggered, and therefore may facilitate manufacture of the optical system.
Another embodiment of an optical structure is shown in Fig. 37C. The structure of Fig. 37C is similar to that of Fig. 37B, except that a graded index element (also called a "GRIN" element) 729 is used in place of the staggered optical flats 719. An array of smaller lenses 721, as in Fig. 37B, is non-staggered. The smaller lenses 721 are positioned in front of the GRIN element 729, which is positioned in front of a large collimating lens 727 set before the CCD detector 722. Scan lines 728 form a colli¬ mated pattern, providing dense coverage of a plane. The GRIN element 729 is formed of a material having a graded index, so that the path length differs along its axis. Thus, towards one end 730 of the GRIN element 729, the path is long, while towards the other end 731 of the GRIN element 729, the path is short. Accordingly, the field depths 725 are staggered along the collimated scan lines 728, as shown.
Figure 38A shows another embodiment of an optical structure. The Fig. 38A structure is similar to that of 51
Fig. 37B, except that some of the individual elements of Fig. 37B have been incorporated into a single element. A poly-optic lens 751 incorporates aspects of the large collimating lens 717, the staggered optical flats 719, and the array of smaller lenses 711 shown in Fig. 37B into a single structure. The back portion 753 of the poly-optic lens 751 is curved in the shape of the collimating lens 717 of Fig. 37B. The array of smaller lenses 711 from Fig. 37B appear as individual lens elements 752 in Fig. 38A. The individual lens elements 752 are staggered, so that varying amounts of optical material are located between each individual lens element 752 and the collimated lens backing 753. This varying amount of optical material acts in a fashion similar to the staggered optical flats 719 of Fig. 37B, thereby providing for differing focal lengths and staggered field depths 755 along collimated scan lines 758. The poly-optic lens 751 of Fig. 38A therefore has a composite depth of field similar to that of Fig. 37B. Another embodiment of an optical structure is shown in Fig. 38B. The Fig. 38B structure is similar to that of Fig. 37C, except that some of the individual elements of Fig. 37C have been incorporated into a single element. A poly-optic lens 761 incorporates aspects of the large collimating lens 727, the GRIN element 729, and the array of smaller lenses 721 shown in Fig. 37C into a single structure. The back portion of the poly-optic lens 761 is curved in the shape of the collimating lens 727 of Fig. 37C. The array of smaller lenses 721 from Fig. 37C appear as individual lens elements 762 in Fig. 38B. A GRIN ele¬ ment 764 is embedded between the individual lens elements 762 and the collimated lens backing 763, thereby providing for differing focal lengths and staggered field depths 765 along collimated scan lines 768. The poly-optic lens 761 of Fig. 38B therefore has a composite depth of field similar to that of Fig. 37C. While Figs. 36A through 39B are shown from a side view of the CCD detector, the CCD detector may also be rotated 90 degrees so that the lenses or other light collection elements are horizontally rather than vertically stacked, and so that the photodiodes receive the collected light in a manner similar to that shown in Fig. 35A. In general, horizontal stacking results in greater light collection, but vertical stacking results in less interference between lenses or lens elements. Figures 39A and 39B are alternative embodiments of optical structures using curved mirrors or similar ele¬ ments. The optical system of Fig. 39A comprises an array of curved (i.e., focusing) mirrors 790 positioned in a diagonal line 791 at an angle θ with respect to a CCD detector 792. The array of curved mirrors with different focal lengths or different optical path lengths provide staggered field depths 793 along collimated scan lines 794 as shown. The structure of Fig. 39A minimizes the effects of chromatic aberration. Figure 39B shows a similar optical structure to Fig. 39A, but utilizes a single multi-faceted curved surface 796 rather than array of curved mirrors 790. The single multi-faceted curved surface 796 is positioned at a diagonal angle with respect to the CCD detector 792, and thereby provides staggered field depths 797 in a manner similar to the Fig. 39A structure along collimated scan lines 798 as shown. The Fig. 39B structure also minimizes the effects of chromatic aberration.
Another embodiment of a poly-optic lens system is shown in Figs. 40A-40D. Figure 40A is a diagram of a side view of a poly-optic lens 780 positioned in front of a CCD detector 782. As shown, light from a distant point is collected by the poly-optic lens 780 and focused on the CCD detector 782. Figure 40B is a diagram showing a top view of the same poly-optic system. Light from distant points on each of the collimated scan lines is collected by each individual lens element 783, and focused on the CCD detector 782. Preferably, each of the lens elements 783 has a different focal length. Figure 40C is a diagram showing a top view of the same poly-optic system of Fig. 4OB, but shows a target scan line 784 for several of the individual lens elements 783. Light reflected from a target scan line 784 passes through an axial point 785 in an individual lens element 783, as shown, and is projected onto the CCD detector 782. The target scan line 784 for each individual lens element 783 is slightly offset from the other target scan lines 784, due to the relative positions of the lens elements 783. Figure 40D shows an oblique view of the same target scan line capture as Fig. 40C. A preferred multi-focal lens is depicted in Figs. 41A-41C. An oblique view of a preferred lens 801 is shown in Fig. 1A. The lens 801 may be circularly symmetric, as shown. The back surface 801a of the lens 801 is preferably spherical, while the front surface 801b has four different aspheric curvatures as described below. The lens 801 of Fig. 41A may be provided with a rectangular opaque mask which provides a similar function to an aperture. Alternatively, the effect of the aperture can be obtained by cutting the lens 801 to obtain a cut- away lens 810, as shown in Figs. 4IB and 41C. Figure 4IB shows an oblique view of the cut-away lens 810, while the same cut-away portion is shown from a top view in Fig. 41C. The cut-away angle is selected so as to preferably allow for a 45-degree field angle. The lens 801 comprises four zones 802, including a center zone 803 (Zone 1) and three outer ring-shaped zones 804, 805 and 806 (Zones 2 through 4, respectively) . Zone 1 is preferably 0.426 inches wide, while Zones 2 through 4 are preferably 0.213 inches wide. The base radius (i.e., radius of the back spherical portion of the lens) is preferably 84.967 millimeters. The distance from the lens 801 to an imager (such as a CCD detector) is preferably 34.85272 millimeters. Preferred focus distances appear in Table 41-1.
Table 41-1
Zone Focus Distance (mm)
1 48.3
2 76.2
3 122.0
4 191.0
The material of the lens 801 preferably comprises a plastic such as polystyrene, with an index of refraction of 1.59453059. Following are preferred dimensions and other characteristics according to lens zone 802.
Zone 1:
Effective ! Focal Length 23.4508mm
Base Radius 16.80842mm
Conic -0.8092686
Aspheric Coefficients r2 2.57577xl0"3 r4 -1.923724X10-5 r6 5.448919X10"7 r8 -1.765304X10"8 rιo -2.931042X10-10 r12 -5.210294X10-13 r14 1.652903x10-"
-.16 8.778694X10"15
Zone 2:
Effective Focal Length 28.087mm Base Radius 19.64043mm
Conic -1.212117 Aspheric Coefficients r2 4.823852X10-4 r4 -4.101818X10'7 r6 3.891872X10"9
Figure imgf000057_0001
£.10 -5.040134X10'14 r12 -6.846745X10-17 r14 4.655883X10"19 r16 7.854166X10'21
Zone 3:
Effective Focal Length 32.0515mm
Base Radius 26.37444mm
Conic 0.1965319
Aspheric Coefficients r2 2.988976xl0"3 r4 -1.127446X10'5 r6 -5.863206xl0"9 r8 8.905697xl0-13
„10 -1.472282X10"14
—12 -7.391061X10-17
Zone 4:
Effective Focal Length 35.2546mm
Base Radius 27.69962mm Conic 0.1344306 Aspheric Coefficients r2 1.1298197xl0'3 r4 -8.30275X10"6 r6 -4.769432X10"9 r8 -6.74277X10'12
—10 -1.675321X10'15 r12 -1.827999X10'17 r14 -4.208912X10"21
A preferred aperture width is 0.043 inches. Another embodiment of a multiple-focus optical system comprises a poly-optic lens wherein the longer focal length zones have a restricted angular field of view, as may be accomplished by intentionally leaving some of the aberrations of the zone uncorrected, thereby restricting 56 the angular field of view for that zone over which a clear image of the barcode is produced on the CCD detector. The effect of this structure is shown in Fig. 57, which depicts a poly-optic lens 1160 having three zones and positioned in front of a CCD detector 1161. By selectively leaving coma, astigmatism, and lateral chromatic aberrations uncorrected, a poly-optic lens 1160 is provided having three different fields of view 1162, 1163 and 1164, one for each zone. The field of view 1162 for the first zone is wide to allow reading of wide images
(e.g., bar code labels) close in. The field of view 1163 for the second zone is narrower than the first field of view 1162, and the field of view 1164 for the third zone is yet narrower than the second field of view 1163. By progressively narrowing the field of view for the more distant focal regions, the poly-optic lens 1160 of Fig. 57 offers the advantage of minimizing inadvertent reading of bar codes or other images far from the bar code reader, as well as improving the aberration correction in the regions where reading is desired thereby improving resolution.
Figures 42A-42C are alternative embodiments of optical structures using shutters to reduce or eliminate inter-zonal interference. These embodiments use either mechanical or electro-optic shutters so that only one (or several) lenses are active simultaneously.
Figure 42A shows an arrangement of lenses 830 in a circular pattern, forming a poly-optic array. The lenses 830 are positioned in front of a CCD detector 832. A mechanical shutter 834 is positioned in front of the lens array 830. The shutter 834 may be rotated mechanically around a motor shaft (not shown) . The shutter 834 has an opening 835 (or aperture) which revolves in a circular fashion in response to a motive force applied to the motor shaft. In operation, the shutter 834 screens out all of the lenses 830 except one lens 830 (or several lenses 830, if desired) which collects light through the opening 835 in the shutter 834. The shutter 834 is rotated so that the opening 835 unscreens each lens 830 (or group of lenses 830) sequentially in the circular array. In this manner, the lenses 830 are prevented from interfering with one another. Figure 42B is a diagram of an alternate embodiment employing a shutter mechanism. In Fig. 42B, a poly-optic array of lenses 840 are arranged in a linear fashion. The lenses 840 are positioned in front of a CCD detector 842. A rectangular or other appropriately shaped shutter 844 is positioned in front of the array of lenses 840. The shutter 844 has an opening 845 (i.e., aperture) enabling a single lens 840 (or several lenses 840, if desired) to operate at a given time. The shutter 844 is movable along the axis 846 parallel to the array of lenses 840, and may be driven by a solenoid, a linear motor, or a mechanical cam (not shown) . In such a manner, the position of the opening 845 may be shifted from one lens 840 to the next. In operation, the shutter 844 is driven along the axis 846 parallel to the lenses 840 so as to position the opening 845 in front of the desired lens 840 or lenses 840 in sequence. In this manner, the lenses 840 are prevented from interfering with one another.
Figure 42C is a diagram of another embodiment employing a shuttering technique. In Fig. 42C, a poly- optic array of lenses 850 forms a circular cluster pattern as shown, or may form any other suitable pattern matched to a particular LCD shutter configuration. An LCD array
854 is positioned in front of the array of lenses 850. The lenses 850 are positioned in front of a CCD detector 852. The LCD array 854 comprises a plurality of elec¬ tronically operated shutters 855. Each shutter 855 is individually responsive to an electronic shutter command from a shutter controller (not shown) . The LCD shutters
855 are sequenced electronically so that one lens 850 (or several lenses 850, if desired) is operating at given time. In this manner, the lenses 850 do not interfere with one another. An advantage of the LCD array 854 is that it has no moving parts, and therefore may increase reliability.
The above described shuttering techniques may be implemented with a multi-focal plane optical collection system comprising, for example, either a single poly-optic element or a multi-element optical system.
The effect of the shuttering techniques described in Figs. 42A-42C may be explained with reference to Figs. 23B and 23C, which, as noted previously, are graphs plotting minimum renderable bar size against distance from a lens. If all the lenses in the previously described poly-optic arrays of Figs. 42A-42C operate simultaneously, then dead zones 472 such as shown in Fig. 23C may occur in the composite depth of field due to inter-zonal interference, such that a target image may be misidentified or missed entirely by the reading device. The shuttering techniques of Figs. 42A-42C provide for sequential operation of the individual lenses, thereby minimizing or eliminating inter-zonal interference. The depth of field for each lens, as shown in Fig. 23B, is established separately for each individual lens. However, after sequencing through all the desired lenses using the described shuttering techniques, the composite depth of field becomes the aggregate of the individual depths of field, without the effect of inter-zonal interference. The result is a much larger effective depth of field than is possible with only a single lens, or without shuttering.
Another embodiment of a multi-focus optical system in which inter-zonal interference is reduced is shown in Figs. 56A and 56B. The optical system as shown in Fig. 56B comprises a multi-element lens system 1155 and a plurality of photo-sensitive detectors such as linear CCD arrays 1156. The multi-element lens system 1155 comprises three lens elements 1155a-c, each having a different focal length and different field of view. A comparison of the respective fields of views 1151 through 1153 are shown in Fig. 56A. The closest field of view 1151 is associated with the first lens element 1155a; the middle field of view 1152 is associated with the second lens element 1155b; and the most distant field of view 1153 is associ¬ ated with the third lens element 1155c. The lens elements 1155a-c may be embodied as a single poly-optic lens, or as three separate lenses adjacently positioned. While three lens elements 1155a-c are shown, more or fewer lens ele¬ ments may be used as needed depending on the total depth of field requirements. In operation, each lens element 1155a-c focuses light on a different one of the three linear CCD arrays 1156a-c. Thus, the first lens element 1155a has a focal length Fl and focuses light on a first linear CCD array 1156a; the second lens element 1155b has a focal length F2 and focuses light on a second linear CCD array 1156b; and the third lens element 1155c has a focal length F3 and focuses light on a third linear CCD array 1156c. Each linear CCD array 1156a-c thereby receives focused light from only one of the lens elements 1155a-c, and receives an image from only one focal plane. This technique achieves a result of non-interference between lens elements 1155a-c similar to the shuttering techniques described above with respect to Figs. 42A-42C, with the advantage of shorter image reading times since all of the focal planes are imaged simultane- ously or in parallel by the linear CCD arrays 1156a-c.
The poly-optic and shuttering lens techniques described are preferably used in conjunction with the aperturing techniques described elsewhere herein. Some of the various embodiments may be compared in speed and depth of field characteristics as follows. Use of a small, oblong aperture in connection with a single-focus increases speed over a small, round aperture, as the oblong aperture collects more light. Both the oblong aperture and the round aperture result in a relatively good depth of field. Use of a poly-optic lens configura¬ tion with large apertures results in even faster opera¬ tion, but may result in a "perforated" field depth due to intra-lens or inter-lens interference. The perforation in the depth of field may be eliminated by using shutters or multiple CCD arrays (each associated with an independent lens element, as described previously) . Use of a poly- optic lens configuration with oblong apertures results in even faster operation, and also may result in a perforated field depth. Again, shutters or multiple CCD arrays with separate lenses may be used to reduce or eliminate the perforation in the depth of field. A CCD detector used in conjunction with some of the above lens and/or aperture configurations may be used at a small fraction of saturation (e.g., 1% of full scale) in order allow it to run faster (i.e., achieve faster reading speeds) . Running the CCD detector at less than full scale may result in a higher noise system, so correlated double sampling may be used to lower the noise level in the system. Correlated double sampling may also be beneficial where there is no self-generated illumination source being used because the incoming signal level may be reduced. Thus, by gathering enough light and operating in a low noise environment, a CCD detector may use ambient light to read images such as bar code labels.
The optical techniques described herein may be used in whole or part in conjunction with the adjustable-focus techniques described in U.S. Patent Nos. 5,317,211 and
5,479,011, both of which are incorporated by reference as if fully set forth herein.
The optical system can in a variety of embodiments be configured to allow multidirectional imaging using a single detector array. Multidirectional scanning is particularly desirable in fixed reader applications where the relative orientations of bar code and reader may vary widely. One embodiment of an optical system allowing variously oriented images (e.g., bar codes) to be read in shown in Fig. 58. In the Fig. 58 embodiment, two cylindrical optics (such as cylindrical lenses 1172 and 1174) are used to image the bar code 1175. A cylindrical optic (also called a cylindrical rotational filter) produces significant image modulation only in one axis, and therefore has resolution only in the direction perpendicular to the cylinder axis. On the other hand, many types of codes have uniaxial properties in that they contain modulation information only in the direction perpendicular to the bars. In order for a cylindrical optical element to resolve a bar code, the direction of resolution must be adequately aligned with the bar code's direction of modulation. Use of two cylindrical optics at different orientations allows multiple images to be over¬ lapped with less interference, and thereby allows multi¬ directional reading of images in a configuration such as shown in Fig. 58. Accordingly, in Fig. 58, the cylindrical lenses 1172, 1174 are preferably oriented so that their axes are sub¬ stantially perpendicular to one another (i.e., ortho¬ gonal) , so that one cylindrical lens 1174 produces an image well resolved in the horizontal direction while the other cylindrical lens 1172 produces an image well resolved in the vertical direction. The light collected by the horizontal axis cylindrical lens 1174 is rotated approximately 90° by a plurality of image-rotating mirrors 1176 and then combined with the light from the vertical axis cylindrical lens 1172 by a beamsplitter 1178.
The net result is two substantially orthogonal virtual imaging lines in front of the bar code reader, or equivalently, a superposition of two substantially ortho¬ gonal images of the bar code 1175 on the CCD detector 1170. One of these images will be a well resolved image of the bar code 1175, while the other will not be resolved and will result in substantially uniform illumination (i.e., an average light level) upon the CCD detector 1170, and will therefore not interfere with the detection of the well resolved image. In this manner bar codes 1175 (or other images) oriented either horizontally or vertically can be read. Various modifications may be made to the Fig. 58 embodiment to alter its operation in various manners. For example, rather than a linear CCD detector 1170, a two- dimensional CCD array may be used instead. Also, the number of cylindrical optics is not limited to two, but may be any suitable number; likewise, their angle of orientation with respect to one another need not be 90° but may be any suitable angle. The cylindrical optics may be reflective rather than refractive elements, or combina- tions of cylindrical and non-cylindrical optical elements may be used to modify the acceptance angles of the rota¬ tional filters. Further, the focal distances of a multi- axis system may be offset such that different orientation bar codes may focus at different distances away from the reader for the purpose of enhancing depth of field.
A simple one-axis filter is shown in Fig. 59 which may be used singly or multiply to detect when a bar code is aligned in a particular orientation for the purpose of, e.g., controlling other bar code reading equipment which utilizes such information. The one-axis filter of Fig. 59 comprises a vertical axis cylindrical lens 1181 positioned in front of a linear CCD detector 1180. When a bar code 1182 is oriented along the cylindrical axis of the cylin¬ drical lens 1181, the bar code image is at its best resolution. The signal output from the CCD detector 1180 can be compared against a threshold value to determine when a bar code 1182 is properly oriented in front of the CCD detector 1180 so that it may be read.
Another embodiment of an optical system allowing variously oriented barcodes to be read is shown in Figs. 60A-60C, with various alternative area and linear sensor configurations shown in Figs. 60D-60F. Figure 60A depicts an oblique view of the optical system, while Fig. 60B depicts a top view, and Fig. 60C depicts a side view. The embodiment of Figs. 60A-60C comprises a beamsplitter cube 1202 (or alternatively a pellicle beamsplitter or other similar device), a pair of mirrors 1203 and 1204, and imaging regions 1215, 1216 on which are mounted one or more imaging elements (such as shown in any of Figs. 60D- 60F) . The beamsplitter cube 1202 comprises a pair of triangular prism-shaped beamsplitters joined along their diagonal sides, so as to form a cube shape as shown in Fig. 60A. The beamsplitter cube 1202 splits the incoming bar code image into two paths 1207 and 1208. The first path 1208 continues in a vertical line and reaches imaging region 1216. The second path 1207 reaches mirror 1203 which causes a 90° rotation of the image, and then reaches mirror 1204 which causes another 90° rotation of the image so that it also is directed in a vertical line 1209, parallel to the first path 1208. The second vertical path 1209 reaches imaging region 1215. The net effect of the optical structure is shown in Fig. 60B, which shows two versions of the image rotated at 90° with respect to one another, and each respectively being directed down one of the vertical paths 1208 and 1209.
Several different imaging configurations as shown in Figs. 60D-60F may be used in conjunction with the Fig. 60A-60C optical system. Figure 60D depicts two CCD area sensors (i.e., two-dimensional CCD arrays) 1217, 1218 located on imaging regions 1215 and 1216, respectively. The CCD area sensors 1217, 1218 allow detection of a bar code or other image by any of a number of manners well known in the art.
Figure 60E depicts an alternative arrangement wherein two CCD linear arrays 1219, 1220 are disposed on imaging regions 1215 and 1216, respectively. The two CCD linear arrays 1219, 1220 are likewise used to detect the received image. The CCD linear arrays 1219, 1220 are lined up in parallel because the image projected on the first CCD linear array 1219 will be oriented at 90° with respect to the image projected on the second CCD linear array 1220; placing them in parallel therefore increases the chance that the image will be properly oriented on at least one of the CCD linear arrays 1219, 1220. Figure 60F depicts another alternative arrangement wherein a single CCD linear array 1221 straddles both imaging regions 1215, 1216 along their common diagonal. The Fig. 60F embodiment generally requires a larger CCD linear array 1221 than the CCD linear arrays 1219, 1220 of the Fig. 60E embodiment, but may require less processing or fewer components because only one CCD linear array needs to be read out and processed.
Another multi-directional optical system is shown in Fig. 61A. The optical system 1225 receives light reflected from an image (e.g., a bar code) and rotates the image 90° by using an angle adjustment mirror 1226. The rotated image is then reflected off a mirror basket 1228 which divides the single projected image into several segments, resulting in a multi-segment pattern projected onto a photodetector 1227 (such as one or more linear CCD array(s) or CCD area sensor) . Viewed from another perspective, an imaging line from the photodetector 1227 is reflected off the mirror basket 1228 and divided into several segments, resulting in a star shaped pattern. The star shaped pattern 1229 is rotated 90° by the angle adjustment mirror 1226 and projected into the imaging volume in front of the bar code reader to create multi¬ directional virtual imaging lines across an image to be read.
To obtain the desired pattern, the mirror basket 1228 preferably has a first mirrored region 1230 and a pair of second mirrored regions 1231 attached to either side of the first mirrored region 1230 and bent therefrom by an angle of 30°. The mirror basket 1228 is positioned at a 45° angle with respect to the horizontal in order to rotate the image 90° while at the same time generating the star shaped pattern 1229.
While the Fig. 61A embodiment involves several compo- nents in the collection optics, it offers the advantage that the images associated with the variously oriented virtual imaging lines are projected onto different seg- ments of the photodetector 1227, reducing or eliminating the possibility of interference between different imaging lines.
The use of a mirror basket can be particularly useful in the embodiment of Fig. 6IB, whereby a handheld bar code reader 1235 which has a single virtual imaging line in normal handheld operation can be made to have multiple virtual imaging lines in a fixed configuration. The handheld bar code reader 1235 is placed in a cradle (not shown) adjacent to a fixed optical projection unit 1236. The optical projection unit 1236 comprises an angle adjustment mirror 1238 and a mirror basket 1237 (e.g., such as mirror basket 1228) positioned at a 45° angle for rotating an image (such as from an object held above the fixed optical projection unit 1236) and projecting it (in a star-shaped pattern) onto the imaging sensor within the handheld bar code reader 1235. The optical projection unit 1236 thereby gives the handheld bar code reader 1235 effective multi-directional imaging capability, even if the handheld bar code reader 1235 is only a single-line imaging device.
Another embodiment of a multi-direction reader which allows multi-directional reading employs a kaleido¬ scope mirror for projecting multiple rotated images of a bar code onto a detector array, such as shown in Figs. 62A-62F. As shown in Figs. 62A and 62B, a kaleidoscope mirror 1250 comprises two mirror elements 1253, 1254 connected along one edge and oriented at a 45° angle with respect to one another, with the reflective portions of the mirror elements 1253 and 1254 facing one another. An image 1251 projected into the "tunnel" formed by the kaleidoscope mirror 1250 causes a plurality of images to be generated at a detector 1252 at the opposite end of the kaleidoscope mirror, as shown in Figs. 62B and 62C. A plurality of imaging "zones" 1259 are defined over the region of the detector 1252, each having a version of the original image 1256 oriented at a different direction. The total number (Nz) of duplicate projected images 1257 is generally given by the equation N- = 360°/AK, where Aκ is the angle of the kaleidoscope mirrors 1253, 1254 with respect to one another. In the present example, the angle Aκ is 45°, so the number of projected images is eight.
Where the detector 1252 is embodied as a CCD area sensor, the kaleidoscope mirror embodiment of Fig. 62A-62B has the advantage of allowing relatively simple signal processing of the CCD area sensor output without requiring complex algorithms for determining the image orientation. The data from the CCD area sensor may be read out linearly in a fixed direction (i.e., line-by-line), and there is a high probability that one of the duplicate projected images 1257 will be at a sufficiently horizontal angle with respect to the CCD area array as to allow recognition of the image (such as, e.g., a bar code).
Figures 62D-62F are examples showing some of the various relative orientations that the duplicate projected images 1257 may take. The straight arrow symbol "-->" in Figs. 62D-62F represents the angle of orientation of each duplicate projected image. In Fig. 62D the original image 1263 is oriented in a vertical direction with respect to the CCD area sensor, but a duplicate image 1261 is projected at a horizontal orientation by the kaleidoscope mirror 1250. In Fig. 62E the original image 1264 is oriented at a lesser vertical direction (e.g., about 67°), and another duplicate image 1262 from a different image zone 1259 is directed in a close to horizontal direction so as to allow recognition using the CCD sensor array. Figure 62F shows the "worst case" scenario in which the original image is oriented at 45° with respect to the horizontal direction. In this case, all of the duplicate projected images 1257 are about equally difficult to recognize using the CCD sensor array, and the ability to recognize any of the projected images 1257 will depend in part on the tolerances of the signal processing and recognition electronics and intelligence. For a bar code reader to function properly with ambient illumination, sufficient distance should prefer¬ ably be maintained between the bar code and the optical system to allow ambient light to reach the bar code. If the reader and bar code are brought too close together, the bar code may be shadowed to such an extent by either the reader or the object bearing the bar code that it becomes unreadable in ambient light. Several modifica¬ tions may be made to the bar code reader to increase the amount of ambient light reaching the bar code label. One possible modification as illustrated in Fig. 63 comprises an optically transmissive spacer 1303 extending forward from the housing 1301 which encases the optical system. The optically transmissive spacer 1303 ensures that a minimum adequate distance is maintained between the reader and the bar code, while at the same time allowing ambient light to reach the bar code. The optically transmissive spacer 1303 may be transparent, or may comprise opaque or solid portions with missing gaps (e.g., two opaque or solid walls and two missing walls) . Ambient light 1311 entering through the transparent or opaque material, or the gaps, of the optically transmissive spacer 1303 reflects off an optically scattering surface 1306 (which is a more reflective surface) in front of the housing 1301 and thereby results in ambient light 1312 being reflected away from the housing 1301 and onto the bar code 1310. The ambient light reflected from the bar code 1310 is received into the receiver aperture 1305 and collected by the optical system. The optically transmissive spacer 1303 may optionally be configured with a transparent window 1304 across its front end (i.e., the end opposite the housing 1301). In hand-held readers, the optically transmissive spacer 1303 may take the form of a transparent "nose" on the front end of the reader. Alternatively, an opaque "nose" may be employed in which sections of the nose are missing to allow entrance of ambient light into the volume between the reader and the bar coded object 1309. In either of these cases the window 1304 is not required, but never¬ theless is optional to provide protection. In fixed bar code reader applications, particularly where the bar code reader has horizontally positioned optical front-end and objects to be read are held above it, a transparent window 1304 is preferred so as to keep objects off of the optical front-end. In either case, the optically transmissive spacer 1303 mechanically prevents the bar coded object 1309 from being brought too close to the reader to cut off all ambient light, and the transparent, translucent, or missing portions of the optically transmissive spacer 1303 allow ambient light 1311 to enter the space between the bar coded object 1309 and the optical collection system and illuminate the affixed bar code 1310. The spacer 1303 also limits the required range for depth of focus.
A second structure for increasing ambient light at close range or for shaded bar codes, and which may be used alone or in combination with the optically transmissive spacer 1303, is an optically diffusive front surface placed on the barcode reader (excluding the optical collection system) , such as shown for example in Fig. 64A. The optically diffusive front surface 1320 shown in Fig. 64A may be fabricated from a variety of materials, including, but not limited to, roughened white plastic, roughened clear plastic, or brushed metal. As shown in Fig. 64B, the optically diffusive front surface 1320 scatters a portion of the incident ambient light back in the direction of the bar code 1321, thereby increasing the level of illumination at the bar code 1321. An opening 1325 in the diffuser 1320 allows the reflected ambient light to be collected by a lens 1326 (such as any of the multi-focal or poly-optic lenses herein described) and be focused onto a CCD detector 1327. Figure 64C compares a situation wherein an ambient light source 1330 hits the back of a bar coded object 1331, thereby causing the bar code 1332 to be shaded with respect to the bar code reader 1335 and preventing sufficient ambient light to allow the reader to properly operate.
The optical collection and imaging system can in one aspect be regarded as projecting one or more images of the detector array(s) into the imaged volume of space in front of the barcode reader, creating one or more virtual imag¬ ing lines. In order to read a bar code one of these imaging lines is preferably projected across the bar code. This targeting can become increasingly difficult as the depth of field of the bar code reader is lengthened by, for example, use of the multi-focal imaging techniques described herein. To facilitate proper relative posi¬ tioning of the bar code and the reader, one or more LEDs or similar devices may be used as pointers. A preferred LED pointer system is depicted in Figs. 65A and 65B, which are side view and front view diagrams, respectively.
The Fig. 65A-65B embodiment comprises an LED 1350
(such as a "super bright" style LED) , a housing wall 1352 with an aperture or slit 1351, a holographic (or other type of) diffuser 1353 positioned so as to cover the slit 1351, and a plastic lens 1355 positioned in front of the slit 1351 and in alignment with the LED 1350. In opera¬ tion, light from the LED 1350 is diffused by the holo¬ graphic diffuser 1353 and thereby uniformly illuminates the area of the slit 1351. Light passing through the slit 1351 and the holographic diffuser 1353 thereafter passes through the plastic lens 1355 which serves as a pointer optic or optical system to image the light from the illuminated slit 1351 at a point which coincides in space with the projected imaging line. Without the holographic diffuser 1353, which provides a diffused and relatively uniform outgoing illumination from the slit 1351, the LED output may otherwise appear nonuniform and may generate an image of the LED chip (including the bond wires) used to support the LED 1350.
Preferably, two pointer LEDs such as LED 1350 are employed, one on either side of the imager (as per Fig. 46A or 48) , or alternatively positioned above or below the imager. The pointer LEDs are not used to illuminate the bar code for reading, but merely to assist the bar code reader operator in properly positioning the bar coded object with respect to the reader. The proper positioning can be recognized by the appearance of a more focused image of the pointer slit 1351 near the bar code. Figure 65C is a diagram illustrating various pointer slit shapes at different distances. As shown in Fig. 65C, the illumi- nated pointer slits appear as relatively small, spaced apart dots 1360 at near field, as larger oblong more closely spaced shapes 1361 at mid field, and as a single larger oblong shape 1362 at far field. The pointer LED(s) 1350 are preferably cycled on and off corresponding to the image capture cycle, being off while an image of the bar code is being captured by the imager so the pointer beams do not interfere with the image detection. Alternatively, an optical notch filter (not shown) may be used to prevent backscattered light from the LED pointer from reaching the CCD detector array.
Various modifications or alternative structures may be employed for the LED pointing system described above. For example, one alternative embodiment employs at least two pointers focused at different points in space. Another alternative embodiment employs a compound para¬ bolic concentrator optical assembly (not shown) instead of a diffuser 1353. Several other alternative embodiments are particularly directed to increasing the efficiency of coupling of the LED light through the pointer optical system. One such embodiment comprises a cylindrical lens (not shown) positioned in front of the LED 1350 but behind the diffuser 1353 and a slit shaped pointer aperture 1351, the cylindrical lens serving to direct a larger fraction of the light emitted by the LED 1350 onto the pointer aperture 1351 (thereby reducing the LED power needed) . Another embodiment wherein a similar objective is achieved comprises a light pipe (not shown) covering the LED 1350 and shaping the light to match the shape of the pointer aperture 1351. The light pipe may have a diffusing end so that it could also replace the diffuser 1353.
The techniques described herein may be used in conjunction with area imaging techniques such as those described, for example, in U.S. Patent No. 5,446,271 entitled "Omni-directional Scanning Method and Apparatus, " which is hereby incorporated by reference as if fully set forth herein. Thus, an area imager may be provided having increased speed in various lighting conditions, or being capable of operation in ambient light without a self- generated illumination source. Also, the multi¬ directional reading techniques using a multi-dimensional CCD area sensor and virtual imaging lines can be employed in the embodiments described herein using CCD area sensors.
After a target image is processed through the optical front-end and captured by the CCD detector such that a CCD video output signal is produced, the CCD signal is pro- cessed in a variety of manners. The CCD output- signal itself may be used as part of an auto-exposure or adaptive exposure system for adjusting the exposure time of the CCD detector, as further described herein. The CCD output signal may also be processed to determine the locations of bars and spaces and recognize bar codes or other symbols, as also discussed further herein.
Figure 43A depicts one embodiment of an adaptive exposure system in accordance with one or more aspects of the present invention. Figure 43B is a diagram of exemp- lary waveforms in accordance with the Fig. 43A embodiment.
In Fig. 43A, a pulse train oscillator 900 is shown comprising a Schmitt trigger U7A having an input node 901 connected through a capacitor C9 to ground 902. The input node 901 of the Schmitt trigger U7A is connected through a diode D3 and a resistor R26 positioned in series to the output node 903 of the Schmitt trigger U7A. The input node 901 is also connected to another diode D2 configured in the opposite direction of diode D3, and thereafter to a resistor R22 and then to the output node 903 of the Schmitt trigger U7A.
The output of the pulse train oscillator 900 at point A comprises a periodic waveform 920 such as shown in Fig. 43B. The periodic waveform 920 may have a period of, for example, 30 milliseconds and a 200 microsecond negative- going pulse 921. Capacitor C9 may have a capacitance of 0.1 μF , resistor R22 may have a resistance of 910 ohms, and resistor R26 may have a resistance of 4.3 kΩ. The output pulse of waveform 920 is generally positive when capacitor C9 is charging through resistor R22, and is negative when the capacitor C9 discharges through resistor R26. The pulse train oscillator 900 is connected to a monostable multivibrator 905. The monostable multi¬ vibrator 905 comprises a capacitor C2 connected to one end of a photodiode Dl as shown. The other end of the photo¬ diode Dl is connected to ground 902. The photodiode Dl is preferably located adjacent to the CCD detector or other¬ wise nearby so as to be able to detect the amount of light being received from the optical system. The capacitor C2 is connected to the input node 906 of a Schmitt trigger U7B. Capacitor C2 may have a capacitance of 0.02 μF. In operation, the pulse train oscillator 900 drives the capacitor C2 and the photodiode Dl. The waveform 924 at point B and the resulting output waveform 926 from the monostable multivibrator 905 at point C are shown in Fig. 43B. The conductance of the photodiode Dl controls the output pulse width 927 of the monostable multivibrator 905, and therefore controls the duty cycle of oscillation at point C. When the photodiode Dl receives a relatively large amount of light, the output waveform 926 at point C will have a relatively narrow negative pulse (e.g, almost 100% duty cycle) . When the photodiode Dl receives a rela¬ tively small amount of light, the output waveform 926 will have a relatively wide negative pulse (e.g., near 0% duty cycle) .
The output of the monostable multivibrator 905 is connected to two other monostable multivibrators 908 and 910. Both of monostable multivibrators 908, 910 have a similar configuration. Monostable multivibrator 908 comprises a capacitor C13 at its input . The capacitor C13 is shunted by a resistor Rll to ground, and is connected through another resistor R13 to a Schmitt trigger U7F. Capacitor C13 may have a capacitance of 0.001 μF, and each of resistors Rll and R13 may have a resistance of 1 kΩ. The other monostable multivibrator 910 is configured in a similar manner with capacitor CIO, resistors R23 and R25, and Schmitt trigger U7D, except that resistor R25 has a larger resistance than resistor Rll, such as 3.3 kΩ.
The output of monostable multivibrator 908 is con¬ nected to another monostable multivibrator 912, configured in a similar manner with capacitor C7, resistors R8 and R9, and Schmitt trigger U7C. The output of monostable multivibrator 912 at point F is read-out gate (ROG) signal 913, the function of which is explained below. Monostable multivibrators 908 and 912 preferably each produce a fixed pulse of about 1 microsecond in duration, while monostable multivibrator 910 preferably produces a fixed pulse of about 3 microseconds in duration.
The shutter signal 915 at point H is a function of the outputs of both monostable multivibrators 905 and 910. The output of multivibrator 905 at point C and the output of multivibrator 910 at point G are diode OR'ed together through diodes D4 and D5, respectively, to a shutter out¬ put pulse stage 914. The shutter output pulse stage 914 comprises a transistor network, including a transistor Ql whose base 916 is coupled to each of diodes D4 and D5 through a resistor R27. Diodes D4 and D5 are connected to the reference voltage 917 (e.g., 5 volts) through a resistor R28 and a capacitor C3 connected in parallel. The collector 918 of transistor Ql, which produces the shutter signal 915 at point H, is connected through resistor R30 to the reference voltage 917. The base 916 of the transistor is connected to ground 902 through resistor R29, while the emitter 919 of the transistor Ql is connected directly to ground 902.
Through the OR'ing operation of diodes D4 and D5, the shutter signal 915 (shown as waveform 938 in Fig. 43B) comprises a series of pulses 939 which transition from a low state to a high state when the negative pulse of monostable multivibrator 905 begins, and stay high until the negative pulse of monostable multivibrator 910 times out. The effect is that of a variable duty cycle oscillator, wherein the duty cycle varies according to the light received by the photodiode Dl. The actual exposure period 940 as shown in Fig. 43B is somewhat less than the duration of the pulses 939 in the shutter signal 938. The exposure period 940 is commenced at the start of each pulse 939 of the shutter signal 938, but is terminated by the ROG signal pulse 935 as shown in waveform 934 of Fig. 43B. The ROG pulse 935 terminates the exposure period 940 by causing a transfer of charge from the pixels in the CCD array into the CCD shift registers. When the shutter signal 938 is high, the pixels collect light. When the shutter signal 938 is low, the pixels are shorted; however, the ROG pulse 935 generally terminates the collection of light before the shutter signal 938 transitions from high to low.
Monostables 908, 910 and 912 ensure that the ROG pulse 935 terminating the exposure period 940 falls cleanly just before the end of the shutter pulse 939 and at a time when clocking to the CCD array is disabled. Monostable 910 establishes a 3 microsecond clock disable period 937, at the end of which the shutter pulse 939 generally terminates. At the same time as the 3 micro- second period 937 begins, monostable 908 establishes a 1 microsecond delay 933, after which the ROG pulse 935 is generated by monostable 912. Thus, the ROG pulse 935 falls at some point in the middle of the 3 microsecond period 937 established by monostable 910, and just prior to the end of the shutter pulse 939. The output of monostable 910 disables clocking to the CCD array at the time when the ROG pulse 934 occurs.
The adaptive exposure circuit of Fig. 43A thus provides an exposure period which varies based on the amount of light. When there is more light, the exposure period 940 is smaller, and when there is less light the exposure period 940 is larger (up to, e.g., 30 milliseconds) .
Figures 44A and 44B show alternative configurations of an optical system employing an adaptive exposure circuit. In Fig. 44A, a CCD chip 950 produces a video output signal 956 which is connected to a pre-amplifier 951. The pre-amplified signal 952 is connected to a peak detector 954, which measures the peak-to-peak value of the pre- amplified video signal 952. The peak-to-peak signal 957 is provided to an exposure control circuit 955 which adjusts the exposure period for the CCD chip 950. The exposure control circuit 955 is configured in a similar manner to the exposure control circuit of Fig. 43A, but uses the peak-to-peak signal 957 to control the exposure time rather than the collected light as measured by a photodiode Dl.
Figure 44B shows an alternative adaptive exposure technique that is similar to the technique of Fig. 44A, but the peak-to-peak signal level of the first derivative of the video signal is used to control the exposure time of the CCD chip instead. In Fig. 44B, a CCD chip 960 produces a video output signal 967 which is connected to a pre-amplifier 961. The pre-amplified signal 966 is connected to a signal processing block 962. The signal processing block 962 generates a first derivative of the pre-amplified signal 966. The first derivative signal 963 is connected to a peak detector 964, which measures the peak-to-peak value of the first derivative signal 963. The peak-to-peak signal 968 is provided to an exposure control circuit 965 which adjusts the exposure period for the CCD chip 960. The exposure control circuit 965 is configured in a similar manner to the exposure control circuit of Fig. 43A, but uses the peak-to-peak signal 968 generated from the first derivative of the video signal to control the exposure time rather than the collected light as measured by a photodiode Dl. Once the video signal is obtained, transitions in the video signal may be detected by any of a number of means as known in the art. For example, transitions in the video signal may be detected by taking a second derivative of the video signal, and finding zero crossings of the second derivative signal during selected timing intervals. One such technique is described, for example, in U.S. Patent No. 4,000,397 entitled "Signal Processor Method and Apparatus" issued in the name of Hebert et al., which patent is hereby incorporated by reference as if fully set forth herein.
A second-derivative detection circuit is shown in Fig. 45A. Corresponding waveforms for the Fig. 45A circuit are shown in Fig. 45B. In Fig. 45A, a low pass reconstruction filter 971 receives a video input signal 972 from a CCD chip 973. The video input signal 972 may comprise a plurality of stepped voltage levels, like stair steps, corresponding to the contents of the CCD array, such as shown in waveform 990 of Fig. 45B. The video input signal 972 may be passed through an optional gain stage 974, before being fed into the low pass filter 971. The low pass reconstruction filter 971 smooths the video input signal 972 into a smoother, more continuous waveform, such as waveform 992 of Fig. 45B.
The low-pass filtered signal 975 is coupled to a differentiator circuit 976, which generates a first derivative 979 of the low-pass filtered signal 975. The first derivative 979 is coupled to another differentiator circuit 978, which takes another derivative, thereby gen¬ erating a second derivative 980 of the low-pass filtered signal 975. The first derivative signal 979 corresponds to waveform 994 of Fig. 45B, while the second derivative signal 980 corresponds to waveform 996 of Fig. 45B.
The first derivative signal 979 is also coupled to a peak detector circuit 981, which detects the peak value of the first derivative signal 979 and generates a threshold signal 977 thereby. The threshold signal 977, the first derivative signal 979, and the second derivative signal 980 are provided to a comparator logic block 982. The comparator logic block 982 detects zero crossings of the second derivative signal 996 (as shown in Fig. 45B) by comparing the second derivative signal with a zero voltage level. The zero crossings are considered valid transi¬ tions if they occur when the first derivative signal 979 exceeds the level (positive or negative) of the threshold signal 977. Thus, the comparator logic block 982 compares the first derivative signal 979 with the threshold levels (positive and negative) 987, 988 defined by threshold signal 977 and determines a valid transition when the second derivative signal 996 crosses the zero level during a time when the first derivative signal exceeds the level of either threshold signal 987, 988. A reconstructed bar signal waveform 998 such as may be obtained by the method described above is shown in Fig. 45B. The polarity of the bar signal waveform 998 depends on the sign of the first derivative signal 994 at the time of the second derivative zero crossing. Further details regarding second derivative detection techniques are described in U.S. Patent No. 4,000,397, cited above. Transitions in the video signal may also be detected by use of techniques described in U.S. Patent No. 5,463,211 entitled "Method and Apparatus for Detecting Transitions in a Time Sampled Input Signal," and/or in copending application Serial No. 08/548,436 filed on October 26, 1995 and bearing the same title, both of which are assigned to the assignee of the present invention and are incorporated by reference as if fully set forth herein. Further related techniques may be found in U.S Patent No. 5,298,728 entitled "Signal Processing Method and Apparatus," which is also incorporated by reference as if fully set forth herein. The delay-line based signal processing technique described in the '211 patent refer¬ enced above may be workable, for example, without the need for an initial low pass reconstruction filter, and may in some circumstances be configured to operate directly on the CCD video output signal.
Derivative-based signal processing performance can be enhanced by employing AC coupled signal processing tech¬ niques as described further below. Because derivatives are noisier than the original signal, derivative-based signal processors sometimes may require a signal to noise ratio on the order of 10 to 1 for optimal operation. Moreover, if the bar code is blurred or is misaligned with respect to the reader, the transitions between low and high signal levels may be gradual ("soft edges") rather than abrupt ("hard edges"), resulting in low amplitude derivative signals despite adequate signal levels.
In order to provide reliable signal processing even when soft edges are present, a preferred embodiment of the signal processing electronics comprises an AC coupled sig¬ nal processing unit based on (1) a high pass FIR (finite impulse response) filter with a relatively low cutoff frequency; and (2) a comparator with hysteresis. A block diagram of such a signal processor is shown in Fig. 67. The Fig. 67 embodiment comprises a high pass FIR filter 1401 and a hysteretic comparator 1402 which includes an operational amplifier 1405 with resistors 1406, 1407 and 1408 having values selected to provide a suitable amount of hysteresis. An input signal 1410 received from a detector array is connected to the FIR filter 1401 that removes the low frequency and DC bias (i.e., non-barcode signals) . The comparator 1402 trips to the BAR (high) state if the output of the FIR filter 1401 is positive, and trips to the SPACE (low) state if the output of the FIR filter 1401 is negative. A hysteresis band is used to keep noise from triggering the comparator 1402. An optimal FIR filter 1401 consists of a Gaussian impulse response high pass filter, whose coefficients for an llth-order filter are tabulated and graphed in Figure 68B. A simplified FIR filter may also be implemented in the form of a moving average filter. Coefficients for an exemplary llth-order moving average high pass FIR filter are tabulated and graphed in Figure 648. The Fig. 68A filter has the advantage of being easily implemented in software.
The operation of the Fig. 67 signal processor depends in part on the speed of the bar code signal (i.e., time between transitions) , and is illustrated in Fig. 69A for bar widths approaching the time constant of the FIR filter 1401 (i.e., fast bar code signals). In this situation, the FIR filter 1401 acts as a DC restoration circuit so as to AC-couple the input signal 1410 (shown as waveform 1421 in Fig. 69A) around zero volts. The amplitude of the coupled signal 1422 (i.e., the output of the FIR filter 1401) is nearly as large as that of the input waveform 1421; therefore, there is no additional noise introduced by the action of the FIR filter 1401. The hysteresis band 1424 of the comparator 1402, comprising positive hyster¬ esis threshold 1425 and negative hysteresis threshold 1426, is set to be just larger than the noise floor. In CCD devices, the noise floor is usually constant and predictable for a constant signal level and can therefore be set appropriately if the expected input signal level is known in advance, or otherwise may be set by experiment. The comparator 1402 trips when the FIR filter output signal 1422 rises above the positive hysteresis threshold 1425. It resets when the FIR filter output signal 1422 drops below the negative hysteresis threshold 1426. The comparator 1402 will no longer switch states when the signal level remains within the hysteresis band 1424. The resulting output signal 1428 is shown in Fig. 69A.
When the bars are much larger (i.e., the input signal is slower) , the waveforms are substantially different, as shown in Figure 69B. The input signal 1431 has wider spacing between transitions. The FIR filter output signal 1432 looks similar to a derivative. However, the ampli¬ tude of the FIR output signal 1432 is still almost the same as that of the input signal 1431. The comparator 1402 trips to a high state when the filter output signal 1432 is above one hysteresis threshold 1435. The derivative-like FIR filter output signal 1432 declines toward zero quite quickly, but the comparator 1402 does not trip to its other state until another edge appears due to the effect of the hysteresis band 1434. When another edge appears, the FIR filter output signal 1432 passes the negative hysteresis threshold 1436, and the comparator 1402 trips to its other state.
Accordingly, the Fig. 67 circuit is capable of effectively reshaping input signal bars from nearly infi¬ nite width to a width on the order of the time constant of the FIR filter 1401. When the optics of the optical system go out of focus or the reader becomes misaligned with respect to the bar code, the rise time of the analog input signal 1410 will generally increase. The Fig. 67 circuit, however, is still able to detect and respond to bar transitions.
The linear phase properties of FIR filters are desirable in the signal processing system. The imple- mentation of FIR filters is relatively straightforward in software. Hardware implementations are more problematic, consisting of transversal filters, LC delay lines, or sample and hold delay lines, but lend themselves to integration onto the same integrated circuit as a APS detector array. As an alternative to an FIR filter, an IIR (infinite impulse response) filter may be used, which is much easier to implement in hardware (with simple RC circuits or active filters) . Figure 66 is a diagram of an exemplary signal processor using a single pole/zero high pass network. The Fig. 66 embodiment comprises a high pass filter 1380, a buffer 1381, and a hysteretic compara- tor 1382. The high pass filter comprises a capacitor 1386 and resistor 1387, the buffer 1381 comprises an opera¬ tional amplifier 1388 in a voltage-follower configuration, and the comparator 1382 comprises an operational amplifier 1391 with resistors 1389, 1390 and 1392 having values selected to provide a suitable amount of hysteresis. A possible drawback of the Fig. 66 embodiment is that it may show increasing amounts of bar distortion as the bar widths approach the filter time constant.
Information regarding the location or timing of the transitions in the video signal is provided to the decoder 107 (such as shown, e.g., in Fig. 1A) , which interprets the transition data by any of a variety of means such as those well known in the art. Thus, the decoder 107 may determine the relative widths of bars and spaces by measuring the relative time between each transition. The decoder may use the bar and space measurement data to determine which of a variety of bar codes have been read, and/or the particular characters of the bar code. Further details regarding decoding may be found, for example, in U.S. patent application Serial No. 07/995,270 entitled "Method and Apparatus for Recognizing and Assembling Optical Code Labels" filed on December 22, 1992, in the name of inventors Craig D. Cherry and Donald D. Dieball, which application is incorporated by reference as if fully set forth herein.
In many barcode reading and decoding systems, the decoding step is the slow step in the procedure. It is therefore desirable to minimize the amount of time spent by the decoder 107 trying to decode invalid signals. This time minimization may be accomplished in a preferred embodiment by performing preprocessing of image data prior to presentation to the decoder 107. By so doing, voltage 82 sequences which are unlikely to correspond to a valid bar code can be rejected without tying up the decoder 107.
Preprocessing may be performed very rapidly by a simple edge counter. A typical bar code will have on the order of 30 edges. A signal sequence may be classified by the edge counter as "interesting" if it exhibits more than a predetermined minimum number of edges, 20 or 25 for example. Those signal sequences deemed "interesting" are then read out more slowly to the decoder 107 for decoding. However, data readout from a CCD detector or other type of photosensitive array is typically destructive--i.e. , the readout destroys the data. The preprocessing edge counter therefore is preferably used in conjunction with a memory storage element such as an analog shift register to enable "rescanning", that is, a second look at the same data, this time by the decoder 107.
A system including rescanning capability is shown in Fig. 72. Rescanning is accomplished through the use of an analog shift register 1702 to store the image data from a CCD detector 1703. After exposure of the CCD detector 1703 to an image, the pixel charges are transferred to from the CCD detector 1703 the analog shift register 1702. A rescan multiplexer 1704 selects the input from the CCD detector 1703. After an image is transferred to the analog shift register 1702, the image data is rapidly read out of the analog shift register 1702 for preprocessing with the "rescan" function enabled. The "rescan" function causes the data from the analog shift register 1702 to be circularly queued, so that after each pixel charge is read at the output end of the analog shift register 1702, it is shuttled back to the other end of the analog shift register 1702 through rescan multiplexer 1704 as the remaining charges are shifted forward by one position.
The image data is provided to a signal processor (e.g., edge detector) 1705, which then provides its infor¬ mation to an edge counter 1706 to perform preprocessing. At this time the decoder 1707 has not yet received the image data, and could be working on decoding of a previous scan. If after preprocessing by the edge counter 1706 the signal qualifies as "interesting", then the image data is read out from the analog shift register 1703 a second time at a slower speed (by altering the speed of the readout clock 1709) commensurate with the data acceptance rate of the decoder 1707, this time with the rescan function disabled. Disabling the rescan function allows the analog shift register 1703 to empty in preparation for the next image, by preventing the image data from being queued in a circular fashion to the opposite end of the register 1703. Simultaneously, a new image may be read into the analog shift register 1703 as the old image is read out. If the original image data did not qualify as "interest- ing" in the first place, then the analog shift register 1703 is emptied in preparation for the next scan, either with a drain or by rapidly reading out with the rescan disabled.
A number of the various rescan components can be combined onto a single integrated chip, as indicated in Fig. 72. The analog shift register 1703 commonly exists in CCD chips in the form of an image transfer register.
The preprocessing and rescan capabilities described in the immediately preceding paragraphs are particularly useful in applications involving a plurality of detector arrays. Each of the arrays could have its own prepro¬ cessor to determine when an "interesting" signal is present, thereby allowing more efficient division of decoder time among only those detector arrays having signals likely to represent a bar code.
Further details of the preferred embodiment of a bar code reader of Fig. 1 are shown in Figs. 70A-70C, with a preferred electronic block diagram and associated wave¬ forms shown in Figs. 71A-71D. The bar code reader 1501 depicted in Fig. 1 and Figs. 70A-70D is a handheld unit with relatively large depth of field that can operate in ambient light by employing a preferred combination of the techniques and components described previously herein.
The scan head 1522 and optics chassis 1506 of the Fig. 1 bar code reader are shown in more detail in Figs. 70A-70C, along with other optical system and related components. Attached to the analog board 1504 and within the optical chassis 1506 is a CCD detector array 1512 of the masked variety such as depicted in Fig. 52A. The CCD detector preferably has rectangular pixels having a 10:1 aspect ratio, such as shown in and described with respect to the example of Fig. 50A. The CCD detector 1512 is connected to an adaptive exposure system such as shown in Fig. 43A and resident on the analog circuit board 1504 along with a microcontroller that generates CCD timing signals, implements an adaptive exposure feedback system, and provides various signals to the decoder (such as described with respect to the electronics in Fig. 71A below) .
Towards the front of the optical chassis 1506 is located an imaging lens 1509, which is shown in more detail in Figs. 73A (front view) and 73B (side view) . In Fig. 73A is shown an imaging lens 1749 (corresponding to imaging lens 1509) that comprises a multi-focus lens portion 1753 such as the Fig. 41A multi-zone lens having four concentric annular zones 1750a-d of varying focal lengths, and which form images on the CCD detector 1512. The multi-focus lens portion 1753 is mounted on a cylinder lens 1755. The cylinder lens 1755 and multi-focus lens portion 1753 preferably comprise a single optical element and may be formed of solid glass or plastic. The cylinder lens 1755 has a concave back surface 1752 which allows correction for field of view aberrations as necessary.
Preferred dimensions (in millimeters) are provided in Figs. 73A and 73B, and are also provided in the table shown in Fig. 73C along with other preferred character¬ istics. Preferably, the tolerance of the center thickness is no more than +0.05 millimeters, the tolerance of the surface irregularity is no more than 10 microns from a perfect figure, and the tolerance of the surface power is no more than 15 microns from a perfect figure. The back surface 1752 minimum clear aperture is 6 millimeters. Returning to the description of Figs. 70C-70C, a high aspect ratio slit aperture 1508 (i.e., a series of indi¬ vidual apertures formed into a single "continuous" aper¬ ture) such as shown in Figs. 22A-22C is located in front of the multi-zone lens 1509. Mounted on or in the optical chassis 1506 are two pointer LEDs 1520, each illuminating a holographic diffuser 1519 and sending LED pointing beams through a pointer slit 1518 and pointer lens 1507. The LED pointing system is similar to that described with respect to Fig. 65A. The optical chassis 1506 is rigidly connected to the analog circuit board 1504 in order to fix the relative positions of the slit aperture 1508, multi- zone lens 1509, CCD detector 1512, and pointer LEDs 1520.
The preferred bar code reader 1501 further comprises a derivative based signal processor for converting a video signal from the CCD detector 1512 to a bar/space signal for the decoder, with the signal processor being imple¬ mented on the analog circuit board 1504 while the decoder resides on the digital circuit board 1503. The bar code reader 1501 also has an optically transmissive spacer such as described with respect to Fig. 63 at the input end of the scan head 1522 with a missing top section for allowing sufficient ambient illumination to reach the bar code.
In operation, light from the pointer LEDs 1520 passes through holographic diffuser 1519 to uniformly illuminate pointer slits 1518 such as described with respect to Fig. 65A. Pointing beams from the pointer slits 1518 are imaged by the pointer lenses 1507 in order to allow an operator to aim the unit at the bar code and to position the bar code within the depth of field of the bar code reader 1501. The purpose of the LEDs, as previously mentioned, is not to illuminate the bar code, which is illuminated with ambient light, but simply to assist the user in properly positioning and aiming the bar code reader 1501. The LEDs 1520 are preferably turned off during the exposure interval to avoid nonuniform illumi¬ nation of the bar code and to conserve power in the unit. Ambient light reflects off a target (e.g., an encoded symbol) located along a virtual imaging line. The spacer at the front of the bar code reader 1501 prevents the unit from being brought so close to the bar code that shadowing prevents adequate ambient illumination from reaching the bar code. The reflected ambient light is focused by the imaging lens 1509 onto the CCD chip 1512. The CCD chip 1512 translates light intensity information into an electrical signal by any of the techniques described previously herein. The electrical signal may be processed and decoded to provide recognition of the encoded symbol.
In hand held embodiments such as the preferred version of bar code reader 1501, it is often important to minimize power consumption while maintaining a high read rate. Power consumption may be reduced by keeping the analog circuit board powered down until needed to read a bar code. The difficulty that arises in doing so is that the CCD pixels and transfer registers located on the analog board 1504 are full of charge upon power up, and must be emptied before an image can be acquired. This emptying is preferably accomplished by pulsing a shutter pin to clear the pixels of the CCD chip 1512, and by clocking out the transfer register several times at a very high clock speed. Once the transfer register is empty, it may be read out at normal speed for image acquisition. Also upon power up of the analog board 1504 it is desirable to stabilize the exposure feedback system rapidly to facilitate fast bar code reading, and an expo¬ nential control loop is implemented for this purpose. The pointer LEDs 1520 can consume considerable power, and preferably do not illuminate the bar code during image acquisition; therefore the LEDs 1520 are preferably turned off during image acquisition, during the powering of other high power components of the reader (beepers, indicator LEDs, etc.), and as rapidly as possible after a successful decode.
A preferred electrical block diagram for the bar code reader 1501 of Fig. 1 and Figs. 70A-70C is shown in Fig. 71A. The electronics system 1601 comprises a CCD control¬ ler 1602 which may be embodied as a MC68HC05 microchip as manufactured by Motorola, Inc. The CCD controller 1602 provides ROG and SHUT clock signals to a CCD chip 1603. An oscillator 1607 also provides a clock signal to the CCD chip 1603. The CCD chip outputs a video signal to a signal processor 1604. The video signal is also provided to a peak detector and compare circuit 1606, which sends an STB signal back to the CCD controller 1602 to indicate when the signal is too big such that clipping might occur. The signal processor 1604 and CCD controller 1602 are both connected to an interface 1608. The signal processor 1604 and CCD controller 1602 collectively output a data valid (DV) signal, start of scan (SOS) signal, something in volume (SIV) signal, and bar/space (BAR) signal to a decoder (not shown) . The decoder returns with a good read (GR) signal when it has recognized a bar code, and illumi¬ nates a good read LED 1610 visible to the operator of the bar code reader 1501. To facilitate the decoding process, the DV signal is generated by the CCD controller 1602 and, at the end of each exposure cycle signals the decoder to temporarily cease any decoding operations and await a new data stream from the signal processor 1604. Upon completion of the data transfer and bar width digitization to the decoder from the signal processor 1604, the DV signal is removed and decoding operations can resume. In this manner very high pixel readout rates are used to transfer the bar/ space data to the decoder in a fraction of each exposure cycle, leaving more time available for typically slower symbology decoding operations. Another control signal utilized is referred to as "something in volume" or SIV signal. The SIV signal is used to prevent multiple reads of the same bar code, and is particularly useful in fixed reader applications. The SIV signal is output from an edge counter (internal to the CCD controller 1602) and switches state whenever it detects a number of edges substantially different (larger or smaller) than the usual number present when no object is in front of the bar code reader 1501. Upon reception of the SIV signal, the decoder begins digitization and decoding operations until a successful decode occurs. After a successful decode the decoder is disabled until the SIV signal is removed and reasserted. The SIV signal is removed only when the edge count returns to its background level, and is only reasserted when a new bar code label is placed in front of the bar code reader 1501. In this way multiple reading of the same label is prevented.
A description of each of the signals associated with the Fig. 71A circuit embodiment as follows. The "mode select" (MODE) signal selects different modes of the device, each of which may, for example, be associated with a different speed requirement.
The "read out gate" (ROG) signal is an active high signal that drives the CCD chip's ROG input to transfer charge from the photodetectors to the transfer register.
The "shutter" (SHUT) signal is an active high signal that performs a shutter function by draining the charge from the photodetectors on the CCD chip 1603. The period of time from the release of this signal (i.e., going low) to the end of the ROG signal going high is the exposure time.
The "LED flash" (LED) signal is an active high signal that turns on the pointer LEDs 1520. The LED signal is preferably not active during the exposure time.
The "data valid" (DV) signal is, as explained previously, an active high signal that is on during the readout of pixel information from the CCD chip 1603. The DV signal being high represents the time when useful image data exists.
The "start of scan" (SOS) signal toggles after each read by the CCD chip 1603. The DV signal is approximately centered with the SOS toggling.
The "something in volume" (SIV) signal is, as explained previously, an active high signal that pulses briefly when there is a transition from the background BAR pattern to something different (such as when a label has just appeared) .
The "bar input" (BAR) signal is an active high signal which has high signal levels corresponding to relatively darker portions of a read image and low signal levels corresponding to relatively lighter portions of a read image.
The "signal too big" (STB) signal is an active high signal that is on when the video output signal from the CCD chip 1602 is larger than the desired value, set by a Vref input on a comparator. The exposure time is preferably reduced whenever this signal is high and increased whenever this signal is low.
The "speed up" (SUP) signal, when made a high impedance input, will cause the pixel clock to increase to a much higher frequency. When pulled low as an output, the pixel clock will resume its normal frequency. The SUP signal is used on power up to drain the CCD chip 1603 of stored charge.
Exemplary signal waveform diagrams associated with the Fig. 71A embodiment are shown in Figs. 71B-71D. Figure 71B shows the relationship between the ROG signal, DV signal, video signal (i.e., CCD chip output signal), and BAR signal, and exemplary timing characteristics. Figure 71C shows an enlarged timescale view of these same signals, along with the LED signal, SHUT signal, SOS signal, SIV signal, and STB signal. Figure 71D shows these same signals at start up timing (such as when the analog board 1504 is energized) .
An embodiment of a particular CCD detection module 1001 is depicted from various angles of view in Figs. 46A- 46D. Figure 46A shows a top view of the CCD detection module 1001, while Fig. 46B shows a front view, Fig. 46C shows a side view, and Fig. 46D shows an oblique view of the CCD detection module 1001. In Figs. 46A-46D, the CCD detection module 1001 comprises a CCD chip 1003, one or more light emitting diodes (LEDs) 1005, an imaging lens 1006, and one or more focusing or pointer lenses 1007. The pointer lenses 1007 provide focusing for the LEDs 1005. The imaging lens 1006 may be a single zone lens, a high aspect ratio aperture lens, a multi-zone lens, or a combination thereof.
The system may comprise one or more circuit boards 1008 on which to mount the various components of the system, and may further comprise a power supply (not shown) , a CCD driver (not shown) , and signal processing electronics (not shown) . The CCD driver and signal processing electronics may comprise embodiments previously described herein.
Alternatively, in other preferred embodiments, the CCD detection module 1001 may be configured to operate without the LEDs 1005 and pointer lenses 1007 for providing aiming beams.
Figures 47 and 48 show two different embodiments of systems incorporating the CCD detection module 1001 of Figs. 46A-46D. Figure 47 is a diagram of an embodiment of a handheld imaging device 1020 such as may be used in point of sale applications or for inventory control. In the handheld imaging device of Fig. 47, the CCD detection module 1001 is located in a head portion 1022 of the hand¬ held imaging device 1020 and is connected to a handheld scanner controller (not shown) . The CCD video output signal from the CCD detection module 1001 may be directly provided to the handheld scanner controller, or may be conditioned and/or processed first by any of the tech¬ niques described previously herein. If pointer LEDs are used, the pointer LEDs 1005 may be controlled by a signal from a trigger 1021 which is routed through the handheld scanner controller.
Figure 48 is a diagram of an embodiment of a portable data terminal 1030 incorporating the CCD detection 1001 module of Figs. 46A-46D. In Fig. 48, the portable data terminal 1030 comprises a keypad 1032 for entering data and a display screen 1031 for displaying data in a manner known in the art. The CCD detection module 1001 is con¬ nected to a decoder (not shown) internal to the portable data terminal 1030. The CCD video output signal from the CCD detection module 1001 may be directly connected to signal conditioning and/or processing electronics in the portable data terminal 1030, or may be conditioned and/or processed within the CCD detection module 1001 by tech¬ niques described previously herein, and thereafter pro¬ vided to the decoder of the portable data terminal 1030. The pointer LEDs 1005 also may be controlled by the portable data terminal 1030.
The CCD detection module 1001 may also be mounted in a fixed fashion (e.g., on a "slim-stand" or support stand) for retail or industrial applications. Alternatively, the CCD detection module 1001 may be placed on a hand or wrist mounted imaging device for applications such as point of sale operations or inventory control.
Figure 49 is a diagram of an embodiment of a multiple-CCD imaging device which may be employed to read a multi-dimensional area. In Fig. 49, an imaging device 1050 comprises a plurality of CCD detectors 1051. Each CCD detector 1051 has a corresponding lens 1052 for collecting light. The lens 1052 may be a poly-optic lens such as described previously herein, and may be apertured or shaped in accordance with techniques also described previously herein. A transparent element 1054 is placed in the top of the imaging device 1050, such that items bearing encoded symbols may be passed over the imaging device 1050. Light is collected by each lens 1052 along one of a plurality of virtual imaging lines 1053 associa¬ ted with the lens 1052. By placing the CCD detectors 1051 at different angles (such as shown in the example of Fig. 49) , a multi-dimensional region may be covered by the imaging device 1050.
The multiple CCD detectors 1051 may be controlled by a single CCD driver configured in a manner as in Fig. 7, for example. Thus, the CCD detectors 1051 may be polled in series by a single signal processing unit. The imaging device 1050 may be incorporated into a fixed scanning system and may be used for point-of-sale applications, such as retail grocery checkout applications. The techniques described herein may be used in combination or, in some instances, separately so as to provide an image reader having increased capacity to operate in various lighting conditions. The described techniques may allow the image reader either to run faster where a self-generated illumination source is used, or to operate in ambient light in the absence of a self- generated illumination source.
It is intended that various alternative embodiments of the present invention may be constructed and operated using various combinations of the devices and methods set forth herein. Additions, omissions, and/or modifications of one or more of these devices and methods may be made without departing from the scope of the invention. For example, the bar code reading methods and apparatus described herein may be constructed with CMOS active pixel sensor arrays in place of CCD arrays without fundamentally changing the inventive concept set forth herein, namely, methods and apparatus for reading a bar code under ambient illumination. CMOS active pixel sensor arrays are described in more detail in, e.g., copending provisional application Serial No. 60/003,256 entitled "Single Chip Bar Code Scanner Using CMOS Active Pixel Sensor Technology," filed on August 25, 1995, and hereby incorporated by reference as if fully set forth herein.
The present invention has been set forth in the form of its preferred embodiments. It is nevertheless intended that modifications to the disclosed optical reading tech¬ niques and devices may be made by those skilled in the art without departing from the scope and spirit of the present invention. Moreover, such modifications are considered to be within the purview of the appended claims.

Claims

Claims
1. A bar code reader comprising a multi-focus ambient light gathering system comprising a lens having a plurality of zones, each zone having a different focal length; means for reducing interference between said zones caused by simultaneous gathering of ambient light; a photo-sensitive detector outputting a signal having an amplitude varying in response to lighter and darker portions of a read image; and a decoder coupled to said signal.
2. A handheld bar code reader comprising a photosensitive detector outputting a signal indicative of a level of received light; a multi-focus lens having a plurality of concentric zones, said multi-focus lens positioned so as to focus ambient light on said photosensitive detector; an aperture positioned along a central axis of said lens, said aperture having a non-unity aspect ratio,- and an edge detector coupled to said signal.
3. The handheld bar code reader of claim 2 wherein said photosensitive detector comprises a- CCD detector having array of rectangular pixels.
4. The handheld bar code reader of claim 2 further comprising a mask positioned in front of said CCD detector.
5. The handheld bar code reader of claim 2 further comprising a housing including a handle portion and a head portion, said head portion comprising an imaging window positioned to allow ambient light to be collected by said multi-focus lens and further comprising an optically transmissive spacer in front of said imaging window.
6. The handheld bar code reader of claim 2 further comprising an adaptive exposure circuit connected to said photosensitive detector.
7. The handheld bar code reader of claim 2 further comprising means connected to said photosensitive detector for reducing common mode noise.
8. The handheld bar code reader of claim 2 further comprising means for enhancing transitions in said signal.
9. The handheld bar code reader of claim 8 wherein said means for enhancing transitions in said signal com¬ prises a finite impulse response filter coupled to a comparator.
10. The handheld bar code reader of claim 8 wherein said means for enhancing transitions in said signal com- prises an infinite impulse response filter coupled to a comparator.
11. The handheld bar code reader of claim 2 further comprising a decoder connected to said edge detector.
12. The handheld bar code reader of claim 2 further comprising at least one light emitting diode for generat¬ ing an aiming beam.
13. The handheld bar code reader of claim 12 further comprising an optical diffuser positioned in the path of said aiming beam.
14. The handheld bar code reader of claim 2 wherein each zone of said multi-focus lens provides a different field of view, each field of view progressively narrowing from the zone with the shortest focal distance to the zone with the longest focal distance.
15. The handheld bar code reader of claim 2 further comprising means for focusing along multiple axes by selectively magnifying said image along each of said axes and combining the magnified images.
16. The handheld bar code reader of claim 2 further comprising a kaleidoscope mirror, wherein said photo¬ sensitive detector comprises an area array for receiving multiple images reflected from said kaleidoscope mirror.
17. The handheld bar code reader of claim 2 further comprising a mirror basket for receiving and reproducing an image in a repeating pattern.
18. A bar code reader comprising a multi-focus ambient light gathering system comprising a lens having a plurality of lens elements, each lens element having a different focal length; a plurality of photosensitive detectors, one for each lens element, each outputting a detection signal having an amplitude varying in response to lighter and darker portions of an image,- and a decoder coupled to said photosensitive detectors.
19. A bar code reader comprising: a light sensitive detector; a multi-focus optical imaging system having multiple focal distances and capable of gathering and focusing ambient light reflected from a bar code label onto said light sensitive detector, wherein said light sensitive detector is capable of generating in the absence of self- generated illumination a signal containing information corresponding to relatively lighter portions and relatively darker portions of the bar code, and a decoder coupled to said light sensitive detector.
20. The bar code reader of claim 19 wherein said multi-focus optical imaging system comprises a poly-optic lens.
21. The bar code reader of claim 20 further comprising an aperture positioned on a side of said poly- optic lens opposite from said light sensitive detector.
22. The bar code reader of claim 20 wherein said poly-optic lens comprises a cut-away portion of a round poly-optic lens.
23. The bar code reader of claim 19 wherein said multi-focus optical imaging system comprises a lens array and a collimating optical element.
24. The bar code reader of claim 23 wherein said lens array is radially disposed in relation to said light sensitive detector.
25. The bar code reader of claim 23 wherein individual lenses of said lens array are disposed parallel to a surface of said light sensitive detector, and wherein said multi-focus optical imaging system further comprises graded index element positioned between said lens array and said light sensitive detector.
26. The bar code reader of claim 25 wherein said lens array, said graded index element and said collimating optical element are combined in a composite lens.
27. The bar code reader of claim 23 wherein said multi-focus optical imaging system further comprises an array of staggered optical flats positioned between said lens array and said light sensitive device.
28. The bar code reader of claim 27 wherein said lens array, said array of staggered optical flats and said collimating optical element are combined in a composite lens.
29. The bar code reader of claim 19 wherein said multi-focus optical imaging system comprises a plurality of curved reflective surfaces each having a different curvature corresponding to a different focal length, whereby light reflected from said bar code label is directed by said curved reflective surfaces onto said light sensitive detector.
30. The bar code reader of claim 19 wherein said multi-focus optical imaging system comprises a plurality of lenses having different focal lengths and means for sequentially shuttering said lenses.
31. The bar code reader of claim 19 wherein said light sensitive detector comprises a charge coupled device.
32. The bar code reader of claim 31 further comprising a correlated double sampling circuit connected to said charge coupled device.
33. The bar code reader of claim 32 wherein said correlated double sampling circuit comprises a dual slope processor.
34. The bar code reader of claim 32 wherein said correlated double sampling circuit comprises a dual sampler processor.
35. The bar code reader of claim 32 further comprising means for varying an exposure time of said charge coupled device as a function of an amount of light received by said charge coupled device.
36. The bar code reader of claim 19 wherein said multi-focus optical imaging system has a depth of field of at least 48.3 millimeters to 191 millimeters.
37. A bar code reader comprising: a light sensitive detector; a multi-focus optical imaging system having multiple focal distances for gathering and focusing onto said light sensitive detector ambient light reflected from a bar code label without use of a self-generated illumination source, said light sensitive detector generating a signal contain¬ ing information corresponding to relatively lighter por¬ tions and relatively darker portions of the bar code, and a decoder coupled to said light sensitive detector.
38. The bar code reader of claim 37 wherein said multi-focus optical imaging system comprises a poly-optic lens positioned between an aperture and said light sensitive detector.
39. The bar code reader of claim 37 wherein said light sensitive detector comprises an array of photo¬ diodes, and further comprising means for varying an exposure time of said photodiodes in response to the strength of received light.
40. The bar code reader of claim 37 further comprising means for reducing common mode noise in said signal generated by said light sensitive detector.
41. A bar code reader comprising: a multi-focus optical imaging system having multiple focal distances and capable of gathering and focusing light reflected from a bar code label; a light sensitive detector positioned to receive light focused by said multi-focus optical imaging system, said light sensitive detector generating from ambient light in the absence of self-generated illumination a signal containing information corresponding to relatively lighter portions and relatively darker portions of the bar code; and a decoder coupled to said light sensitive detector.
42. The bar code reader of claim 41 wherein said multi-focus optical imaging system comprises a poly-optic lens positioned between an aperture and said light sensitive detector.
43. The bar code reader of claim 41 wherein said light sensitive detector comprises an array of photo- diodes, and further comprising means for varying an exposure time of said photodiodes in response to the strength of received light.
44. The bar code reader of claim 41 further comprising means for reducing common mode noise in said signal generated by said light sensitive detector.
45. A system for reading encoded images comprising a multi-focus imaging system, a charge-coupled device for receiving light collected by said multi-focus imaging system, and generating a CCD output signal thereby, means responsive to said CCD output signal for varying an amount of time said charge-coupled device receives said collected light, means for reducing common mode noise in said CCD output signal, and a decoder coupled to said means for reducing noise.
46. The system of claim 45 wherein said collected light is reflected from a bar code label.
47. The system of claim 45 wherein said means for reducing noise comprises a correlated double sampling circuit.
48. The system of claim 47 wherein said correlated double sampling circuit comprises a dual sampler processor.
49. The system of claim 47 wherein said correlated double sampling circuit comprises a dual slope processor.
50. The system of claim 45 wherein said multi-focus imaging system comprises a poly-optic lens.
51. The system of claim 50 wherein said poly-optic lens is oblong.
52. The system of claim 51 wherein said poly-optic lens is rectangular.
53. The system of claim 45 wherein said multi-focus imaging system comprises an aperture having a high aspect ratio.
54. The system of claim 53 wherein said aperture is oblong.
55. The system of claim 54 wherein said aperture is rectangular.
56. The system of claim 45 wherein said multi-focus imaging system comprises a plurality of lenses, at least two of said lenses having different focal lengths, and 102 means for shuttering at least one of said lenses when collecting light.
57. The system of claim 45 further comprising an edge detector coupled between said means for reducing noise and said decoder.
58. The system of claim 57 wherein said edge detector generates a second derivative of an output of said means for reducing noise and detects zero crossings of said second derivative.
59. A method for reading images comprising the steps of collecting light along a plurality of collimated scan lines, focusing said collimated scan lines onto an array of photodiodes, and generating a signal thereby, removing common mode noise from said signal, detecting edge transitions in said signal, said edge transitions corresponding to transitions between lighter and darker portions of a read image.
60. The method of claim 59 wherein each of said collimated scan lines is associated with a lens element, and at least two of said lens elements have different focal lengths.
61. The method of claim 60 wherein said lens elements collectively comprise an array of lenses.
62. The method of claim 59 further comprising the step of adjusting an exposure time of said array of photodiodes in response to said signal.
63. The method of claim 59 further comprising the step of decoding.
64. The method of claim 59 wherein the step of removing common mode noise comprises charging a capacitor to a reset voltage for a predetermined period, after said predetermined period, applying said signal to said capacitor, and sampling an output of said capacitor.
65. An apparatus for reading images comprising a charge-coupled device for receiving collected light and generating a CCD output signal thereby, and means responsive to said CCD output signal for varying an amount of time said charge-coupled device receives said collected light.
66. The apparatus of claim 65 wherein said varying means comprises a peak detector coupled to said CCD output signal, and said amount of time is varied in response to a measured peak value in said CCD output signal.
67. The apparatus of claim 65 wherein said varying means comprises a differentiator coupled to said CCD output signal, said differentiator generating a first derivative of said CCD output signal, and a peak detector coupled to said first derivative, and said amount of time is varied in response to a measured peak value in said first derivative.
68. An apparatus for reading encoded images comprising a multi-focus imaging system, a charge-coupled device for receiving light collected by said multi-focus imaging system, and generating a CCD output signal thereby, a correlated double sampling circuit connected to said CCD output signal, and a decoder.
69. The apparatus of claim 68 wherein said multi- focus imaging system comprises a poly-optic lens.
70. The apparatus of claim 68 wherein said multi- focus imaging system comprises an array of lenses, at least two of said lenses having a different focal length.
71. The apparatus of claim 70 wherein each of said lenses is associated with a scan line, and wherein said multi-focus imaging system further comprises a collimating lens for collimating said scan lines.
72. An apparatus for reading encoded images comprising a poly-optic lens having a first optical portion for generating collimated scan lines, and a second optical portion for providing different focal lengths along at least two of said collimated scan lines, and a charge-coupled device for receiving light collected by said poly-optic lens and generating an output signal thereby.
73. The apparatus of claim 72 further comprising an oblong aperture positioned in front of said poly-optic lens.
74. The apparatus of claim 71 wherein said poly- optic lens is oblong.
75. A multiple-CCD system for reading encoded images comprising a plurality of charge-coupled devices, each of said charge-coupled devices having a CCD output signal, a plurality of poly-optic lenses, one for each of said charge-coupled devices, for focusing light on said charge-coupled devices, a multiplexer connected to each of said CCD output signals, a signal processor connected to said multiplexer, and a decoder connected to said signal processor.
76. The multiple-CCD system of claim 75 wherein said signal processor comprises a correlated double sampling circuit.
77. The multiple-CCD system of claim 75 further comprising means for varying an amount of exposure time of each of said charge-coupled devices.
78. The multiple-CCD system of claim 75 further comprising an oblong aperture positioned in front of one or more of said poly-optic lenses.
79. The multiple-CCD system of claim 75 wherein at least one of said poly-optic lenses is oblong.
80. The multiple-CCD system of claim 75 wherein at least one of said poly-optic lenses is shuttered periodically.
81. The multiple-CCD system of claim 75 wherein said charge-coupled devices are positioned so as to receive light over a two-dimensional region.
PCT/US1995/016828 1994-12-23 1995-12-22 Method and apparatus for reading images without need for self-generated illumination source WO1996020454A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
JP8520561A JPH09512372A (en) 1994-12-23 1995-12-22 Method and apparatus for reading an image without the need for a self-generated illumination source
DE19581524T DE19581524B4 (en) 1994-12-23 1995-12-22 Method and system for reading images without requiring their own illumination source and bar code reader therefor
GB9617173A GB2301691B (en) 1994-12-23 1995-12-22 Method and apparatus for reading images without need for self-generated illumination source

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US08/363,258 US5770847A (en) 1994-12-23 1994-12-23 Bar code reader with multi-focus lens
US08/363,258 1994-12-23
US08/576,203 US5814803A (en) 1994-12-23 1995-12-21 Image reader with multi-focus lens
US08/576,203 1995-12-21

Publications (1)

Publication Number Publication Date
WO1996020454A1 true WO1996020454A1 (en) 1996-07-04

Family

ID=27001984

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1995/016828 WO1996020454A1 (en) 1994-12-23 1995-12-22 Method and apparatus for reading images without need for self-generated illumination source

Country Status (5)

Country Link
US (3) US5814803A (en)
JP (1) JPH09512372A (en)
DE (1) DE19581524B4 (en)
GB (1) GB2301691B (en)
WO (1) WO1996020454A1 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1998047377A1 (en) * 1997-04-21 1998-10-29 Intermec Scanner Technology Center Optoelectronic device for image acquisition, in particular of code bars
US6732930B2 (en) 1999-12-23 2004-05-11 Intermec Ip Corp. Optoelectronic device and process for acquiring symbols, such as bar codes, using a two-dimensional sensor
DE102008009404A1 (en) * 2008-02-15 2009-09-03 MARSON TECHNOLOGY CO., LTD., Hsin-Tien Polydirectional radiating barcode reader for reading linear bar codes or two-dimensional bar codes, has housing which is provided at side with through-hole
US7770799B2 (en) 2005-06-03 2010-08-10 Hand Held Products, Inc. Optical reader having reduced specular reflection read failures
US7780089B2 (en) 2005-06-03 2010-08-24 Hand Held Products, Inc. Digital picture taking optical reader having hybrid monochrome and color image sensor array
US8720781B2 (en) 2005-03-11 2014-05-13 Hand Held Products, Inc. Image reader having image sensor array
US8978985B2 (en) 2005-03-11 2015-03-17 Hand Held Products, Inc. Image reader comprising CMOS based image sensor array
CN105684437B (en) * 2013-11-03 2021-10-29 杜比实验室特许公司 System and method for local dimming in multi-modulation displays

Families Citing this family (237)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8352400B2 (en) 1991-12-23 2013-01-08 Hoffberg Steven M Adaptive pattern recognition based controller apparatus and method and human-factored interface therefore
US20030043463A1 (en) * 1992-03-30 2003-03-06 Yajun Li Athermalized plastic lens
US7387253B1 (en) 1996-09-03 2008-06-17 Hand Held Products, Inc. Optical reader system comprising local host processor and optical reader
US6708883B2 (en) 1994-06-30 2004-03-23 Symbol Technologies, Inc. Apparatus and method for reading indicia using charge coupled device and scanning laser beam technology
US5672858A (en) 1994-06-30 1997-09-30 Symbol Technologies Inc. Apparatus and method for reading indicia using charge coupled device and scanning laser beam technology
US6073846A (en) * 1994-08-17 2000-06-13 Metrologic Instruments, Inc. Holographic laser scanning system and process and apparatus and method
US6758402B1 (en) 1994-08-17 2004-07-06 Metrologic Instruments, Inc. Bioptical holographic laser scanning system
US7051922B2 (en) 1994-08-17 2006-05-30 Metrologic Instruments, Inc. Compact bioptical laser scanning system
GB2308267B (en) 1995-08-25 2000-06-28 Psc Inc Optical reader with imaging array having reduced pattern density
US6705526B1 (en) * 1995-12-18 2004-03-16 Metrologic Instruments, Inc. Automated method of and system for dimensioning objects transported through a work environment using contour tracing, vertice detection, corner point detection, and corner point reduction methods on two-dimensional range data maps captured by an amplitude modulated laser scanning beam
US6260763B1 (en) * 1996-02-06 2001-07-17 Psc Scanning, Inc. Integral illumination source/collection lens assembly for data reading system
US6141048A (en) * 1996-08-19 2000-10-31 Eastman Kodak Company Compact image capture device
US6173894B1 (en) 1996-08-23 2001-01-16 Psc Inc. Optical reader with addressable pixels
US6166375A (en) 1996-10-08 2000-12-26 Psc Scanning, Inc. Offset optical axes for bar code scanner
US6073849A (en) * 1996-11-01 2000-06-13 Psc Scanning, Inc. Electronic edge detection system using a second derivative signal processor
US6310730B1 (en) 1997-10-02 2001-10-30 Raytheon Company Optical system with asymmetric optical corrector
US6028712A (en) * 1997-10-02 2000-02-22 Raytheon Company Optical system and method for providing corrected optical images
US6313951B1 (en) 1997-10-02 2001-11-06 Raytheon Company Optical system with zernike-shaped corrector
US6318635B1 (en) 1997-12-02 2001-11-20 Telxon Corporation Multi-focal length imaging based portable dataform reader
US6053408A (en) 1997-12-02 2000-04-25 Telxon Corporation Multi-focal length imaging based portable dataform reader
US6575964B1 (en) 1998-02-03 2003-06-10 Sciton, Inc. Selective aperture for laser delivery system for providing incision, tissue ablation and coagulation
US6229964B1 (en) * 1998-02-26 2001-05-08 Eastman Kodak Company Image with sound playback apparatus
US7369327B1 (en) * 1998-06-09 2008-05-06 Olympus Corporation Variable optical-property element and optical apparatus including the same
US6340114B1 (en) * 1998-06-12 2002-01-22 Symbol Technologies, Inc. Imaging engine and method for code readers
US6275388B1 (en) * 1998-07-08 2001-08-14 Welch Allyn Data Collection, Inc. Image sensor mounting system
US20030019934A1 (en) * 1998-07-08 2003-01-30 Hand Held Products, Inc. Optical reader aiming assembly comprising aperture
US20040000592A1 (en) * 2002-02-20 2004-01-01 Welch Allyn, Inc. Adjustable illumination system for a barcode scanner
US6659350B2 (en) 2000-11-01 2003-12-09 Hand Held Products Adjustable illumination system for a barcode scanner
US6607128B1 (en) 1998-07-08 2003-08-19 Welch Allyn Data Collection Inc. Optical assembly for barcode scanner
US6164544A (en) 1998-07-08 2000-12-26 Welch Allyn Data Collection, Inc. Adjustable illumination system for a barcode scanner
US6547139B1 (en) * 1998-07-10 2003-04-15 Welch Allyn Data Collection, Inc. Method and apparatus for extending operating range of bar code scanner
US6133862A (en) * 1998-07-31 2000-10-17 Intel Corporation Method and apparatus to reduce row reset noise in photodiode
US6634558B1 (en) 1998-08-12 2003-10-21 Symbol Technologies, Inc. Optical code reader with hand mounted imager
US7440993B1 (en) 1998-09-11 2008-10-21 Lv Partners, L.P. Method and apparatus for launching a web browser in response to scanning of product information
US6860424B1 (en) * 1998-09-11 2005-03-01 L.V. Partners, L.P. Optical reader and use
US7392945B1 (en) 1998-09-11 2008-07-01 Lv Partners, L.P. Portable scanner for enabling automatic commerce transactions
US6745234B1 (en) 1998-09-11 2004-06-01 Digital:Convergence Corporation Method and apparatus for accessing a remote location by scanning an optical code
US7386600B1 (en) 1998-09-11 2008-06-10 Lv Partners, L.P. Launching a web site using a personal device
US6704864B1 (en) 1999-08-19 2004-03-09 L.V. Partners, L.P. Automatic configuration of equipment software
US7191247B1 (en) 1998-09-11 2007-03-13 Lv Partners, Lp Method for connecting a wireless device to a remote location on a network
US6636896B1 (en) 1998-09-11 2003-10-21 Lv Partners, L.P. Method and apparatus for utilizing an audibly coded signal to conduct commerce over the internet
US6868433B1 (en) 1998-09-11 2005-03-15 L.V. Partners, L.P. Input device having positional and scanning capabilities
US7379901B1 (en) 1998-09-11 2008-05-27 Lv Partners, L.P. Accessing a vendor web site using personal account information retrieved from a credit card company web site
US6823388B1 (en) * 1998-09-11 2004-11-23 L.V. Parners, L.P. Method and apparatus for accessing a remote location with an optical reader having a programmable memory system
US7206849B1 (en) 1998-10-05 2007-04-17 Symbol Technologies, Inc. Communication in a wireless communications network when a mobile computer terminal may be unreachable
US6336587B1 (en) * 1998-10-19 2002-01-08 Symbol Technologies, Inc. Optical code reader for producing video displays and measuring physical parameters of objects
DE69803459T2 (en) * 1998-10-30 2003-02-13 Datalogic Spa Optical device and method for sighting and visually displaying a readout area
DE19852464C2 (en) * 1998-11-13 2002-10-24 Wincor Nixdorf Gmbh & Co Kg Device for reading a bar code
US7904187B2 (en) 1999-02-01 2011-03-08 Hoffberg Steven M Internet appliance system and method
US6830190B2 (en) * 1999-02-02 2004-12-14 Metrologic Instruments, Inc. Multipath scan data signal processor having multiple signal processing paths with different operational characteristics to enable processing of signals having increased dynamic range
US6651888B1 (en) * 1999-02-02 2003-11-25 Symbol Technologies, Inc. Beam shaping system and diverging laser beam for scanning optical code
US6345129B1 (en) * 1999-02-03 2002-02-05 Oren Aharon Wide-field scanning tv
EP1028344A3 (en) * 1999-02-11 2002-02-27 UMM Electronics, Inc. A device for one-dimensional lens imaging
IL133243A0 (en) * 1999-03-30 2001-03-19 Univ Ramot A method and system for super resolution
US6160250A (en) * 1999-03-31 2000-12-12 Hewlett-Packard Company Integrated optical imaging assembly
US6656373B1 (en) * 1999-07-09 2003-12-02 Wavefront Sciences, Inc. Apodized micro-lenses for Hartmann wavefront sensing and method for fabricating desired profiles
CN1391680A (en) * 1999-10-04 2003-01-15 韦尔奇·埃林数据采集公司 Imaging module for optical reader
US7270274B2 (en) 1999-10-04 2007-09-18 Hand Held Products, Inc. Imaging module comprising support post for optical reader
US6832725B2 (en) 1999-10-04 2004-12-21 Hand Held Products, Inc. Optical reader comprising multiple color illumination
US6375075B1 (en) * 1999-10-18 2002-04-23 Intermec Ip Corp. Method and apparatus for reading machine-readable symbols including color symbol elements
JP2001209751A (en) * 2000-01-28 2001-08-03 Olympus Optical Co Ltd Adaptive binarization system
US6520415B1 (en) * 2000-02-10 2003-02-18 Psc Scanning, Inc. Barcode reading device with photodetector assembly including a lenticular array
US7137555B2 (en) 2000-02-28 2006-11-21 Psc Scanning, Inc. Multi-format bar code reader
US6912076B2 (en) * 2000-03-17 2005-06-28 Accu-Sort Systems, Inc. Coplanar camera scanning system
JP4330762B2 (en) 2000-04-21 2009-09-16 富士フイルム株式会社 Multi-beam exposure system
US6502751B1 (en) * 2000-04-26 2003-01-07 Ncr Corporation Methods and apparatus for dual thresholding in processing of barcode signals
US6382511B1 (en) * 2000-04-26 2002-05-07 Ncr Corporation Methods and apparatus for digitizing and processing of analog barcode signals
US6689998B1 (en) 2000-07-05 2004-02-10 Psc Scanning, Inc. Apparatus for optical distancing autofocus and imaging and method of using the same
EP1205873B1 (en) * 2000-11-14 2006-09-27 Datalogic S.P.A. Method for locating the transitions between the elements of a bar code
US6909554B2 (en) * 2000-12-27 2005-06-21 Finisar Corporation Wafer integration of micro-optics
US6556349B2 (en) 2000-12-27 2003-04-29 Honeywell International Inc. Variable focal length micro lens array field curvature corrector
ATE335246T1 (en) 2001-01-22 2006-08-15 Hand Held Prod Inc OPTICAL READER WITH PARTIAL IMAGE CUTTING FUNCTION
US7268924B2 (en) 2001-01-22 2007-09-11 Hand Held Products, Inc. Optical reader having reduced parameter determination delay
US6478224B1 (en) 2001-03-02 2002-11-12 Psc Scanning, Inc. Symbology-independent method and apparatus for identifying potential bar code data
US6743221B1 (en) 2001-03-13 2004-06-01 James L. Hobart Laser system and method for treatment of biological tissues
US6685092B2 (en) * 2001-06-15 2004-02-03 Symbol Technologies, Inc. Molded imager optical package and miniaturized linear sensor-based code reading engines
US6621063B2 (en) * 2001-06-21 2003-09-16 Psc Scanning, Inc. Omni-directional optical code reader using scheimpflug optics
US6770069B1 (en) 2001-06-22 2004-08-03 Sciton, Inc. Laser applicator
US7331523B2 (en) 2001-07-13 2008-02-19 Hand Held Products, Inc. Adaptive optical image reader
US6758403B1 (en) 2001-09-11 2004-07-06 Psc Scanning, Inc. System for editing data collection device message data
US6837431B2 (en) * 2002-04-09 2005-01-04 Symbol Technologies, Inc. Semiconductor device adapted for imaging bar code symbols
US6811085B2 (en) 2001-10-26 2004-11-02 Symbol Technologies, Inc. Miniature imager
TWI227447B (en) * 2001-10-26 2005-02-01 Symbol Technologies Inc Semiconductor device adapted for imaging bar code symbols
GB2418512B (en) * 2001-10-26 2006-07-05 Symbol Technologies Inc Semiconductor device adapted for imaging bar code symbols
US6789737B2 (en) * 2001-12-19 2004-09-14 Riotec Corp. Light source mechanism of barcode scanner
US7344082B2 (en) * 2002-01-02 2008-03-18 Metrologic Instruments, Inc. Automated method of and system for dimensioning objects over a conveyor belt structure by applying contouring tracing, vertice detection, corner point detection, and corner point reduction methods to two-dimensional range data maps of the space above the conveyor belt captured by an amplitude modulated laser scanning beam
US8717456B2 (en) 2002-02-27 2014-05-06 Omnivision Technologies, Inc. Optical imaging systems and methods utilizing nonlinear and/or spatially varying image processing
US6824059B2 (en) * 2002-04-30 2004-11-30 Hewlett-Packard Development Company, L.P. Apparatus for capturing images and barcodes
US7083609B2 (en) * 2002-06-13 2006-08-01 Visx, Incorporated Corneal topography-based target warping
CA2390781C (en) * 2002-06-14 2009-09-22 Institut National D'optique Line generator optical apparatus
US7014113B1 (en) 2002-08-19 2006-03-21 The Code Corporation Versatile graphical code reader for reading different types of graphical codes
TWI222325B (en) * 2002-09-24 2004-10-11 Mstar Semiconductor Inc Interface circuitry for display chip
TW582164B (en) * 2002-10-16 2004-04-01 Veutron Corp Method for optimizing arrangement of light source array
EP1473657A1 (en) * 2003-04-29 2004-11-03 Sicpa Holding S.A. Method and device for the authentication of documents and goods
US7105793B2 (en) * 2003-07-02 2006-09-12 Micron Technology, Inc. CMOS pixels for ALC and CDS and methods of forming the same
US9070031B2 (en) 2003-10-24 2015-06-30 Cognex Technology And Investment Llc Integrated illumination assembly for symbology reader
US7823789B2 (en) 2004-12-21 2010-11-02 Cognex Technology And Investment Corporation Low profile illumination for direct part mark readers
US9536124B1 (en) 2003-10-24 2017-01-03 Cognex Corporation Integrated illumination assembly for symbology reader
US7604174B2 (en) 2003-10-24 2009-10-20 Cognex Technology And Investment Corporation Method and apparatus for providing omnidirectional lighting in a scanning device
US7823783B2 (en) 2003-10-24 2010-11-02 Cognex Technology And Investment Corporation Light pipe illumination system and method
US7874487B2 (en) 2005-10-24 2011-01-25 Cognex Technology And Investment Corporation Integrated illumination assembly for symbology reader
US7017812B1 (en) 2003-11-26 2006-03-28 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Variable distance angular symbology reader
US6942152B1 (en) 2004-01-21 2005-09-13 The Code Corporation Versatile graphical code reader that is configured for efficient decoding
US7707039B2 (en) 2004-02-15 2010-04-27 Exbiblio B.V. Automatic modification of web pages
US8442331B2 (en) 2004-02-15 2013-05-14 Google Inc. Capturing text from rendered documents using supplemental information
US7812860B2 (en) 2004-04-01 2010-10-12 Exbiblio B.V. Handheld device for capturing text from both a document printed on paper and a document displayed on a dynamic display device
US10635723B2 (en) 2004-02-15 2020-04-28 Google Llc Search engines and systems with handheld document data capture devices
US7481536B2 (en) * 2004-02-19 2009-01-27 Amo Manufacturing Usa, Llc Methods and systems for differentiating left and right eye images
US7990556B2 (en) 2004-12-03 2011-08-02 Google Inc. Association of a portable scanner with input/output and storage devices
US20060081714A1 (en) 2004-08-23 2006-04-20 King Martin T Portable scanning device
US20060098900A1 (en) 2004-09-27 2006-05-11 King Martin T Secure data gathering from rendered documents
US8081849B2 (en) 2004-12-03 2011-12-20 Google Inc. Portable scanning and memory device
US9143638B2 (en) 2004-04-01 2015-09-22 Google Inc. Data capture from rendered documents using handheld device
US9116890B2 (en) 2004-04-01 2015-08-25 Google Inc. Triggering actions in response to optically or acoustically capturing keywords from a rendered document
US9008447B2 (en) 2004-04-01 2015-04-14 Google Inc. Method and system for character recognition
US7894670B2 (en) 2004-04-01 2011-02-22 Exbiblio B.V. Triggering actions in response to optically or acoustically capturing keywords from a rendered document
US8146156B2 (en) 2004-04-01 2012-03-27 Google Inc. Archive of text captures from rendered documents
US8713418B2 (en) 2004-04-12 2014-04-29 Google Inc. Adding value to a rendered document
US8620083B2 (en) 2004-12-03 2013-12-31 Google Inc. Method and system for character recognition
US8874504B2 (en) 2004-12-03 2014-10-28 Google Inc. Processing techniques for visual capture data from a rendered document
US8489624B2 (en) 2004-05-17 2013-07-16 Google, Inc. Processing techniques for text capture from a rendered document
EP1615045A1 (en) * 2004-07-07 2006-01-11 STMicroelectronics Limited Optical detection of relative motion
US8346620B2 (en) 2004-07-19 2013-01-01 Google Inc. Automatic modification of web pages
US7490770B2 (en) * 2004-08-12 2009-02-17 Datalogic Scanning, Inc. System and method of optical reading with enhanced depth of field collection
US7290711B2 (en) * 2004-09-30 2007-11-06 Symbol Technologies, Inc. System and method for reducing motion sensitivity in an imager based optical code reader
US7721966B2 (en) 2004-10-18 2010-05-25 Datalogic Scanning, Inc. System and method of optical reading employing virtual scan lines
US20090080037A1 (en) * 2004-10-19 2009-03-26 Kenichi Hayashi Condenser Lens and Optical Scanning Device
US20060087580A1 (en) * 2004-10-21 2006-04-27 Symagery Microsystems Inc. Variable focal distance image reader
EP1828958B1 (en) * 2004-12-01 2012-05-16 Datalogic ADC, Inc. Triggering illumination for a data reader
WO2006062818A2 (en) * 2004-12-08 2006-06-15 Symbol Technologies, Inc. Swipe imager scan engine
US7617984B2 (en) 2004-12-16 2009-11-17 Cognex Technology And Investment Corporation Hand held symbology reader illumination diffuser
US9292724B1 (en) 2004-12-16 2016-03-22 Cognex Corporation Hand held symbology reader illumination diffuser with aimer optics
US7163150B2 (en) * 2005-01-10 2007-01-16 Psion Teklogix Systems Inc. Targeting system for a portable image reader
US7499090B2 (en) 2005-01-27 2009-03-03 Datalogic Scanning, Inc. Rolling-reset imager with optical filter
US7215493B2 (en) * 2005-01-27 2007-05-08 Psc Scanning, Inc. Imaging system with a lens having increased light collection efficiency and a deblurring equalizer
US7224540B2 (en) * 2005-01-31 2007-05-29 Datalogic Scanning, Inc. Extended depth of field imaging system using chromatic aberration
US7273177B2 (en) * 2005-03-10 2007-09-25 Microsoft Corporation Camera-based barcode recognition
EP3300359B1 (en) 2005-03-11 2021-06-30 Hand Held Products, Inc. Bar code reading device with global electronic shutter control
US7237721B2 (en) * 2005-05-24 2007-07-03 Nokia Corporation Image processing for pattern detection
JP4701013B2 (en) * 2005-06-01 2011-06-15 キヤノン株式会社 Conveying member and sheet processing apparatus
US20060278713A1 (en) * 2005-06-13 2006-12-14 Psc Scanning, Inc. Method and system for high speed optical scanning
US8286877B2 (en) * 2005-06-13 2012-10-16 Datalogic ADC, Inc. System and method for data reading using raster scanning
US20070063043A1 (en) * 2005-09-21 2007-03-22 Intermec Ip Corp. Optoelectronic reader and method for reading machine-readable symbols
KR100735367B1 (en) * 2005-10-27 2007-07-04 삼성전기주식회사 Optical System Having Multiple curvature Lens And Forming Method Thereof
US7614563B1 (en) 2005-12-29 2009-11-10 Cognex Technology And Investment Corporation System and method for providing diffuse illumination in a symbology reader
US7717343B2 (en) * 2006-01-12 2010-05-18 Hand Held Products, Inc. High-efficiency illumination in data collection devices
US20070164115A1 (en) * 2006-01-17 2007-07-19 Symboltechnologies, Inc. Automatic exposure system for imaging-based bar code reader
US8181878B2 (en) * 2006-01-25 2012-05-22 Cognex Technology And Investment Corporation Method and apparatus for providing a focus indication for optical imaging of visual codes
ATE512422T1 (en) * 2006-04-03 2011-06-15 Omnivision Cdm Optics Inc OPTICAL IMAGING SYSTEMS AND METHODS WITH NON-LINEAR AND/OR SPATIALLY VARYING IMAGE PROCESSING
US7461790B2 (en) * 2006-04-04 2008-12-09 Datalogic Scanning, Inc. Data reader and methods for imaging targets exposed to high intensity illumination
US20070267584A1 (en) * 2006-05-19 2007-11-22 Psc Scanning, Inc. Optical code reader using an anamorphic Scheimpflug optical system
US7549582B1 (en) 2006-05-30 2009-06-23 Cognex Technology And Investment Corporation Method and apparatus for adjusting focus in a symbology reader
US7784696B2 (en) 2006-06-09 2010-08-31 Hand Held Products, Inc. Indicia reading apparatus having image sensing and processing circuit
US7740176B2 (en) * 2006-06-09 2010-06-22 Hand Held Products, Inc. Indicia reading apparatus having reduced trigger-to-read time
CN101449194B (en) * 2006-06-20 2012-06-13 数据逻辑扫描公司 Imaging scanner with multiple image fields
US20080023555A1 (en) * 2006-07-31 2008-01-31 Edward Barkan Aperture stop in imaging reader
EP2067119A2 (en) 2006-09-08 2009-06-10 Exbiblio B.V. Optical scanners, such as hand-held optical scanners
US20080067250A1 (en) * 2006-09-19 2008-03-20 Igor Vinogradov Imaging reader and method with optically modified field of view
US8033670B2 (en) * 2006-12-29 2011-10-11 Cognex Corporation Manually adjustable ruggedized focus mechanism
WO2008086498A2 (en) * 2007-01-11 2008-07-17 Datalogic Scanning, Inc. Methods and systems for optical code reading using virtual scan lines
US7852519B2 (en) 2007-02-05 2010-12-14 Hand Held Products, Inc. Dual-tasking decoder for improved symbol reading
US20080239509A1 (en) * 2007-03-30 2008-10-02 Symbol Technologies, Inc. Compact imaging lens assembly for an imaging-based bar code reader
US20080277472A1 (en) * 2007-05-10 2008-11-13 Tohru Takahashi Two-dimensional code scanner with guide illumination
US7918398B2 (en) 2007-06-04 2011-04-05 Hand Held Products, Inc. Indicia reading terminal having multiple setting imaging lens
TWI475874B (en) * 2007-06-15 2015-03-01 Camtek Ltd Optical inspection system using multi-facet imaging
US8201743B2 (en) * 2007-06-28 2012-06-19 Symbol Technologies, Inc. Bar code reader with improved lens for imaging with balanced astigmatism
EP2183635B1 (en) 2007-08-04 2015-09-16 Omnivision Technologies, Inc. Multi-region imaging systems
EP2031714B1 (en) * 2007-08-31 2010-08-18 EM Microelectronic-Marin SA Optoelectronic circuit with a photoreceiver and a laser diode, and a module comprising the same
US7886978B2 (en) 2007-09-20 2011-02-15 Microsoft Corporation Techniques for decoding images of barcodes
US20090084856A1 (en) * 2007-09-28 2009-04-02 Igor Vinogradov Imaging reader with asymmetrical magnification
US20090140048A1 (en) * 2007-11-30 2009-06-04 Symbol Technologies, Inc. CPC Illumination Apparatus for an Imaging-Based Bar Code Reader
US7967210B2 (en) * 2007-11-30 2011-06-28 Symbol Technologies, Inc. Imaging bar code reader having light emitting diode for generating a field of view
US8302864B2 (en) 2007-12-28 2012-11-06 Cognex Corporation Method and apparatus using aiming pattern for machine vision training
US8646689B2 (en) 2007-12-28 2014-02-11 Cognex Corporation Deformable light pattern for machine vision system
US8353457B2 (en) * 2008-02-12 2013-01-15 Datalogic ADC, Inc. Systems and methods for forming a composite image of multiple portions of an object from multiple perspectives
US8608076B2 (en) * 2008-02-12 2013-12-17 Datalogic ADC, Inc. Monolithic mirror structure for use in a multi-perspective optical code reader
US8678287B2 (en) * 2008-02-12 2014-03-25 Datalogic ADC, Inc. Two-plane optical code reader for acquisition of multiple views of an object
WO2009102616A2 (en) 2008-02-12 2009-08-20 Datalogic Scanning, Inc. Systems and methods for forming a composite image of multiple portions of an object from multiple perspectives
US8033471B2 (en) * 2008-05-08 2011-10-11 Psion Teklogix Inc. Optical image reader
JP4743553B2 (en) * 2008-09-29 2011-08-10 京セラ株式会社 Lens unit, imaging device, and electronic device
US8436909B2 (en) * 2008-10-21 2013-05-07 Stmicroelectronics S.R.L. Compound camera sensor and related method of processing digital images
US8628015B2 (en) 2008-10-31 2014-01-14 Hand Held Products, Inc. Indicia reading terminal including frame quality evaluation processing
US8393543B2 (en) * 2008-12-10 2013-03-12 Optoelectronics Co., Ltd. Method and apparatus for scanning optical code
US8083148B2 (en) 2008-12-16 2011-12-27 Hand Held Products, Inc. Indicia reading terminal including frame processing
CN102349079B (en) 2008-12-26 2015-09-16 数据逻辑扫描公司 For obtaining two planar wave code readers of multiple views of object
US8322621B2 (en) 2008-12-26 2012-12-04 Datalogic ADC, Inc. Image-based code reader for acquisition of multiple views of an object and methods for employing same
US8261990B2 (en) 2008-12-26 2012-09-11 Datalogic ADC, Inc. Data reader having compact arrangement for acquisition of multiple views of an object
US8134116B2 (en) * 2009-01-12 2012-03-13 Cognex Corporation Modular focus system for image based code readers
CN105930311B (en) 2009-02-18 2018-10-09 谷歌有限责任公司 Execute method, mobile device and the readable medium with the associated action of rendered document
WO2010105245A2 (en) * 2009-03-12 2010-09-16 Exbiblio B.V. Automatically providing content associated with captured information, such as information captured in real-time
US8447066B2 (en) 2009-03-12 2013-05-21 Google Inc. Performing actions based on capturing information from rendered documents, such as documents under copyright
US8622302B2 (en) * 2009-03-17 2014-01-07 Datalogic ADC, Inc. Systems and methods for compensating for fixed pattern noise
US20100316291A1 (en) * 2009-06-11 2010-12-16 Shulan Deng Imaging terminal having data compression
JP2012163581A (en) * 2009-06-19 2012-08-30 Sharp Corp Display device
US8373108B2 (en) * 2009-08-12 2013-02-12 Hand Held Products, Inc. Indicia reading terminal operative for processing of frames having plurality of frame featurizations
US8295601B2 (en) * 2009-08-12 2012-10-23 Hand Held Products, Inc. Indicia reading terminal having multiple exposure periods and methods for same
US8256678B2 (en) 2009-08-12 2012-09-04 Hand Held Products, Inc. Indicia reading terminal having image sensor and variable lens assembly
US8587595B2 (en) 2009-10-01 2013-11-19 Hand Held Products, Inc. Low power multi-core decoder system and method
US8225747B2 (en) * 2009-10-14 2012-07-24 The Kong Company, Llc Treat dispenser for animals and method
US9081799B2 (en) 2009-12-04 2015-07-14 Google Inc. Using gestalt information to identify locations in printed information
US9323784B2 (en) 2009-12-09 2016-04-26 Google Inc. Image search using text-based elements within the contents of images
CN102131044B (en) * 2010-01-20 2014-03-26 鸿富锦精密工业(深圳)有限公司 Camera module
US20110186639A1 (en) * 2010-02-04 2011-08-04 Metrologic Instruments, Inc. Contact aperture for imaging apparatus
JP5670481B2 (en) * 2010-02-19 2015-02-18 デュアル・アパーチャー・インコーポレーテッド Multi-aperture image data processing
EP2550801B1 (en) 2010-03-22 2017-05-03 Hologic Inc. Decorrelated channel sampling for digital imaging
US8816284B2 (en) * 2010-03-30 2014-08-26 Lawrence Livermore National Security, Llc. Room-temperature quantum noise limited spectrometry and methods of the same
EP2558903A1 (en) * 2010-04-13 2013-02-20 University Court Of The University Of St Andrews Optical eigenmode imaging
US8939374B2 (en) 2010-12-30 2015-01-27 Hand Held Products, Inc. Terminal having illumination and exposure control
DE102011001880B4 (en) 2011-04-07 2014-02-27 Sick Ag Optoelectronic barcode reader and lens for it
US8777106B2 (en) * 2011-06-23 2014-07-15 Symbol Technologies, Inc. Imaging reader with non-uniform magnification within a field of view
US8985459B2 (en) 2011-06-30 2015-03-24 Metrologic Instruments, Inc. Decodable indicia reading terminal with combined illumination
US8387883B1 (en) * 2011-08-26 2013-03-05 Honeywell International Inc. Bar code imagers
US8844823B2 (en) * 2011-09-15 2014-09-30 Metrologic Instruments, Inc. Laser scanning system employing an optics module capable of forming a laser beam having an extended depth of focus (DOF) over the laser scanning field
US8729653B2 (en) 2011-10-26 2014-05-20 Omnivision Technologies, Inc. Integrated die-level cameras and methods of manufacturing the same
US8629926B2 (en) 2011-11-04 2014-01-14 Honeywell International, Inc. Imaging apparatus comprising image sensor array having shared global shutter circuitry
US10498933B2 (en) 2011-11-22 2019-12-03 Cognex Corporation Camera system with exchangeable illumination assembly
US8947590B2 (en) 2011-11-22 2015-02-03 Cognex Corporation Vision system camera with mount for multiple lens types
US11366284B2 (en) 2011-11-22 2022-06-21 Cognex Corporation Vision system camera with mount for multiple lens types and lens module for the same
US20130138094A1 (en) 2011-11-30 2013-05-30 Amo Development, Llc System and method for ophthalmic surface measurements based on sequential estimates
US9518864B2 (en) * 2011-12-15 2016-12-13 Facebook, Inc. Controllable optical sensing
US9004363B2 (en) * 2012-04-27 2015-04-14 Symbol Technologies, Inc. Diffuser engine for barcode imaging scanner
US9746636B2 (en) 2012-10-19 2017-08-29 Cognex Corporation Carrier frame and circuit board for an electronic device
JP2014103353A (en) * 2012-11-22 2014-06-05 Samsung R&D Institute Japan Co Ltd Recognition device, recognition method, mounting device, and mounting method
US9128356B1 (en) 2013-03-13 2015-09-08 Lockheed Martin Corporation System and method for providing a polarization projector utilizing a polarization sensitive lens
JP6396638B2 (en) * 2013-03-29 2018-09-26 マクセル株式会社 Phase filter, imaging optical system, and imaging system
US9016578B2 (en) * 2013-05-14 2015-04-28 Symbol Technologies, Inc. Apparatus for and method of electro-optically reading a target in the presence of ambient light by suppressing the ambient light
TWI487951B (en) 2013-05-27 2015-06-11 Largan Precision Co Ltd Plastic lens
WO2016102988A1 (en) 2014-12-24 2016-06-30 Datalogic Automation, Inc. Multiline scanner
USD791137S1 (en) * 2015-07-22 2017-07-04 Hand Held Products, Inc. Scanner
US10368021B2 (en) * 2016-08-26 2019-07-30 Mems Start, Llc Systems and methods for derivative sensing using filtering pixels
US10958885B2 (en) 2016-08-26 2021-03-23 Mems Start, Llc Filtering imaging system including a light source to output an optical signal modulated with a code
US9798912B1 (en) * 2016-09-26 2017-10-24 Symbol Technologies, Llc Imaging module and reader for, and method of, reading targets by image capture with a substantially constant resolution over an extended range of working distances
WO2018134654A1 (en) 2017-01-17 2018-07-26 Six Degrees Space Ltd Wide field of view optical module for linear sensor
CN110383114B (en) 2017-02-24 2020-12-29 富士胶片株式会社 Lens, zoom lens and imaging lens
US10817685B2 (en) * 2017-09-28 2020-10-27 Datalogic Ip Tech S.R.L. System and method for illuminating a target of a barcode reader
US10679024B2 (en) 2018-07-24 2020-06-09 Cognex Corporation System and method for auto-focusing a vision system camera on barcodes
US11333748B2 (en) 2018-09-17 2022-05-17 Waymo Llc Array of light detectors with corresponding array of optical elements

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US2811081A (en) * 1952-07-03 1957-10-29 Henry L Praeger Universal focal length photo-objective system
US4743773A (en) * 1984-08-23 1988-05-10 Nippon Electric Industry Co., Ltd. Bar code scanner with diffusion filter and plural linear light source arrays
US4843222A (en) * 1986-05-29 1989-06-27 Eastman Kodak Company Bar code reader for reading bar code symbols at different distances
US4874933A (en) * 1987-08-21 1989-10-17 Recognition Equipment Incorporated Ambient illumination bar code reader
US5210398A (en) * 1991-06-14 1993-05-11 Symbol Technologies, Inc. Optical scanner with extended depth of focus
WO1994015314A1 (en) * 1992-12-21 1994-07-07 Norand Corporation Portable optical reader system for reading optically readable information located within a three dimensional area

Family Cites Families (65)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3004470A (en) * 1956-07-28 1961-10-17 Zeiss Ikon A G Stuttgart Multiple focal length lens
US3801775A (en) * 1972-08-07 1974-04-02 Scanner Method and apparatus for identifying objects
US4000397A (en) * 1975-03-21 1976-12-28 Spectra-Physics, Inc. Signal processor method and apparatus
US4354101A (en) * 1977-04-15 1982-10-12 Msi Data Corporation Method and apparatus for reading and decoding a high density linear bar code
US4135161A (en) * 1977-05-31 1979-01-16 Torrieri Don J Method and means for pulse detection
GB2048606B (en) * 1979-02-28 1983-03-16 Barr & Stroud Ltd Optical scanning system
US4335302A (en) * 1980-08-20 1982-06-15 R.L.S. Industries, Inc. Bar code scanner using non-coherent light source
DE3243009A1 (en) * 1982-11-20 1984-05-24 Scantron Gmbh METHOD AND DEVICE FOR IDENTIFYING OBJECTS
JPS59147306A (en) * 1983-02-10 1984-08-23 Sony Corp Focus error detector
JPS60140597A (en) * 1983-12-28 1985-07-25 Giichi Kuze Eprom programmer
JPH0762858B2 (en) * 1985-06-06 1995-07-05 株式会社サトー Bar code reading method and device
US4697098A (en) * 1985-06-10 1987-09-29 Priam Corporation Composite gate generator circuit for detecting valid data signals
JPS61285404A (en) * 1985-06-12 1986-12-16 Mitsubishi Electric Corp Optical device
US4740675A (en) * 1986-04-10 1988-04-26 Hewlett-Packard Company Digital bar code slot reader with threshold comparison of the differentiated bar code signal
US4782220A (en) * 1986-09-29 1988-11-01 Mars, Incorporated Method and apparatus for bar code data autodiscrimination
US4818886A (en) * 1986-11-12 1989-04-04 Quential, Inc. Method and apparatus for self-referencing and self-focusing a bar-code reader
US4879456A (en) * 1987-06-18 1989-11-07 Spectra-Physics, Inc. Method of decoding a binary scan signal
US4749879A (en) * 1987-06-18 1988-06-07 Spectra-Physics, Inc. Signal transition detection method and system
JPS6474687A (en) * 1987-09-16 1989-03-20 Yamatake Honeywell Co Ltd Reading method for crossed square-shaped code
US4861972B1 (en) * 1987-11-05 2000-12-05 Spectra Physics Scanning Syst Bar code scanner and method of programming
FR2631476B1 (en) * 1988-05-10 1992-01-03 Bertin & Cie METHOD AND DEVICE FOR READING A BAR CODE ON A SUBSTANTIALLY IMMOBILE MEDIUM
US4978860A (en) * 1988-09-21 1990-12-18 Hewlett-Packard Company Optical system for a large depth-of-field bar code scanner
US5103209A (en) * 1989-01-09 1992-04-07 Checkpoint Systems, Inc. Electronic article surveillance system with improved differentiation
US5010241A (en) * 1989-01-12 1991-04-23 Hewlett-Packard Company Sensor array and illumination system for a large depth-of-field bar code scanner
US5319181A (en) * 1992-03-16 1994-06-07 Symbol Technologies, Inc. Method and apparatus for decoding two-dimensional bar code using CCD/CMD camera
FR2646252B1 (en) * 1989-04-21 1991-12-13 France Etat PROCESS AND DEVICE HOLOGRAPHIC IN INCOHERENT LIGHT FOR THE STUDY OF TERRESTRIAL RELIEF
FR2646251B1 (en) * 1989-04-21 1991-08-09 France Etat IMPROVED HOLOGRAPHIC DEVICE IN INCOHERENT LIGHT
US4939356A (en) * 1989-05-02 1990-07-03 Spectra-Physics, Inc. Bar code scanner with asterisk scan pattern
US5408081A (en) * 1989-06-16 1995-04-18 Symbol Technologies, Inc. Digital circuit for a laser scanner using a first derivative signal and a comparison signal
US5446272A (en) * 1989-06-16 1995-08-29 Symbol Technologies, Inc. System for digitizing a scanned signal indicating changes in signal intensity
JPH0827890B2 (en) * 1989-09-07 1996-03-21 株式会社テック Boarding pass reader
US4963719A (en) * 1989-11-01 1990-10-16 Spectra-Physics Bar code scanner and method of scanning
US5272323A (en) * 1989-11-22 1993-12-21 Symbol Technologies, Inc. Digitizer for barcode scanner
WO1991012551A1 (en) * 1990-02-14 1991-08-22 Massachusetts Institute Of Technology Lens/zone plate combination for chromatic dispersion correction
US5144118A (en) * 1990-03-05 1992-09-01 Spectra-Physics, Inc. Bar code scanning system with multiple decoding microprocessors
US5155343A (en) * 1990-03-28 1992-10-13 Chandler Donald G Omnidirectional bar code reader with method and apparatus for detecting and scanning a bar code symbol
JPH0415774A (en) * 1990-05-02 1992-01-21 Eastman Kodak Japan Kk Bar code reader
US5121009A (en) * 1990-06-15 1992-06-09 Novatel Communications Ltd. Linear phase low pass filter
US5229797A (en) * 1990-08-08 1993-07-20 Minnesota Mining And Manufacturing Company Multifocal diffractive ophthalmic lenses
US5071207A (en) * 1990-09-25 1991-12-10 The United States Of America As Represented By The United States Department Of Energy Broadband diffractive lens or imaging element
US5124537A (en) * 1990-10-29 1992-06-23 Omniplanar, Inc. Omnidirectional bar code reader using virtual scan of video raster scan memory
US5103080A (en) * 1990-12-10 1992-04-07 Symbol Technologies, Inc. Digitizer signal processing circuit for a bar code
US5144460A (en) * 1990-12-11 1992-09-01 The Dz Company High contrast-resolution camera
US5200597A (en) * 1991-02-07 1993-04-06 Psc, Inc. Digitally controlled system for scanning and reading bar codes
US5296690A (en) * 1991-03-28 1994-03-22 Omniplanar, Inc. System for locating and determining the orientation of bar codes in a two-dimensional image
DE4213910A1 (en) * 1991-05-03 1992-11-05 Minnesota Mining & Mfg HOLOGRAPHY MIRROR WITH SUPERZONES
CA2053460A1 (en) * 1991-09-13 1993-03-14 Eugene Bernard Joseph Analog waveform decoder
US5298728A (en) * 1991-11-01 1994-03-29 Spectra-Physics Scanning System, Inc. Signal processing apparatus and method
US5202784A (en) * 1992-01-10 1993-04-13 Spectra-Physics Scanning Systems, Inc. Optical system for data reading applications
WO1993017396A1 (en) * 1992-02-21 1993-09-02 Spectra-Physics Scanning Systems, Inc. Bar code scanner
US5349172A (en) * 1992-02-27 1994-09-20 Alex Roustaei Optical scanning head
JP2755516B2 (en) * 1992-03-10 1998-05-20 アルプス電気株式会社 Automatic identification decoding device
US5327171A (en) * 1992-05-26 1994-07-05 United Parcel Service Of America, Inc. Camera system optics
JP3187936B2 (en) * 1992-05-29 2001-07-16 オリンパス光学工業株式会社 Barcode reader
US5343028A (en) * 1992-08-10 1994-08-30 United Parcel Service Of America, Inc. Method and apparatus for detecting and decoding bar code symbols using two-dimensional digital pixel images
US5329105A (en) * 1992-08-10 1994-07-12 United Parcel Service Of America, Inc. Method and apparatus for determining the width of elements of bar code symbols
US5479011A (en) * 1992-12-18 1995-12-26 Spectra-Physics Scanning Systems, Inc. Variable focus optical system for data reading
US5347121A (en) * 1992-12-18 1994-09-13 Spectra-Physics Scanning Systems, Inc. Variable focus optical system for data reading
JP3003897B2 (en) * 1993-02-02 2000-01-31 株式会社東芝 Transport device and ticket processing device
US5317211A (en) * 1993-03-05 1994-05-31 Motorola, Inc. Programmable pin for use in programmable logic devices
US5463211A (en) * 1993-05-07 1995-10-31 Spectra-Physics Scanning Systems, Inc. Method and apparatus for detecting transitions in a time sampled input signal
US5304787A (en) * 1993-06-01 1994-04-19 Metamedia Corporation Locating 2-D bar codes
US5386105A (en) * 1993-06-07 1995-01-31 Psc Inc. Diffractive optical beam shaping methods and apparatus for providing enhanced depth of working range of bar code scanners
US5446271A (en) * 1993-08-06 1995-08-29 Spectra-Physics Scanning Systems, Inc. Omnidirectional scanning method and apparatus
US5471515A (en) * 1994-01-28 1995-11-28 California Institute Of Technology Active pixel sensor with intra-pixel charge transfer

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US2811081A (en) * 1952-07-03 1957-10-29 Henry L Praeger Universal focal length photo-objective system
US4743773A (en) * 1984-08-23 1988-05-10 Nippon Electric Industry Co., Ltd. Bar code scanner with diffusion filter and plural linear light source arrays
US4843222A (en) * 1986-05-29 1989-06-27 Eastman Kodak Company Bar code reader for reading bar code symbols at different distances
US4874933A (en) * 1987-08-21 1989-10-17 Recognition Equipment Incorporated Ambient illumination bar code reader
US5210398A (en) * 1991-06-14 1993-05-11 Symbol Technologies, Inc. Optical scanner with extended depth of focus
WO1994015314A1 (en) * 1992-12-21 1994-07-07 Norand Corporation Portable optical reader system for reading optically readable information located within a three dimensional area

Cited By (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1998047377A1 (en) * 1997-04-21 1998-10-29 Intermec Scanner Technology Center Optoelectronic device for image acquisition, in particular of code bars
US6375076B1 (en) 1997-04-21 2002-04-23 Intermec Ip Corp. Optoelectronic device for acquisition of images, in particular of bar codes
US6732930B2 (en) 1999-12-23 2004-05-11 Intermec Ip Corp. Optoelectronic device and process for acquiring symbols, such as bar codes, using a two-dimensional sensor
US7143941B2 (en) 1999-12-23 2006-12-05 Intermec Ip Corp. Optoelectronic device and process for acquiring symbols, such as bar codes, using a two-dimensional sensor
US7413125B2 (en) 1999-12-23 2008-08-19 Intermec Ip Corp. Optoelectronic device and process for acquiring symbols, such as bar codes, using a two-dimensional sensor
US11317050B2 (en) 2005-03-11 2022-04-26 Hand Held Products, Inc. Image reader comprising CMOS based image sensor array
US10958863B2 (en) 2005-03-11 2021-03-23 Hand Held Products, Inc. Image reader comprising CMOS based image sensor array
US11323649B2 (en) 2005-03-11 2022-05-03 Hand Held Products, Inc. Image reader comprising CMOS based image sensor array
US9578269B2 (en) 2005-03-11 2017-02-21 Hand Held Products, Inc. Image reader comprising CMOS based image sensor array
US11863897B2 (en) 2005-03-11 2024-01-02 Hand Held Products, Inc. Image reader comprising CMOS based image sensor array
US8720781B2 (en) 2005-03-11 2014-05-13 Hand Held Products, Inc. Image reader having image sensor array
US8978985B2 (en) 2005-03-11 2015-03-17 Hand Held Products, Inc. Image reader comprising CMOS based image sensor array
US11323650B2 (en) 2005-03-11 2022-05-03 Hand Held Products, Inc. Image reader comprising CMOS based image sensor array
US10735684B2 (en) 2005-03-11 2020-08-04 Hand Held Products, Inc. Image reader comprising CMOS based image sensor array
US9305199B2 (en) 2005-03-11 2016-04-05 Hand Held Products, Inc. Image reader having image sensor array
US10721429B2 (en) 2005-03-11 2020-07-21 Hand Held Products, Inc. Image reader comprising CMOS based image sensor array
US10171767B2 (en) 2005-03-11 2019-01-01 Hand Held Products, Inc. Image reader comprising CMOS based image sensor array
US9465970B2 (en) 2005-03-11 2016-10-11 Hand Held Products, Inc. Image reader comprising CMOS based image sensor array
US9576169B2 (en) 2005-03-11 2017-02-21 Hand Held Products, Inc. Image reader having image sensor array
US8720785B2 (en) 2005-06-03 2014-05-13 Hand Held Products, Inc. Apparatus having hybrid monochrome and color image sensor array
US11238252B2 (en) 2005-06-03 2022-02-01 Hand Held Products, Inc. Apparatus having hybrid monochrome and color image sensor array
US9454686B2 (en) 2005-06-03 2016-09-27 Hand Held Products, Inc. Apparatus having hybrid monochrome and color image sensor array
US10691907B2 (en) 2005-06-03 2020-06-23 Hand Held Products, Inc. Apparatus having hybrid monochrome and color image sensor array
US9438867B2 (en) 2005-06-03 2016-09-06 Hand Held Products, Inc. Digital picture taking optical reader having hybrid monochrome and color image sensor array
US9092654B2 (en) 2005-06-03 2015-07-28 Hand Held Products, Inc. Digital picture taking optical reader having hybrid monochrome and color image sensor array
US10949634B2 (en) 2005-06-03 2021-03-16 Hand Held Products, Inc. Apparatus having hybrid monochrome and color image sensor array
US9058527B2 (en) 2005-06-03 2015-06-16 Hand Held Products, Inc. Apparatus having hybrid monochrome and color image sensor array
US11625550B2 (en) 2005-06-03 2023-04-11 Hand Held Products, Inc. Apparatus having hybrid monochrome and color image sensor array
US10002272B2 (en) 2005-06-03 2018-06-19 Hand Held Products, Inc. Apparatus having hybrid monochrome and color image sensor array
US11238251B2 (en) 2005-06-03 2022-02-01 Hand Held Products, Inc. Apparatus having hybrid monochrome and color image sensor array
US11604933B2 (en) 2005-06-03 2023-03-14 Hand Held Products, Inc. Apparatus having hybrid monochrome and color image sensor array
US7780089B2 (en) 2005-06-03 2010-08-24 Hand Held Products, Inc. Digital picture taking optical reader having hybrid monochrome and color image sensor array
US7770799B2 (en) 2005-06-03 2010-08-10 Hand Held Products, Inc. Optical reader having reduced specular reflection read failures
DE102008009404B4 (en) * 2008-02-15 2013-01-24 Marson Technology Co., Ltd. All-round radiating bar code reader
DE102008009404A1 (en) * 2008-02-15 2009-09-03 MARSON TECHNOLOGY CO., LTD., Hsin-Tien Polydirectional radiating barcode reader for reading linear bar codes or two-dimensional bar codes, has housing which is provided at side with through-hole
CN105684437B (en) * 2013-11-03 2021-10-29 杜比实验室特许公司 System and method for local dimming in multi-modulation displays

Also Published As

Publication number Publication date
US6073851A (en) 2000-06-13
US6042012A (en) 2000-03-28
GB2301691A (en) 1996-12-11
JPH09512372A (en) 1997-12-09
DE19581524T1 (en) 1997-06-05
GB9617173D0 (en) 1996-09-25
GB2301691B (en) 1998-09-30
US5814803A (en) 1998-09-29
DE19581524B4 (en) 2007-10-11

Similar Documents

Publication Publication Date Title
US5814803A (en) Image reader with multi-focus lens
US5770847A (en) Bar code reader with multi-focus lens
US8286877B2 (en) System and method for data reading using raster scanning
CA2347420C (en) Variable focus optical system
CA2343326C (en) Optical focusing device
CN101449194B (en) Imaging scanner with multiple image fields
US5484994A (en) Optical scanning head with improved resolution
US5354977A (en) Optical scanning head
EP3905099B1 (en) Bar code reading device with global electronic shutter control
JP2001522103A (en) Electronic edge detection device
US20060278712A1 (en) Method and system for data reading using raster scanning
US20090159685A1 (en) Optimizing Optical Quality of a Sensor in a Bar Code Reader
US6293468B1 (en) Pulsed barcode scanner
US8056806B2 (en) Machine-readable symbol reader and method employing an ultracompact light concentrator with adaptive field of view
JPH0628511A (en) Image recognition apparatus and image detection method
FR2831712A1 (en) SEMICONDUCTOR DEVICE FOR IMAGE FORMATION OF OPTICAL CODES SYMBOLS AND IMAGER
WO2006138165A2 (en) System and method for data reading using raster scanning
EP1891575A2 (en) System and method for data reading using raster scanning
EP0680641A1 (en) Optical scanning head
JP2004145633A (en) Light converging method for data reader

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): DE GB JP

WWE Wipo information: entry into national phase

Ref document number: 9617173.1

Country of ref document: GB

RET De translation (de og part 6b)

Ref document number: 19581524

Country of ref document: DE

Date of ref document: 19970605

WWE Wipo information: entry into national phase

Ref document number: 19581524

Country of ref document: DE

REG Reference to national code

Ref country code: DE

Ref legal event code: 8607

REG Reference to national code

Ref country code: DE

Ref legal event code: 8607