US8118226B2 - High-resolution optical code imaging using a color imager - Google Patents

High-resolution optical code imaging using a color imager Download PDF

Info

Publication number
US8118226B2
US8118226B2 US12/704,313 US70431310A US8118226B2 US 8118226 B2 US8118226 B2 US 8118226B2 US 70431310 A US70431310 A US 70431310A US 8118226 B2 US8118226 B2 US 8118226B2
Authority
US
United States
Prior art keywords
light
sensor elements
optical code
infrared
reading device
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US12/704,313
Other versions
US20100200658A1 (en
Inventor
Bryan L. Olmstead
Alan Shearin
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Datalogic ADC Inc
Original Assignee
Datalogic Scanning Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Datalogic Scanning Inc filed Critical Datalogic Scanning Inc
Priority to US12/704,313 priority Critical patent/US8118226B2/en
Assigned to DATALOGIC SCANNING, INC. reassignment DATALOGIC SCANNING, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: OLMSTEAD, BRYAN L., SHEARIN, ALAN
Publication of US20100200658A1 publication Critical patent/US20100200658A1/en
Application granted granted Critical
Publication of US8118226B2 publication Critical patent/US8118226B2/en
Assigned to Datalogic ADC, Inc. reassignment Datalogic ADC, Inc. MERGER (SEE DOCUMENT FOR DETAILS). Assignors: DATALOGIC SCANNING, INC.
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K7/00Methods or arrangements for sensing record carriers, e.g. for reading patterns
    • G06K7/10Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
    • G06K7/10544Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum
    • G06K7/10712Fixed beam scanning
    • G06K7/10722Photodetector array or CCD scanning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K7/00Methods or arrangements for sensing record carriers, e.g. for reading patterns
    • G06K7/10Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
    • G06K7/10544Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum
    • G06K7/10712Fixed beam scanning
    • G06K7/10722Photodetector array or CCD scanning
    • G06K7/10732Light sources
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K7/00Methods or arrangements for sensing record carriers, e.g. for reading patterns
    • G06K7/10Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
    • G06K7/12Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using a selected wavelength, e.g. to sense red marks and ignore blue marks

Definitions

  • the field of this disclosure relates generally to systems and methods of data reading, and more particularly but not exclusively to reading of optical codes (e.g., bar codes).
  • optical codes e.g., bar codes
  • Optical codes encode useful, optically-readable information about the items to which they are attached or otherwise associated.
  • an optical code is the bar code.
  • Bar codes are ubiquitously found on or associated with objects of various types, such as the packaging of retail, wholesale, and inventory goods; retail product presentation fixtures (e.g., shelves); goods undergoing manufacturing; personal or company assets; and documents.
  • a bar code typically serves as an identifier of an object, whether the identification be to a class of objects (e.g., containers of milk) or a unique item.
  • a typical linear or one-dimensional bar code such as a UPC code, consist of alternating bars (i.e., relatively dark areas) and spaces (i.e., relatively light areas).
  • the pattern of alternating bars and spaces and the widths of those bars and spaces represent a string of binary ones and zeros, wherein the width of any particular bar or space is an integer multiple of a specified minimum width, which is called a “module” or “unit.”
  • a bar code reader must be able to reliably discern the pattern of bars and spaces, such as by determining the locations of edges demarking adjacent bars and spaces from one another, across the entire length of the bar code.
  • Bar codes are just one example of the many types of optical codes in use today.
  • Higher-dimensional optical codes such as, two-dimensional matrix codes (e.g., MaxiCode) or stacked codes (e.g., PDF 417), which are also sometimes referred to as “bar codes,” are also used for various purposes.
  • optical code reading devices are available for capturing an optical code and for decoding the information represented by the optical code.
  • image-based readers are available that include imagers, such as charge coupled devices (CODs) or complementary metal oxide semiconductor (CMOS) imagers, that generate electronic image data that represent an image of a captured optical code.
  • Image-based readers are used for reading one-dimensional optical codes and higher-dimensional optical codes. Because optical codes most often include dark and light patterns (e.g., black and white) that represent binary data, imagers of image-based readers are typically monochrome so that uniform sensitivity for each pixel of the imager is achieved.
  • typical image-based readers include light sources that illuminate the image-based reader's field of view with narrowband visible light to achieve high optical resolution by avoiding chromatic aberration and polychromatic diffraction effects.
  • Narrowband light sources typically used for imaging include laser diodes, having a bandwidth on the order of 5 nanometers (nm), and light emitting diodes (LEDs), having a bandwidth on the order of 50 nm.
  • This disclosure describes improved optical reading devices and associated methods.
  • One embodiment is directed to an optical code reading device that includes a color image sensor array positioned to sense light reflected from an object in a field of view of the optical code reading device and to produce from the sensed reflected light image data representing an image of the object.
  • the color image sensor array has a first set of sensor elements that are sensitive to a first visible wavelength band of light, and a second set of sensor elements that are sensitive to a second visible wavelength band of light. The first and second sets of sensor elements are also sensitive to light within an infrared wavelength band.
  • the optical code reading device includes an artificial illumination source positioned to illuminate the field of view of the optical code reading device with light that is incident upon and reflected from the object in the field of view toward the image sensor array.
  • the illumination source is operable to produce infrared light having wavelengths within the infrared wavelength band so that, upon illumination of the field of view, at least some sensor elements of each of the first and second sets are sensitive to the infrared light and contribute to production of the image data.
  • FIG. 1 is a diagram of an imaging system according to a preferred embodiment.
  • FIG. 2 is a diagram of a color filter pattern of a color image sensor array of the imaging system of FIG. 1 .
  • FIG. 3 is a graph of the sensitivity of blue, green, and red sensor elements as a function of light wavelength of an illustrative color image sensor array used in the imaging system of FIG. 1 .
  • FIG. 4 is a flowchart showing the operational steps of the imaging system of FIG. 1 .
  • FIG. 5 is a flowchart of an illumination calibration method according to one embodiment.
  • FIG. 1 is a diagram of an imaging device 100 , such as an optical code reading device, according to one embodiment.
  • the imaging device 100 includes a color image sensor array 102 contained within a housing 103 of the imaging device 100 .
  • the color image sensor array 102 includes a first set of sensor elements 104 and a second set of sensor elements 106 .
  • the sensor elements of the color image sensor array 102 may be arranged in a one-dimensional array or, preferably, a two-dimensional array.
  • the color image sensor array 102 may be a CCD, such as a frame-transfer or interline-transfer CCD.
  • the color image sensor array 102 may alternatively be a CMOS imager, such as a global shuttered or rolling-reset CMOS imager.
  • Suitable imagers or image sensor arrays are available, for example, from Aptina Imaging Corporation of San Jose, Calif., USA including, but not limited to, a model MT9V022 VGA color imager. Imagers are available from many manufacturers and are available in various resolutions (numbers of pixels). For higher resolution applications, other imagers from Aptina Imaging Corporation are suitable, including the model MT9M001 with 1.3 megapixels, the model MT9M002 with 1.6 megapixels, and the model MT9P001 with 5 megapixels. As imaging technology advances, imager resolution increases and other imagers may also be suitable.
  • the color image sensor array 102 may include more than two sets of sensor elements.
  • the color image sensor array 102 may include three sets of sensor elements 104 , 106 , and 202 arranged in a Bayer pattern as shown in FIG. 2 . Each set of sensor elements corresponds to a different color.
  • the first set 104 may be sensitive to light having wavelengths that correspond to the color green (G) (wavelengths ranging between about 500 nanometers (nm) and about 600 nm)
  • the second set 106 may be sensitive to light having wavelengths that correspond to the color red (R) (wavelengths ranging between about 600 nm and about 750 nm)
  • the third set 202 may be sensitive to light having wavelengths that correspond to the color blue (B) (wavelengths ranging between about 400 nm and about 500 nm).
  • a color filter associated with each sensor element of the different sets appreciably filters out visible light that does not correspond to its color (i.e., the color filters associated with the first set 104 appreciably block out red and blue light).
  • FIG. 3 is a graph depicting an example of the quantum efficiency percentage versus the wavelength of light incident upon red, green and blue sensor elements of the model MT9V022 VGA color imager available from Aptina Imaging Corporation that may be used as the color image sensor array 102 .
  • a curve 104 ′ corresponding to the spectral sensitivity of the sensor elements of the first set 104 , has a local peak 104 a ′ at a wavelength corresponding to the color green.
  • a curve 106 ′ corresponding to the spectral sensitivity of the sensor elements of the second set 106 , has a local peak 106 a ′ at a wavelength corresponding to the color red.
  • a curve 202 ′ corresponding to the spectral sensitivity of the sensor elements of the third set 202 , has a local peak 202 a ′ at a wavelength corresponding to the color blue.
  • the curves 104 ′, 106 ′, and 202 ′ also have respective local peaks 104 b ′, 106 b ′, and 202 b ′ near a common wavelength that corresponds non-visible light—in this case infrared light.
  • the first set 104 , the second set 106 , and the third set 202 are not only sensitive to green, red, and blue light, respectively, but also to light within an infrared wavelength band.
  • the infrared wavelength band, to which the first set 104 , the second set 106 , and the third set 202 are sensitive may be relatively narrow (e.g., no more than about 100 nm) and may include 850 nm. Also, the quantum efficiency percentage associated with the local peaks 104 b ′, 106 b ′, and 202 b ′ may be substantially the same or within a narrow percentage range.
  • the sensitivity of the sets 104 , 106 , and 202 to infrared light may be substantially equal or within a narrow sensitivity range (e.g., about five percent in quantum efficiency) so that an average intensity value of light sensed by the first set 104 may be substantially equal to an average intensity value of light sensed by the second set 106 and an average intensity value of light sensed by the third set 202 .
  • a narrow sensitivity range e.g., about five percent in quantum efficiency
  • the color image sensor array 102 need not be limited to three sets of sensor elements or the colors red, green, and blue, and the color image sensor array 102 may include color filter patterns other than the Bayer pattern.
  • the color image sensor array 102 may include a cyan, yellow, green, and magenta (CYGM) filter or a red, green, blue, and emerald (RGBE) filter in which the sensor elements of the different colors are also sensitive to light within an infrared wavelength band.
  • CYGM cyan, yellow, green, and magenta
  • RGBE emerald
  • the color filter pattern used on the color filter array 102 may be chosen to achieve accurate color rendition or to improve sensitivity in a color photograph application. While these distinctions are not necessary in the present embodiment, the imaging device 100 and its associated methods are flexible to compensate for the effects of these various filters.
  • the imaging device 100 may also include one or more artificial illumination sources 108 (two illumination sources are depicted in FIG. 1 ).
  • the artificial illumination sources 108 may be mounted to a printed circuit board 110 upon which the color image sensor array 102 is also mounted.
  • the artificial illumination sources 108 are operable to emit infrared illumination.
  • the infrared illumination emitted by the artificial illumination sources 108 may be narrowband infrared illumination (e.g., illumination having a bandwidth less than about 100 nm).
  • the wavelength bandwidth of light emitted by the artificial illumination sources 108 preferably includes 850 nm, when using a color image sensor array with characteristics shown in FIG. 3 .
  • the imaging device 100 typically includes a suitable optical system 112 positioned to focus light upon the color image sensor array 102 .
  • the optical system 112 may include conventional optical components, such as one or more lenses, an aperture, and, in some cases, a mechanical shutter.
  • the color image sensor array 102 may include electronic shuttering means.
  • the optical system 112 may also include one or more optical filters to block out certain wavelengths of light.
  • the optical system 112 excludes an infrared filter, which is operable to block out infrared light, and may include one or more optical filters that are operable to block out light that does not have wavelengths within the infrared wavelength band.
  • the artificial illumination sources 108 are shown as being mounted on the printed circuit board 110 , the artificial illumination sources 108 may be positioned in other convenient locations to provide illumination of the object 114 .
  • the artificial illumination sources 108 illuminate the field of view 116 with infrared illumination (step 402 ) or illumination of another non-visible frequency at which all of the sensor elements have an acceptable response. If an object 114 (e.g., an optical code) is within the field of view 116 of the imaging device, infrared light reflects off the object 114 toward the optical system 112 . Infrared light that is incident on the optical system 112 is focused by the optical system 112 onto the sensor elements of the color image sensor array 102 (step 404 ).
  • Sensor elements of the first set 104 , the second set 106 , and the third set 202 sense the focused infrared light (steps 406 a , 406 b , and 406 c ).
  • the color image sensor array 102 produces image data based upon the infrared light that is incident on, and sensed by, the sensor elements of the first, second, and third sets 104 , 106 , and 202 (step 408 ).
  • An enclosure 118 may cover the color image sensor array 102 except where the optical system 112 is located, so that an appreciable amount of light from sources other than the artificial illumination sources 108 does not reach the color image sensor array 102 .
  • each of the sets 104 , 106 , and 202 are sensitive to infrared light, each of the sets 104 , 106 , and 202 contribute to production of the image data and high-resolution infrared imaging of the object 114 may be achieved. Each of the sets 104 , 106 , and 202 may contribute to the production of the image data to a sufficiently equal extent that no one of the set 104 , 106 , or 202 contributes to the image data appreciably more than the other two sets.
  • the resolution of an infrared image represented by the image data may be substantially equal to a resolution produced by a monochrome image sensor array having the same size of sensor elements and the same number of sensor elements as the sum of the number of sensor elements in the sets 104 , 106 , and 202 .
  • the color image sensor array 102 when illuminated with infrared light, the color image sensor array 102 may achieve a resolution substantially equivalent to a monochrome imager.
  • the artificial illumination sources 108 emit visible light having red, green, and blue light components.
  • the artificial illumination sources 108 emit visible polychromatic (white) light, or a combination of monochromatic or quasi-monochromatic lights having wavelengths corresponding to the colors red, green and blue.
  • the artificial illumination sources 108 include a red light, a green light and a blue light (e.g., red, green and blue light emitting diodes). Light emission intensities of each of the red, green and blue light components are calibrated to compensate for the behavior of the color image sensor array 102 described below.
  • a conventional optical code reading device that includes a monochrome imager
  • data representing an image captured by the monochrome imager are converted to grayscale where the shade of gray produced by a sensor element of the monochrome imager depends upon the light intensity level captured by it.
  • the patterned color filter that covers the array of sensor elements effects the transmittance of light, and, thus, the intensity of light that is incident on the sensor elements of the sets 104 , 106 , and 202 .
  • the transmittance of light associated with the filter portions may be different between colors such that the filter portions of one color may transmit more (or less) light than the filter portions of the other colors.
  • the sensor elements behind the filter portions may be inherently more or less sensitive to certain wavelengths of light (e.g., the sensor elements may be more sensitive to red wavelengths than to blue and green wavelengths).
  • the sensor elements may be more sensitive to red wavelengths than to blue and green wavelengths.
  • the differences in the quantum efficiencies of the sets 104 , 106 and 202 may be determined and the light sources 108 may be calibrated according to an illumination calibration method 500 shown in the flowchart of FIG. 5 .
  • the color image sensor array 102 is uniformly illuminated with a light source, such as the sources 108 or another light source (step 502 ).
  • the color image sensor array 102 is directly illuminated with the light source.
  • a white background such as a white piece of paper
  • an image of the white background is captured by the color image sensor array 102 .
  • Each of the sets 104 , 106 and 202 produce a set of image data representing light intensity levels captured by the sensor elements.
  • the light intensity levels represented in the sets of image data are compared to determine the relative sensitivity of each of the sets 104 , 106 and 202 (step 504 ).
  • an average intensity level for each of the sets 104 , 106 and 202 is computed from the sets of image data and the averages are compared to determine the relative differences in quantum efficiencies between the sets 104 , 106 and 202 .
  • the light emission intensity levels of the red, green and blue light components of the artificial illumination sources 108 are adjusted to compensate for the sensitivity differences between the sets 104 , 106 and 202 (step 506 ). For example, if the quantum efficiency of the second set 106 is greater than the quantum efficiency of the first set 104 , the intensity level of the green component emitted by the artificial illumination sources 108 is selected to be greater than the intensity level of the red component by an amount proportional to the difference between the quantum efficiencies of the first and second sets 104 and 106 . In one example, the intensity levels of the red, green and blue light components are selected so that the average intensity levels of light captured by the sets 104 , 106 and 202 are substantially the same.
  • Adjustment of the light emission intensity levels can be implemented by varying an amount of current supplied to the red, green, and blue lights of the artificial illumination sources 108 .
  • the color image sensor array may include an analog or digital gain for each of the sets 104 , 106 and 202 , which are adjusted instead of the light emission intensity levels of the artificial illumination sources 108 to provide a uniform image output.
  • This alternative embodiment may allow the artificial illumination sources 108 to provide a more pleasing color of illumination, such as white.
  • the calibrated red, green and blue light components illuminate the object 114
  • the sensor elements of the sets 104 , 106 and 202 capture an image of the object 114 and produced image data
  • the image data are converted to grayscale. Similar to the first embodiment, high-resolution imaging may be realized because each of the sets 104 , 106 , and 202 contributes to the production of the image data.
  • the optical system 112 includes a dispersive (intentionally uncorrected chromatic aberration) lens assembly.
  • a dispersive lens assembly For example, U.S. Pat. No. 7,224,540 (“the '540 patent”), the entire contents of which are incorporated herein by reference, describes a dispersive lens assembly.
  • Artificial illumination sources 108 emit white light or multiple quasi-monochromatic lights that includes red, green, and blue light. As discussed in the '540 patent, because of the lens assembly's chromatic aberration, a wavelength dependent focal shift (i.e., longitudinal aberration) of the reflected light occurs.
  • the data from each of the sets 104 , 106 , and 202 of the color image sensor array 102 are processed as separate color sub-images, called color planes (such as red, green, and blue color planes).
  • the depth of field of each color plane is distinct and typically overlapping.
  • the focus quality of each color plane depends on the distance between the object 114 and the imaging device 100 . For example, if the distance between the object 114 and the imaging device 100 provides the best green focus, the green color plane of the first set 104 will be well focused while the red and blue color planes of the second set 106 and the third set 202 will be blurred. Thus, by comparing the color planes of the first, second, and third sets 104 , 106 , and 202 , distance between the object 114 and the imaging device 100 can be estimated by determining which color plane is in best focus.
  • One of several different methods may be used for comparing the sharpness of the color planes. For example, techniques used in passive auto-focusing systems may be used. As additional examples, the amount of high spatial frequency content of the color planes may be measured (such as by computing a Fourier transform), or edge sharpness for the color planes may be measured (i.e., measuring the intensity difference between neighboring pixels). By knowing which color plane is in best focus, and by knowing the depth of field associated with the best focused color plane, distance between the object 114 and the imaging device 100 may be estimated. Distance estimation using an optical code reading device may be useful in a number of applications. For example, estimating distance may be used for measuring the dimensions of an optical code to ascertain whether the optical code is of acceptable size (i.e., whether the optical code was printed properly).
  • the third embodiment also provides an increased overall depth of field for the imaging device 100 because the first, second, and third sets 104 , 106 , and 202 form focused images at different distances, although the resolution of the image from each color plane is lower in the case of infrared illumination, since only one pixel set is capturing a well focused image.
  • certain embodiments may be capable of achieving certain advantages, including (1) enabling utilization of lower cost color imagers in optical code reading devices; (2) achieving higher image resolution by utilizing all of an image sensor's pixels to contribute to an image; (3) avoidance of visible illumination from the illumination source, which can be discernable and annoying to human observers; (4) extending the depth of field of an optical code reader or measuring distance to an object using a chromatically aberrated lens, and (5) compensation for manufacturing variation in the sensitivity of different colored pixels.
  • Other advantages of various embodiments will be apparent upon reading the above sections. Combinations of the above embodiments, and other embodiments not specifically described herein will be apparent to skilled persons upon reviewing the above description.

Abstract

An optical code or other data reading device includes a color image sensor array positioned to sense light reflected from an object, and to produce image data. In one configuration, the color image sensor array has multiple sets (e.g., first and second sets) of sensor elements that are sensitive to corresponding visible wavelength bands of light (e.g., first and second wavelength bands), the sets also being sensitive to light within an infrared wavelength band. An artificial illumination source is positioned to illuminate the field of view with light that is reflected off an object in the field of view toward the image sensor array, the illumination source being operable to produce infrared light having wavelengths within the infrared wavelength band so that, upon illumination, at least some sensor elements of each of the sets are sensitive to the infrared light and contribute to production of the image data.

Description

RELATED APPLICATION
This application claims benefit under 35 U.S.C. §119(e) of U.S. Provisional Application No. 61/151,768, titled “High-Resolution Optical Code Imaging Using a Color Imager,” filed Feb. 11, 2009, the entire contents of which is incorporated herein by reference.
TECHNICAL FIELD
The field of this disclosure relates generally to systems and methods of data reading, and more particularly but not exclusively to reading of optical codes (e.g., bar codes).
BACKGROUND INFORMATION
Optical codes encode useful, optically-readable information about the items to which they are attached or otherwise associated. Perhaps the most common example of an optical code is the bar code. Bar codes are ubiquitously found on or associated with objects of various types, such as the packaging of retail, wholesale, and inventory goods; retail product presentation fixtures (e.g., shelves); goods undergoing manufacturing; personal or company assets; and documents. By encoding information, a bar code typically serves as an identifier of an object, whether the identification be to a class of objects (e.g., containers of milk) or a unique item. A typical linear or one-dimensional bar code, such as a UPC code, consist of alternating bars (i.e., relatively dark areas) and spaces (i.e., relatively light areas). The pattern of alternating bars and spaces and the widths of those bars and spaces represent a string of binary ones and zeros, wherein the width of any particular bar or space is an integer multiple of a specified minimum width, which is called a “module” or “unit.” Thus, to decode the information, a bar code reader must be able to reliably discern the pattern of bars and spaces, such as by determining the locations of edges demarking adjacent bars and spaces from one another, across the entire length of the bar code.
Bar codes are just one example of the many types of optical codes in use today. Higher-dimensional optical codes, such as, two-dimensional matrix codes (e.g., MaxiCode) or stacked codes (e.g., PDF 417), which are also sometimes referred to as “bar codes,” are also used for various purposes.
Different methods and types of optical code reading devices are available for capturing an optical code and for decoding the information represented by the optical code. For example, image-based readers are available that include imagers, such as charge coupled devices (CODs) or complementary metal oxide semiconductor (CMOS) imagers, that generate electronic image data that represent an image of a captured optical code. Image-based readers are used for reading one-dimensional optical codes and higher-dimensional optical codes. Because optical codes most often include dark and light patterns (e.g., black and white) that represent binary data, imagers of image-based readers are typically monochrome so that uniform sensitivity for each pixel of the imager is achieved. Also, typical image-based readers include light sources that illuminate the image-based reader's field of view with narrowband visible light to achieve high optical resolution by avoiding chromatic aberration and polychromatic diffraction effects. Narrowband light sources typically used for imaging include laser diodes, having a bandwidth on the order of 5 nanometers (nm), and light emitting diodes (LEDs), having a bandwidth on the order of 50 nm.
Common imagers made for image capturing devices, such as still cameras and video cameras, however, are color imagers—not monochrome. Because imagers made for many image capturing devices are color, color imagers are generally made in higher volume and have become less expensive than monochrome imagers. Some image-based readers have included color imagers, but the present inventors have recognized that those readers do not effectively achieve high optical resolution comparable to monochrome image-based readers with the same number and size of pixels.
SUMMARY OF THE DISCLOSURE
This disclosure describes improved optical reading devices and associated methods.
One embodiment is directed to an optical code reading device that includes a color image sensor array positioned to sense light reflected from an object in a field of view of the optical code reading device and to produce from the sensed reflected light image data representing an image of the object. The color image sensor array has a first set of sensor elements that are sensitive to a first visible wavelength band of light, and a second set of sensor elements that are sensitive to a second visible wavelength band of light. The first and second sets of sensor elements are also sensitive to light within an infrared wavelength band. The optical code reading device includes an artificial illumination source positioned to illuminate the field of view of the optical code reading device with light that is incident upon and reflected from the object in the field of view toward the image sensor array. The illumination source is operable to produce infrared light having wavelengths within the infrared wavelength band so that, upon illumination of the field of view, at least some sensor elements of each of the first and second sets are sensitive to the infrared light and contribute to production of the image data.
Additional aspects and advantages will be apparent from the following detailed description of preferred embodiments, which proceeds with reference to the accompanying drawings.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a diagram of an imaging system according to a preferred embodiment.
FIG. 2 is a diagram of a color filter pattern of a color image sensor array of the imaging system of FIG. 1.
FIG. 3 is a graph of the sensitivity of blue, green, and red sensor elements as a function of light wavelength of an illustrative color image sensor array used in the imaging system of FIG. 1.
FIG. 4 is a flowchart showing the operational steps of the imaging system of FIG. 1.
FIG. 5 is a flowchart of an illumination calibration method according to one embodiment.
DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
With reference to the above-listed drawings, this section describes particular embodiments and their detailed construction and operation.
FIG. 1 is a diagram of an imaging device 100, such as an optical code reading device, according to one embodiment. The imaging device 100 includes a color image sensor array 102 contained within a housing 103 of the imaging device 100. The color image sensor array 102 includes a first set of sensor elements 104 and a second set of sensor elements 106. The sensor elements of the color image sensor array 102 may be arranged in a one-dimensional array or, preferably, a two-dimensional array. The color image sensor array 102 may be a CCD, such as a frame-transfer or interline-transfer CCD. The color image sensor array 102 may alternatively be a CMOS imager, such as a global shuttered or rolling-reset CMOS imager. Suitable imagers or image sensor arrays are available, for example, from Aptina Imaging Corporation of San Jose, Calif., USA including, but not limited to, a model MT9V022 VGA color imager. Imagers are available from many manufacturers and are available in various resolutions (numbers of pixels). For higher resolution applications, other imagers from Aptina Imaging Corporation are suitable, including the model MT9M001 with 1.3 megapixels, the model MT9M002 with 1.6 megapixels, and the model MT9P001 with 5 megapixels. As imaging technology advances, imager resolution increases and other imagers may also be suitable.
The color image sensor array 102 may include more than two sets of sensor elements. For example, the color image sensor array 102 may include three sets of sensor elements 104, 106, and 202 arranged in a Bayer pattern as shown in FIG. 2. Each set of sensor elements corresponds to a different color. For example, the first set 104 may be sensitive to light having wavelengths that correspond to the color green (G) (wavelengths ranging between about 500 nanometers (nm) and about 600 nm), the second set 106 may be sensitive to light having wavelengths that correspond to the color red (R) (wavelengths ranging between about 600 nm and about 750 nm), and the third set 202 may be sensitive to light having wavelengths that correspond to the color blue (B) (wavelengths ranging between about 400 nm and about 500 nm). Moreover, a color filter associated with each sensor element of the different sets appreciably filters out visible light that does not correspond to its color (i.e., the color filters associated with the first set 104 appreciably block out red and blue light).
FIG. 3 is a graph depicting an example of the quantum efficiency percentage versus the wavelength of light incident upon red, green and blue sensor elements of the model MT9V022 VGA color imager available from Aptina Imaging Corporation that may be used as the color image sensor array 102. A curve 104′, corresponding to the spectral sensitivity of the sensor elements of the first set 104, has a local peak 104 a′ at a wavelength corresponding to the color green. A curve 106′, corresponding to the spectral sensitivity of the sensor elements of the second set 106, has a local peak 106 a′ at a wavelength corresponding to the color red. A curve 202′, corresponding to the spectral sensitivity of the sensor elements of the third set 202, has a local peak 202 a′ at a wavelength corresponding to the color blue. The curves 104′, 106′, and 202′ also have respective local peaks 104 b′, 106 b′, and 202 b′ near a common wavelength that corresponds non-visible light—in this case infrared light. In other words, the first set 104, the second set 106, and the third set 202 are not only sensitive to green, red, and blue light, respectively, but also to light within an infrared wavelength band. The infrared wavelength band, to which the first set 104, the second set 106, and the third set 202 are sensitive, may be relatively narrow (e.g., no more than about 100 nm) and may include 850 nm. Also, the quantum efficiency percentage associated with the local peaks 104 b′, 106 b′, and 202 b′ may be substantially the same or within a narrow percentage range. In other words, the sensitivity of the sets 104, 106, and 202 to infrared light may be substantially equal or within a narrow sensitivity range (e.g., about five percent in quantum efficiency) so that an average intensity value of light sensed by the first set 104 may be substantially equal to an average intensity value of light sensed by the second set 106 and an average intensity value of light sensed by the third set 202.
The color image sensor array 102 need not be limited to three sets of sensor elements or the colors red, green, and blue, and the color image sensor array 102 may include color filter patterns other than the Bayer pattern. For example, the color image sensor array 102 may include a cyan, yellow, green, and magenta (CYGM) filter or a red, green, blue, and emerald (RGBE) filter in which the sensor elements of the different colors are also sensitive to light within an infrared wavelength band. The color filter pattern used on the color filter array 102 may be chosen to achieve accurate color rendition or to improve sensitivity in a color photograph application. While these distinctions are not necessary in the present embodiment, the imaging device 100 and its associated methods are flexible to compensate for the effects of these various filters.
The imaging device 100 may also include one or more artificial illumination sources 108 (two illumination sources are depicted in FIG. 1). The artificial illumination sources 108 may be mounted to a printed circuit board 110 upon which the color image sensor array 102 is also mounted. In a first embodiment, the artificial illumination sources 108 are operable to emit infrared illumination. The infrared illumination emitted by the artificial illumination sources 108 may be narrowband infrared illumination (e.g., illumination having a bandwidth less than about 100 nm). Also, the wavelength bandwidth of light emitted by the artificial illumination sources 108 preferably includes 850 nm, when using a color image sensor array with characteristics shown in FIG. 3.
The imaging device 100 typically includes a suitable optical system 112 positioned to focus light upon the color image sensor array 102. The optical system 112 may include conventional optical components, such as one or more lenses, an aperture, and, in some cases, a mechanical shutter. As an alternative to a mechanical shutter, the color image sensor array 102 may include electronic shuttering means. The optical system 112 may also include one or more optical filters to block out certain wavelengths of light. In one example, when infrared illumination sources are selected for the artificial illumination sources 108, the optical system 112 excludes an infrared filter, which is operable to block out infrared light, and may include one or more optical filters that are operable to block out light that does not have wavelengths within the infrared wavelength band. Although the artificial illumination sources 108 are shown as being mounted on the printed circuit board 110, the artificial illumination sources 108 may be positioned in other convenient locations to provide illumination of the object 114.
A preferred operation of the imaging device 100 will now be described with reference to a flowchart 400 of FIG. 4. The artificial illumination sources 108 illuminate the field of view 116 with infrared illumination (step 402) or illumination of another non-visible frequency at which all of the sensor elements have an acceptable response. If an object 114 (e.g., an optical code) is within the field of view 116 of the imaging device, infrared light reflects off the object 114 toward the optical system 112. Infrared light that is incident on the optical system 112 is focused by the optical system 112 onto the sensor elements of the color image sensor array 102 (step 404). Sensor elements of the first set 104, the second set 106, and the third set 202 sense the focused infrared light ( steps 406 a, 406 b, and 406 c). The color image sensor array 102 produces image data based upon the infrared light that is incident on, and sensed by, the sensor elements of the first, second, and third sets 104, 106, and 202 (step 408). An enclosure 118 may cover the color image sensor array 102 except where the optical system 112 is located, so that an appreciable amount of light from sources other than the artificial illumination sources 108 does not reach the color image sensor array 102. Because each of the sets 104, 106, and 202 are sensitive to infrared light, each of the sets 104, 106, and 202 contribute to production of the image data and high-resolution infrared imaging of the object 114 may be achieved. Each of the sets 104, 106, and 202 may contribute to the production of the image data to a sufficiently equal extent that no one of the set 104, 106, or 202 contributes to the image data appreciably more than the other two sets. The resolution of an infrared image represented by the image data may be substantially equal to a resolution produced by a monochrome image sensor array having the same size of sensor elements and the same number of sensor elements as the sum of the number of sensor elements in the sets 104, 106, and 202. In other words, when illuminated with infrared light, the color image sensor array 102 may achieve a resolution substantially equivalent to a monochrome imager.
In a second alternative embodiment, the artificial illumination sources 108 emit visible light having red, green, and blue light components. For example, the artificial illumination sources 108 emit visible polychromatic (white) light, or a combination of monochromatic or quasi-monochromatic lights having wavelengths corresponding to the colors red, green and blue. In one configuration, the artificial illumination sources 108 include a red light, a green light and a blue light (e.g., red, green and blue light emitting diodes). Light emission intensities of each of the red, green and blue light components are calibrated to compensate for the behavior of the color image sensor array 102 described below. In a conventional optical code reading device that includes a monochrome imager, data representing an image captured by the monochrome imager are converted to grayscale where the shade of gray produced by a sensor element of the monochrome imager depends upon the light intensity level captured by it. For the color image sensor array 102, the patterned color filter that covers the array of sensor elements effects the transmittance of light, and, thus, the intensity of light that is incident on the sensor elements of the sets 104, 106, and 202. Also, the transmittance of light associated with the filter portions may be different between colors such that the filter portions of one color may transmit more (or less) light than the filter portions of the other colors. Moreover, the sensor elements behind the filter portions may be inherently more or less sensitive to certain wavelengths of light (e.g., the sensor elements may be more sensitive to red wavelengths than to blue and green wavelengths). The effects of the color dependent differences in light transmittances and pixel sensitivities can be seen in the example of FIG. 3 in which the quantum efficiencies associated with the local peaks 104 a′, 106 a′ and 202 a′ are different from one another.
Prior to operation, the differences in the quantum efficiencies of the sets 104, 106 and 202 may be determined and the light sources 108 may be calibrated according to an illumination calibration method 500 shown in the flowchart of FIG. 5. First, the color image sensor array 102 is uniformly illuminated with a light source, such as the sources 108 or another light source (step 502). In one example, the color image sensor array 102 is directly illuminated with the light source. In another example, a white background (such as a white piece of paper) is illuminated, and an image of the white background is captured by the color image sensor array 102. Each of the sets 104, 106 and 202 produce a set of image data representing light intensity levels captured by the sensor elements. The light intensity levels represented in the sets of image data are compared to determine the relative sensitivity of each of the sets 104, 106 and 202 (step 504). In one example, an average intensity level for each of the sets 104, 106 and 202 is computed from the sets of image data and the averages are compared to determine the relative differences in quantum efficiencies between the sets 104, 106 and 202.
After the image data are analyzed, the light emission intensity levels of the red, green and blue light components of the artificial illumination sources 108 are adjusted to compensate for the sensitivity differences between the sets 104, 106 and 202 (step 506). For example, if the quantum efficiency of the second set 106 is greater than the quantum efficiency of the first set 104, the intensity level of the green component emitted by the artificial illumination sources 108 is selected to be greater than the intensity level of the red component by an amount proportional to the difference between the quantum efficiencies of the first and second sets 104 and 106. In one example, the intensity levels of the red, green and blue light components are selected so that the average intensity levels of light captured by the sets 104, 106 and 202 are substantially the same. Adjustment of the light emission intensity levels can be implemented by varying an amount of current supplied to the red, green, and blue lights of the artificial illumination sources 108. Alternatively, the color image sensor array may include an analog or digital gain for each of the sets 104, 106 and 202, which are adjusted instead of the light emission intensity levels of the artificial illumination sources 108 to provide a uniform image output. This alternative embodiment may allow the artificial illumination sources 108 to provide a more pleasing color of illumination, such as white.
In operation, the calibrated red, green and blue light components illuminate the object 114, the sensor elements of the sets 104, 106 and 202 capture an image of the object 114 and produced image data, and the image data are converted to grayscale. Similar to the first embodiment, high-resolution imaging may be realized because each of the sets 104, 106, and 202 contributes to the production of the image data.
In a third alternative embodiment, the optical system 112 includes a dispersive (intentionally uncorrected chromatic aberration) lens assembly. For example, U.S. Pat. No. 7,224,540 (“the '540 patent”), the entire contents of which are incorporated herein by reference, describes a dispersive lens assembly. Artificial illumination sources 108 emit white light or multiple quasi-monochromatic lights that includes red, green, and blue light. As discussed in the '540 patent, because of the lens assembly's chromatic aberration, a wavelength dependent focal shift (i.e., longitudinal aberration) of the reflected light occurs. The data from each of the sets 104, 106, and 202 of the color image sensor array 102 are processed as separate color sub-images, called color planes (such as red, green, and blue color planes). The depth of field of each color plane is distinct and typically overlapping. The focus quality of each color plane depends on the distance between the object 114 and the imaging device 100. For example, if the distance between the object 114 and the imaging device 100 provides the best green focus, the green color plane of the first set 104 will be well focused while the red and blue color planes of the second set 106 and the third set 202 will be blurred. Thus, by comparing the color planes of the first, second, and third sets 104, 106, and 202, distance between the object 114 and the imaging device 100 can be estimated by determining which color plane is in best focus.
One of several different methods may be used for comparing the sharpness of the color planes. For example, techniques used in passive auto-focusing systems may be used. As additional examples, the amount of high spatial frequency content of the color planes may be measured (such as by computing a Fourier transform), or edge sharpness for the color planes may be measured (i.e., measuring the intensity difference between neighboring pixels). By knowing which color plane is in best focus, and by knowing the depth of field associated with the best focused color plane, distance between the object 114 and the imaging device 100 may be estimated. Distance estimation using an optical code reading device may be useful in a number of applications. For example, estimating distance may be used for measuring the dimensions of an optical code to ascertain whether the optical code is of acceptable size (i.e., whether the optical code was printed properly). The third embodiment also provides an increased overall depth of field for the imaging device 100 because the first, second, and third sets 104, 106, and 202 form focused images at different distances, although the resolution of the image from each color plane is lower in the case of infrared illumination, since only one pixel set is capturing a well focused image.
As skilled persons will appreciate in light of this disclosure, certain embodiments may be capable of achieving certain advantages, including (1) enabling utilization of lower cost color imagers in optical code reading devices; (2) achieving higher image resolution by utilizing all of an image sensor's pixels to contribute to an image; (3) avoidance of visible illumination from the illumination source, which can be discernable and annoying to human observers; (4) extending the depth of field of an optical code reader or measuring distance to an object using a chromatically aberrated lens, and (5) compensation for manufacturing variation in the sensitivity of different colored pixels. Other advantages of various embodiments will be apparent upon reading the above sections. Combinations of the above embodiments, and other embodiments not specifically described herein will be apparent to skilled persons upon reviewing the above description.
Though the present invention has been set forth in the form of its preferred embodiments, it is nevertheless intended that modifications to the disclosed systems and methods may be made without departing from inventive concepts set forth herein.

Claims (20)

The invention claimed is:
1. A method of imaging an object in a field of view of an optical code reading device, the optical code reading device having a color image sensor array, the method comprising:
illuminating the field of view of the optical code reading device with artificial infrared light that is incident upon the object, thereby producing reflected infrared light, the reflected infrared light having wavelengths within an infrared wavelength band;
sensing the reflected infrared light with a first set of sensor elements of the color image sensor array, the first set of sensor elements being sensitive to the infrared light and to visible light having wavelengths within a first visible wavelength band;
sensing the reflected infrared light with a second set of sensor elements of the color image sensor array, the second set of sensor elements being sensitive to the infrared light and to visible light having wavelengths within a second visible wavelength band different from the first visible wavelength band; and
producing image data from the first and second sets of sensor elements, the image data being derived from the infrared light sensed by the first and second sets of sensor elements and representing a high-resolution infrared image of the object.
2. A method according to claim 1, further comprising:
sensing the reflected infrared light with a third set of sensor elements of the color image sensor array, the third set of sensor elements being sensitive to the infrared light and to visible light having wavelengths within a third visible wavelength band different from each of the first and second visible wavelength bands; and
producing the image data from the first, second and third sets of sensor elements, the image data being derived from the infrared light sensed by the first, second and third sets of sensor elements.
3. A method according to claim 2, wherein the resolution of the infrared image represented by the image data is substantially equal to a resolution produced by a monochrome image sensor array having the same number of sensor elements as the sum of the numbers of sensor elements in the first, second and third sets of sensor elements.
4. A method according to claim 2, wherein the first, second and third visible wavelength bands correspond to the colors red, green and blue.
5. A method according to claim 1, wherein the infrared wavelength band includes 850 nm.
6. A method according to claim 1, wherein upon illumination of the field of view, sensor elements of each of the first and second sets of sensor elements are sensitive to the infrared light to a sufficiently equal extent such that an average intensity value of light sensed by the first set of sensor elements is substantially equal to an average intensity value of light sensed by the second set of sensor elements.
7. A method according to claim 1, further comprising:
intermittently blocking the reflected infrared light from reaching the sensor elements of the first and second sets by a mechanical shutter.
8. An optical code reading device comprising:
a color image sensor array positioned to sense light reflected from an object in a field of view of the optical code reading device and to produce from the sensed reflected light image data representing an infrared image of the object, the color image sensor array comprising:
a first set of sensor elements that are sensitive to a first visible wavelength band of light, and
a second set of sensor elements that are sensitive to a second visible wavelength band of light, wherein the first and second sets of sensor elements are sensitive to light within an infrared wavelength band; and
an artificial illumination source positioned to illuminate the field of view of the optical code reading device with light that is incident upon and reflected from the object in the field of view toward the color image sensor array, the artificial illumination source being operable to produce infrared light having a wavelength within the infrared wavelength band so that, upon illumination of the field of view, at least some sensor elements of each of the first and second sets are sensitive to the infrared light and contribute to production of the image data.
9. An optical code reading device according to claim 8, wherein the color image sensor array includes a third set of sensor elements that are sensitive to a third visible wavelength band of light and to light within the infrared wavelength band.
10. An optical code reading device according to claim 9, wherein the first, second and third visible wavelength bands correspond to the colors red, green and blue.
11. An optical code reading device according to claim 8, wherein the infrared wavelength band includes 850 nm.
12. An optical code reading device according to claim 8, wherein upon illumination of the field of view, sensor elements of each of the first and second sets of sensor elements are sensitive to the infrared light to a sufficiently equal extent such that an average intensity value of light sensed by the first set of sensor elements is substantially equal to an average intensity value of light sensed by the second set of sensor elements.
13. An optical code reading device according to claim 8, wherein the color image sensor array is part of a rolling-reset CMOS imager.
14. An optical code reading device according to claim 8, further comprising:
a mechanical shutter positioned between the color image sensor array and the field of view.
15. An optical code reading device comprising:
a color image sensor array positioned to sense light reflected from an object in a field of view of the optical code reading device and to produce from the sensed reflected light image data representing an image of the object, the color image sensor array having a first set, a second set, and a third set of sensor elements, wherein:
the first set of sensor elements are characterized by a first spectral sensitivity that varies as a first function of wavelength, the first function having a local peak at a first peak sensitivity wavelength and also having a local peak near a first non-visible wavelength;
the second set of sensor elements are characterized by a second spectral sensitivity that varies as a second function of wavelength, the second function having a local peak at a second peak sensitivity wavelength and also having a local peak near the first non-visible wavelength;
the third set of sensor elements are characterized by a third spectral sensitivity that varies as a third function of wavelength, the third function having a local peak at a third peak sensitivity wavelength and also having a local peak near the first non-visible wavelength; and
the first, second and third peak sensitivity wavelengths being different wavelengths; and
an artificial illumination source positioned to illuminate the field of view of the optical code reading device with light that is incident upon and reflected from the object in the field of view toward the image sensor array, the illumination source being operable to produce radiation including light having the first non-visible wavelength.
16. An optical code reading device according to claim 15, wherein the first non-visible wavelength is an infrared wavelength.
17. A method of calibrating light emission of an optical code reading device, the optical code reading device including a color image sensor array and an artificial illumination source, the color image sensor array including a first set of sensor elements that are sensitive to a first visible wavelength band of light and a second set of sensor elements that are sensitive to a second visible wavelength band of light different than the first visible wavelength band of light, the artificial illumination source operable to emit visible light having a first light component corresponding to a wavelength in the first visible wavelength band of light and a second light component corresponding to a wavelength in the second visible wavelength band of light, comprising:
illuminating the color image sensor array with visible light having the first light component and the second light component;
sensing the first light component with the first set of sensor elements, the first set of sensor elements characterized by a first quantum efficiency;
sensing the second light component with the second set of sensor elements, the second set of sensor elements characterized by a second quantum efficiency that is less than the first quantum efficiency;
producing a first set of image data representing light intensities sensed by the first set of sensor elements and a second set of image data representing light intensities sensed by the second set of sensor elements;
determining a difference between the first and second quantum efficiencies by comparing the first and second sets of image data; and
selecting for the artificial illumination source a first light emission intensity level corresponding to the first light component and a second light emission intensity level corresponding to the second light component, the second light emission intensity level being different than the first light emission intensity level by an amount proportional to the difference between the first and second quantum efficiencies of the first and second sets of sensor elements.
18. The method of claim 17, further comprising selecting the first and second light emission intensity levels of the first and second light components so that, upon illumination, an average intensity of light sensed by the first set of sensor elements is substantially the same as an average intensity level of light sensed by the second set of sensor elements.
19. The method of claim 17, wherein the illuminating step further comprises illuminating with the artificial illumination source an object in a field of view of the optical code reading device, the visible light reflecting off the object toward the color image sensor array to thereby illuminate the color image sensor array.
20. An optical code reading device comprising:
a color image sensor array positioned to sense light reflected from an object in a field of view of the optical code reading device, the color image sensor array comprising:
a first set of sensor elements that are sensitive to a first visible wavelength band of light, the first set of sensor elements characterized by a first quantum efficiency; and
a second set of sensor elements that are sensitive to a second visible wavelength band of light, the second set of sensor elements characterized by a second quantum efficiency, the first and second quantum efficiencies having a difference therebetween in which the second quantum efficiency is less than the first quantum efficiency; and
an artificial illumination source positioned to illuminate the field of view of the optical code reading device with light that is incident on and reflected from the object in the field of view toward the color image sensor array, the artificial illumination source being operable to produce light including:
a first light component having a first light emission intensity level and a wavelength in the first visible wavelength band of light, and
a second light component having a second light emission intensity level and a wavelength in the second visible wavelength band of light, wherein the second light emission intensity level is different than the first light emission intensity level of the first light component by an amount that is proportional to the difference between the first and second quantum efficiencies.
US12/704,313 2009-02-11 2010-02-11 High-resolution optical code imaging using a color imager Active 2030-06-13 US8118226B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/704,313 US8118226B2 (en) 2009-02-11 2010-02-11 High-resolution optical code imaging using a color imager

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US15176809P 2009-02-11 2009-02-11
US12/704,313 US8118226B2 (en) 2009-02-11 2010-02-11 High-resolution optical code imaging using a color imager

Publications (2)

Publication Number Publication Date
US20100200658A1 US20100200658A1 (en) 2010-08-12
US8118226B2 true US8118226B2 (en) 2012-02-21

Family

ID=42539596

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/704,313 Active 2030-06-13 US8118226B2 (en) 2009-02-11 2010-02-11 High-resolution optical code imaging using a color imager

Country Status (4)

Country Link
US (1) US8118226B2 (en)
EP (1) EP2396744B1 (en)
CN (1) CN102317951B (en)
WO (1) WO2010093806A2 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8556179B2 (en) * 2011-08-16 2013-10-15 Intermec Ip Corp. Motion tolerant bar code reader
US9173508B2 (en) 2010-07-08 2015-11-03 Itab Scanflow Ab Checkout counter
US9811705B1 (en) 2016-05-10 2017-11-07 Datalogic Ip Tech S.R.L. Illumination system with active element for generating different illumination patterns for a data reader
US20180095307A1 (en) * 2016-10-03 2018-04-05 Xerox Corporation Hyperspectral imaging system
US10452951B2 (en) * 2016-08-26 2019-10-22 Goodrich Corporation Active visual attention models for computer vision tasks

Families Citing this family (47)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7442629B2 (en) 2004-09-24 2008-10-28 President & Fellows Of Harvard College Femtosecond laser-induced formation of submicrometer spikes on a semiconductor substrate
US7057256B2 (en) 2001-05-25 2006-06-06 President & Fellows Of Harvard College Silicon-based visible and near-infrared optoelectric devices
US8800874B2 (en) * 2009-02-20 2014-08-12 Datalogic ADC, Inc. Systems and methods of optical code reading using a color imager
US8886206B2 (en) 2009-05-01 2014-11-11 Digimarc Corporation Methods and systems for content processing
US9749607B2 (en) 2009-07-16 2017-08-29 Digimarc Corporation Coordinated illumination and image signal capture for enhanced signal detection
US9673243B2 (en) 2009-09-17 2017-06-06 Sionyx, Llc Photosensitive imaging devices and associated methods
US9911781B2 (en) 2009-09-17 2018-03-06 Sionyx, Llc Photosensitive imaging devices and associated methods
TWI438699B (en) * 2009-10-09 2014-05-21 Primax Electronics Ltd Processing method of barcode and related apparatus thereof
US8692198B2 (en) 2010-04-21 2014-04-08 Sionyx, Inc. Photosensitive imaging devices and associated methods
CN103081128B (en) 2010-06-18 2016-11-02 西奥尼克斯公司 High-speed light sensitive device and correlation technique
TW201216168A (en) * 2010-10-08 2012-04-16 Primax Electronics Ltd Barcode evaluation method and barcode evaluation apparatus
US20130258112A1 (en) * 2010-12-21 2013-10-03 Zamir Recognition Systems Ltd. Visible light and ir hybrid digital camera
US9007689B2 (en) * 2011-04-04 2015-04-14 General Electric Company Method and apparatus for forming multiple images
US9496308B2 (en) 2011-06-09 2016-11-15 Sionyx, Llc Process module for increasing the response of backside illuminated photosensitive imagers and associated methods
US20130016203A1 (en) * 2011-07-13 2013-01-17 Saylor Stephen D Biometric imaging devices and associated methods
US8523074B2 (en) * 2011-08-26 2013-09-03 Honeywell International Inc. Bar code imagers
US9064764B2 (en) 2012-03-22 2015-06-23 Sionyx, Inc. Pixel isolation elements, devices, and associated methods
CN104412580B (en) * 2012-07-06 2016-04-06 富士胶片株式会社 Color image sensor and camera head
CA2841464C (en) * 2013-02-01 2016-10-25 Centre De Recherche Industrielle Du Quebec Apparatus and method for scanning a surface of an article
JP6466346B2 (en) 2013-02-15 2019-02-06 サイオニクス、エルエルシー High dynamic range CMOS image sensor with anti-blooming characteristics and associated method
KR20140112874A (en) * 2013-03-14 2014-09-24 삼성전자주식회사 Electronic device with camera module
WO2014151093A1 (en) 2013-03-15 2014-09-25 Sionyx, Inc. Three dimensional imaging utilizing stacked imager devices and associated methods
WO2014209421A1 (en) 2013-06-29 2014-12-31 Sionyx, Inc. Shallow trench textured regions and associated methods
CN106575370A (en) * 2014-08-19 2017-04-19 英派尔科技开发有限公司 Machine recognizable pattern generation
FR3026223B1 (en) * 2014-09-22 2016-12-23 Photonis France APPARATUS FOR ACQUIRING PHOTOCATHODE BIMODE IMAGES.
US9727941B1 (en) 2014-11-19 2017-08-08 Digimarc Corporation Optimizing optical scanners for digital watermark detection
EP3166048A1 (en) * 2015-11-06 2017-05-10 ams AG Optical reader device, tag for use on a disposable or replaceable component, optical data validation system and method for optical data validation
WO2018048001A1 (en) * 2016-09-12 2018-03-15 한국생산기술연구원 Smart tag recognition system using infrared-sensitive code and recognition method using same
US10739189B2 (en) * 2018-08-09 2020-08-11 Ouster, Inc. Multispectral ranging/imaging sensor arrays and systems
US10760957B2 (en) 2018-08-09 2020-09-01 Ouster, Inc. Bulk optics for a scanning array
US11537832B2 (en) * 2018-11-12 2022-12-27 Hewlett-Packard Development Company, L.P. Multiple-pattern fiducial for heterogeneous imaging sensor systems
US10496911B1 (en) 2019-03-18 2019-12-03 Capital One Services, Llc Detection of images in relation to targets based on colorspace transformation techniques and utilizing ultraviolet and infrared light
US10496862B1 (en) 2019-03-18 2019-12-03 Capital One Services, Llc Detection of images in relation to targets based on colorspace transformation techniques and utilizing ultraviolet light
US10509991B1 (en) 2019-03-18 2019-12-17 Capital One Services, Llc Detection of images in relation to targets based on colorspace transformation techniques and utilizing infrared light
US10534948B1 (en) 2019-03-18 2020-01-14 Capital One Services, Llc Optimizing detection of images in relation to targets based on colorspace transformation techniques
US10523420B1 (en) 2019-04-18 2019-12-31 Capital One Services, Llc Transmitting encoded data along transmission mediums based on colorspace schemes
US10504013B1 (en) 2019-04-24 2019-12-10 Capital One Services, Llc Colorspace encoding multimedia data on a physical page
US10529300B1 (en) 2019-06-20 2020-01-07 Capital One Services, Llc Adaptive image display based on colorspace conversions
US11062104B2 (en) * 2019-07-08 2021-07-13 Zebra Technologies Corporation Object recognition system with invisible or nearly invisible lighting
US10614635B1 (en) 2019-07-25 2020-04-07 Capital One Services, Llc Augmented reality system with color-based fiducial marker
CN110632735A (en) * 2019-08-16 2019-12-31 俞庆平 Method for searching optimal focal plane in laser direct imaging system
US10833852B1 (en) 2019-10-03 2020-11-10 Capital One Services, Llc Encoded data along tape based on colorspace schemes
US10715183B1 (en) * 2019-10-25 2020-07-14 Capital One Services, Llc Data encoding with error-correcting code pursuant to colorspace schemes
US10867226B1 (en) 2019-11-04 2020-12-15 Capital One Services, Llc Programmable logic array and colorspace conversions
US10762371B1 (en) 2019-11-14 2020-09-01 Capital One Services, Llc Object detection techniques using colorspace conversions
US10878600B1 (en) 2019-12-10 2020-12-29 Capital One Services, Llc Augmented reality system with color-based fiducial marker utilizing local adaptive technology
US11302036B2 (en) 2020-08-19 2022-04-12 Capital One Services, Llc Color conversion between color spaces using reduced dimension embeddings

Citations (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3971065A (en) 1975-03-05 1976-07-20 Eastman Kodak Company Color imaging array
JPH06309486A (en) 1993-04-22 1994-11-04 Tokyo Electric Co Ltd Bar code reader
US5714745A (en) 1995-12-20 1998-02-03 Metanetics Corporation Portable data collection device with color imaging assembly
US5786582A (en) * 1992-02-27 1998-07-28 Symbol Technologies, Inc. Optical scanner for reading and decoding one- and two-dimensional symbologies at variable depths of field
US5804805A (en) 1986-08-08 1998-09-08 Norand Technology Corporation Hand-held optical indicia reader having a controlled oscillating system for optimal indicia reading
US6628330B1 (en) 1999-09-01 2003-09-30 Neomagic Corp. Color interpolator and horizontal/vertical edge enhancer using two line buffer and alternating even/odd filters for digital camera
US6642962B1 (en) 1999-09-01 2003-11-04 Neomagic Corp. Merged pipeline for color interpolation and edge enhancement of digital images
US6722569B2 (en) 2001-07-13 2004-04-20 Welch Allyn Data Collection, Inc. Optical reader having a color imager
US6765703B1 (en) 2000-09-27 2004-07-20 Kabushiki Kaisha Toshiba Method and apparatus for sensing image
US6832729B1 (en) 2001-03-23 2004-12-21 Zih Corp. Portable data collection device for reading fluorescent indicia
US20050011956A1 (en) * 2003-07-17 2005-01-20 Carlson Bradley S. System and method for reading and decoding optical codes using multiple color illumination
US20050056699A1 (en) 2001-07-13 2005-03-17 Timothy Meier Adaptive optical image reader
US6889904B2 (en) 2002-03-20 2005-05-10 Symbol Technologies, Inc. Image capture system and method using a common imaging array
US6976631B2 (en) 1999-05-12 2005-12-20 Tohken Co., Ltd. Code reader and code reading method for color image
US6976629B2 (en) 2002-03-20 2005-12-20 Symbol Technologies, Inc. Image capture system and method
US7014114B2 (en) 2003-10-02 2006-03-21 Symbol Technologies, Inc. Image capture device for and method of electro-optically reading indicia at low ambient light levels
US20060060653A1 (en) 2004-09-23 2006-03-23 Carl Wittenberg Scanner system and method for simultaneously acquiring data images from multiple object planes
US7071978B2 (en) 2001-07-18 2006-07-04 Hewlett-Packard Development Company, L.P. Image mosaic data reconstruction
US20060208083A1 (en) 2003-11-13 2006-09-21 Metrologic Instruments, Inc. Imaging-based bar code symbol reading system permitting modification of system features and functionalities without detailed knowledge about the hardware platform, communication interfaces, or user interfaces
US20060274171A1 (en) 2005-06-03 2006-12-07 Ynjiun Wang Digital picture taking optical reader having hybrid monochrome and color image sensor array
US7163149B2 (en) 2004-03-02 2007-01-16 Symbol Technologies, Inc. System and method for illuminating and reading optical codes imprinted or displayed on reflective surfaces
US7224540B2 (en) 2005-01-31 2007-05-29 Datalogic Scanning, Inc. Extended depth of field imaging system using chromatic aberration
US20080107354A1 (en) 2002-02-27 2008-05-08 Dowski Edward R Jr Optimized Image Processing For Wavefront Coded Imaging Systems
US7387252B2 (en) 2006-08-15 2008-06-17 Hand Held Products, Inc. Optical reader with improved lens focusing system
US20080169347A1 (en) 2007-01-11 2008-07-17 Datalogic Scanning, Inc. Methods and systems for optical code reading using virtual scan lines
US20090159685A1 (en) * 2007-12-20 2009-06-25 Symbol Technologies, Inc. Optimizing Optical Quality of a Sensor in a Bar Code Reader
US7594609B2 (en) * 2003-11-13 2009-09-29 Metrologic Instruments, Inc. Automatic digital video image capture and processing system supporting image-processing based code symbol reading during a pass-through mode of system operation at a retail point of sale (POS) station
US7695608B2 (en) * 2007-03-28 2010-04-13 I-Sens, Inc. Electrochemical biosensor and biosensor measuring device
US20100213259A1 (en) 2009-02-20 2010-08-26 Datalogic Scanning, Inc. Systems and methods of optical code reading using a color imager

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3054151B2 (en) * 1988-07-27 2000-06-19 キヤノン株式会社 Color image forming apparatus and color image processing method
JP2003309707A (en) * 2002-04-16 2003-10-31 Canon Inc Image reader
US6871545B2 (en) * 2003-06-20 2005-03-29 Weyerhaeuser Company System and method for measuring stiffness in standing trees
US7036735B2 (en) * 2004-01-21 2006-05-02 The Code Corporation Graphical code reader having illumination LEDs of different wavelengths
US7770799B2 (en) * 2005-06-03 2010-08-10 Hand Held Products, Inc. Optical reader having reduced specular reflection read failures

Patent Citations (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3971065A (en) 1975-03-05 1976-07-20 Eastman Kodak Company Color imaging array
US5804805A (en) 1986-08-08 1998-09-08 Norand Technology Corporation Hand-held optical indicia reader having a controlled oscillating system for optimal indicia reading
US5786582A (en) * 1992-02-27 1998-07-28 Symbol Technologies, Inc. Optical scanner for reading and decoding one- and two-dimensional symbologies at variable depths of field
JPH06309486A (en) 1993-04-22 1994-11-04 Tokyo Electric Co Ltd Bar code reader
US5714745A (en) 1995-12-20 1998-02-03 Metanetics Corporation Portable data collection device with color imaging assembly
US6976631B2 (en) 1999-05-12 2005-12-20 Tohken Co., Ltd. Code reader and code reading method for color image
US6642962B1 (en) 1999-09-01 2003-11-04 Neomagic Corp. Merged pipeline for color interpolation and edge enhancement of digital images
US6628330B1 (en) 1999-09-01 2003-09-30 Neomagic Corp. Color interpolator and horizontal/vertical edge enhancer using two line buffer and alternating even/odd filters for digital camera
US6765703B1 (en) 2000-09-27 2004-07-20 Kabushiki Kaisha Toshiba Method and apparatus for sensing image
US6832729B1 (en) 2001-03-23 2004-12-21 Zih Corp. Portable data collection device for reading fluorescent indicia
US6722569B2 (en) 2001-07-13 2004-04-20 Welch Allyn Data Collection, Inc. Optical reader having a color imager
US20050056699A1 (en) 2001-07-13 2005-03-17 Timothy Meier Adaptive optical image reader
US7071978B2 (en) 2001-07-18 2006-07-04 Hewlett-Packard Development Company, L.P. Image mosaic data reconstruction
US20080107354A1 (en) 2002-02-27 2008-05-08 Dowski Edward R Jr Optimized Image Processing For Wavefront Coded Imaging Systems
US6976629B2 (en) 2002-03-20 2005-12-20 Symbol Technologies, Inc. Image capture system and method
US6889904B2 (en) 2002-03-20 2005-05-10 Symbol Technologies, Inc. Image capture system and method using a common imaging array
US20050011956A1 (en) * 2003-07-17 2005-01-20 Carlson Bradley S. System and method for reading and decoding optical codes using multiple color illumination
US7014114B2 (en) 2003-10-02 2006-03-21 Symbol Technologies, Inc. Image capture device for and method of electro-optically reading indicia at low ambient light levels
US20060208083A1 (en) 2003-11-13 2006-09-21 Metrologic Instruments, Inc. Imaging-based bar code symbol reading system permitting modification of system features and functionalities without detailed knowledge about the hardware platform, communication interfaces, or user interfaces
US7594609B2 (en) * 2003-11-13 2009-09-29 Metrologic Instruments, Inc. Automatic digital video image capture and processing system supporting image-processing based code symbol reading during a pass-through mode of system operation at a retail point of sale (POS) station
US7163149B2 (en) 2004-03-02 2007-01-16 Symbol Technologies, Inc. System and method for illuminating and reading optical codes imprinted or displayed on reflective surfaces
US20060060653A1 (en) 2004-09-23 2006-03-23 Carl Wittenberg Scanner system and method for simultaneously acquiring data images from multiple object planes
US7224540B2 (en) 2005-01-31 2007-05-29 Datalogic Scanning, Inc. Extended depth of field imaging system using chromatic aberration
US7626769B2 (en) * 2005-01-31 2009-12-01 Datalogic Scanning, Inc. Extended depth of field imaging system using chromatic aberration
US20060274171A1 (en) 2005-06-03 2006-12-07 Ynjiun Wang Digital picture taking optical reader having hybrid monochrome and color image sensor array
US7387252B2 (en) 2006-08-15 2008-06-17 Hand Held Products, Inc. Optical reader with improved lens focusing system
US20080169347A1 (en) 2007-01-11 2008-07-17 Datalogic Scanning, Inc. Methods and systems for optical code reading using virtual scan lines
US7695608B2 (en) * 2007-03-28 2010-04-13 I-Sens, Inc. Electrochemical biosensor and biosensor measuring device
US20090159685A1 (en) * 2007-12-20 2009-06-25 Symbol Technologies, Inc. Optimizing Optical Quality of a Sensor in a Bar Code Reader
US20100213259A1 (en) 2009-02-20 2010-08-26 Datalogic Scanning, Inc. Systems and methods of optical code reading using a color imager

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Frédéric et al., Advances in Camera Phone Picture Quality, Photonics Spectra, Nov. 2007, 2 pp.
International Search Report & Written Opinion, International Application No. PCT/US2010/023918.
Tisse et al., Extended depth-of-field (EDoF) Using Sharpness Transport Across Colour Channels, Proceedings of Electronic Imaging vol. 7250, Jan. 19, 2009, 11 pp.

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9173508B2 (en) 2010-07-08 2015-11-03 Itab Scanflow Ab Checkout counter
US9301626B2 (en) 2010-07-08 2016-04-05 Itab Scanflow Ab Checkout counter
US8556179B2 (en) * 2011-08-16 2013-10-15 Intermec Ip Corp. Motion tolerant bar code reader
US9811705B1 (en) 2016-05-10 2017-11-07 Datalogic Ip Tech S.R.L. Illumination system with active element for generating different illumination patterns for a data reader
US10452951B2 (en) * 2016-08-26 2019-10-22 Goodrich Corporation Active visual attention models for computer vision tasks
US20180095307A1 (en) * 2016-10-03 2018-04-05 Xerox Corporation Hyperspectral imaging system
US10768497B2 (en) * 2016-10-03 2020-09-08 Xerox Corporation Hyperspectral imaging system

Also Published As

Publication number Publication date
EP2396744A2 (en) 2011-12-21
CN102317951B (en) 2015-01-21
WO2010093806A3 (en) 2010-12-02
US20100200658A1 (en) 2010-08-12
EP2396744B1 (en) 2016-06-01
WO2010093806A2 (en) 2010-08-19
EP2396744A4 (en) 2013-12-25
CN102317951A (en) 2012-01-11

Similar Documents

Publication Publication Date Title
US8118226B2 (en) High-resolution optical code imaging using a color imager
US8800874B2 (en) Systems and methods of optical code reading using a color imager
US8505823B2 (en) Noise removal from color barcode images
US9253393B2 (en) Two-dimensional imager with solid-state auto-focus
US8500023B2 (en) System and method for providing sufficient illumination quality for barcodes captured with a color image sensor
US9948868B2 (en) Multi-point spectral system and multi-point spectral module to be mounted on, clipped on or connected to an image capturing module of an electronic device easily
US20030024986A1 (en) Molded imager optical package and miniaturized linear sensor-based code reading engines
KR20170131526A (en) An imaging device having an image distribution unit for generating a spatially coded image
US7729061B2 (en) Imaging system having anamorphic magnification
US8675093B2 (en) Image capturing apparatus and image processing method
US20200007853A1 (en) Imaging systems with depth detection
US20100123009A1 (en) High-resolution interpolation for color-imager-based optical code readers
JP7350119B2 (en) Brightness and color correction of line camera image data
JP6361857B2 (en) Image reading apparatus and image reading program
CN107248242B (en) Watermark enabling scanning
US20210150172A1 (en) Code Reader and Method for Reading Optical Codes
Alvarez‐Cortes et al. Practical Low‐Cost Recovery of Spectral Power Distributions
EP3232367B1 (en) Imaging barcode reader with color separated aimer and illuminator
US20220358625A1 (en) Camera and method for acquiring image data
US20210123806A1 (en) Spectral imaging device and method
CA2757804C (en) System and method for barcode scanning with color image sensors

Legal Events

Date Code Title Description
AS Assignment

Owner name: DATALOGIC SCANNING, INC., OREGON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:OLMSTEAD, BRYAN L.;SHEARIN, ALAN;REEL/FRAME:024316/0099

Effective date: 20100429

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: DATALOGIC ADC, INC., OREGON

Free format text: MERGER;ASSIGNOR:DATALOGIC SCANNING, INC.;REEL/FRAME:033667/0932

Effective date: 20111228

FPAY Fee payment

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 12