WO2000051065A1 - Method and system for identifying a reference region on an image of a dropped-out form - Google Patents

Method and system for identifying a reference region on an image of a dropped-out form Download PDF

Info

Publication number
WO2000051065A1
WO2000051065A1 PCT/US2000/005222 US0005222W WO0051065A1 WO 2000051065 A1 WO2000051065 A1 WO 2000051065A1 US 0005222 W US0005222 W US 0005222W WO 0051065 A1 WO0051065 A1 WO 0051065A1
Authority
WO
WIPO (PCT)
Prior art keywords
data pattern
digital image
data
fit
expected
Prior art date
Application number
PCT/US2000/005222
Other languages
French (fr)
Inventor
Brent R. Smith
Original Assignee
Raf Technology, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Raf Technology, Inc. filed Critical Raf Technology, Inc.
Priority to AU37126/00A priority Critical patent/AU3712600A/en
Publication of WO2000051065A1 publication Critical patent/WO2000051065A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/14Image acquisition
    • G06V30/1444Selective acquisition, locating or processing of specific regions, e.g. highlighted text, fiducial marks or predetermined fields
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition

Definitions

  • This invention relates to the field of automated processing of drop-out forms for use with optical character recognition technology.
  • forms can be printed in drop-out ink (usually red or blue) that the scanner can filter out to leave only data in the scanned image.
  • identification region refers to an area in the digital image of a dropped-out form that corresponds to a pre-defined area on a template form.
  • template form refers to a digital image of the drop-out form that serves as a standardized reference against which later images may be compared.
  • pre-defined area on the template form, to which the identification regions correlates is referred to as a "reference region.
  • the data within the identification region may be used to identify the particular form from among a mixed stream of forms, provide a registration point for use in registering the image, or solve other types of problems encountered when processing drop-out forms that lack fixed landmarks.
  • the steps in this invention may be configured by the user to function on any form type or mixed stream of form types.
  • the first step in implementing this invention is to set up the template form.
  • the user locates and defines the boundaries of a region on the form in which the entered data ideally will have a distinctive and predictable pattern.
  • This region on the template form is the reference region, and the corresponding region on the scanned image is called the identification region.
  • the term "pattern" refers to the formation, shape, or structure represented by the data.
  • One example of a data pattern would be that found in an address section of a filled-out form that uses a standard address format.
  • the first horizontal line typically represents a name; the second line represents a street address: finally, there is a line for a city, state, and zip code.
  • the type of data pattern selected will depend on the particular form and the information requested in the form. For example, a medical insurance claim form may have data fields for patient name, insurance carrier, and policy number. The size, number, distribution, and position relative to one another of data fields such as these define a particular data pattern.
  • defined data pattern refers to the data pattern defined in a template form within the reference region
  • expected data pattern refers to the data pattern found within the identification region in the digital image of a form. The expected data pattern corresponds to the defined data pattern.
  • the term “distinctive” describes a data pattern that is dissimilar to other data patterns on the same form, thus reducing the probability of mistaking another data pattern for the expected data pattern. Also, as used in this specification and in the attached claims, the term “distinctive” describes a data pattern that is dissimilar to other data patterns on the same form, thus reducing the probability of mistaking another data pattern for the expected data pattern. Also, as used in this specification and in the attached claims, the term “distinctive” describes a data pattern that is dissimilar to other data patterns on the same form, thus reducing the probability of mistaking another data pattern for the expected data pattern. Also, as used in this specification and in the attached claims, the term “distinctive” describes a data pattern that is dissimilar to other data patterns on the same form, thus reducing the probability of mistaking another data pattern for the expected data pattern. Also, as used in this specification and in the attached claims, the term “distinctive” describes a data pattern that is dissimilar to other data patterns on the same form, thus reducing the probability of mistaking another
  • predictable describes a data pattern that is expected to be present on substantially all forms that are filled out and possess a fairly standard and constant structure. Because the ink comprising a drop-out form is filtered out during scanning, the data field must be filled out, or there will be nothing to use in identifying the form, locating registration points, or performing other form processing procedures.
  • the defined data pattern should also be unique to one particular type of form.
  • unique describes a data pattern that is at a particular location on only one type of form. Because someone using this invention selects the data pattern such that there is a one- to-one correspondence between the type of form and the particular location of the data pattern, verifying that the data pattern exists at that location verifies the identity of the form. If form identification is not required (for example, if only one type of form is being processed, or if distinguishing the type of form is not necessary), then the selected data pattern does not have to be unique.
  • the user divides the reference region into sub-regions where data (i.e., dark matter) is expected and sub-regions where no data (i.e., white space) is expected.
  • a sub-region is referred to as a "dark zone” if data is expected or a "white zone” if no data is expected.
  • the characteristics of the dark zones and white zones, including their size, number, distribution, and position relative to each other, essentially create the defined data pattern.
  • This defined data pattern acts as a fingerprint of the data expected to be in a form within the reference region. Applying this fingerprint to the identification region on the image of a form, and identifying entered data that matches this fingerprint within the identification region, identifies the location of the expected data pattern in that form. Identifying and knowing the location of the expected data pattern is what allows the form to be processed.
  • the identification region and the dark and white zones are used both to locate the expected data pattern and to calculate an associated confidence level.
  • One method of locating the expected data pattern is by generating a weighted pixel count on each of the dark zones and white zones within the identification region. For example, a pixel within a dark zone can have a weight of 1, and a pixel within a white zone can have a weight of -1. It is not necessary to use positive and negative values, but values should be selected so as to obtain a most favorable score when the data is in the dark zones where it is expected.
  • a score is calculated.
  • the invention uses a sequence of searching steps to maximize this scoring function over a search domain.
  • the search domain is defined by the enclosing boundaries of the identification region.
  • One example of a sequence of searching steps would be to select an initial location for the dark and white zones within the search domain, and then, while keeping the position of the dark and white zones fixed with respect to one another, translating the entire set of dark and white zones to a new location within the search domain.
  • the term "translate” means to move within the space represented by the search domain, without turning or rotating the item being moved. The movement during translation is relative, so translating the defined data pattern with respect to the digital image data achieves the same result as translating the digital image data with respect to the defined data pattern.
  • a weighted pixel count is calculated for each location of the set of dark and white zones until a maximum weighted pixel count has been identified. Once found, the location of the maximum weighted pixel count gives the location of the expected data pattern within the identification region. Knowing the location of the expected data pattern provides the location of a registration point, and hence a solution to the problem of drop-out form registration point location.
  • the value of the maximum weighted pixel count may be compared to a predetermined reference value to calculate a final confidence level.
  • a predetermined reference value include an average weighted pixel count, a minimal acceptable weighted pixel count, a maximum allowable weighted pixel count, or any other suitable reference value.
  • the final confidence level quantifies the degree to which the user is certain the expected data pattern was actually located. If the user chose the defined data pattern to be unique, the expected data pattern will only be at that particular location on one form, so a high final confidence level at that location will only be achieved for that particular form. Hence, the drop-out form identification problem is also solved. Additional objects and advantages of this invention will be apparent from the following detailed description of preferred embodiments thereof which proceeds with reference to the accompanying drawings.
  • FIG. 1 represents a typical page containing data in a pattern.
  • FIG. 2 represents a scanned image of the page of FIG. 1 , wherein the data pattern illustrates image transformation in the scanned image.
  • FIG. 3 illustrates a schematic overview of the overall system.
  • FIG. 4 is a flow diagram overview of the process of identifying an expected data pattern.
  • FIG. 5 depicts a reference region that encloses a portion of a template form containing data with a distinctive and predictable pattern.
  • FIG. 6 further illustrates the subdivision of the reference region of FIG. 5 into dark zones where data is expected to be present, and white zones where no data is expected to be present, in order to define a data pattern.
  • FIG. 7 represents an expected data pattern in a drop-out that fits the dark zones and white zones defining the defined data pattern of FIG. 6.
  • FIGS. 8A-8C depict a translation of the dark zones and white zones making up the defined data pattern to identify the location of the expected data pattern that achieves the best fit to the dark zones and white zones.
  • FIG. 9 represents the individual pixels comprising two characters of an expected data pattern to be found within a dark zone.
  • FIGS. 10A-10C represent a search procedure to achieve a best fit of the expected data pattern, quantified as a maximum weighted pixel count within a dark zone, and FIG. 10C represents the location of the dark zone that obtains the maximum weighted pixel count.
  • This invention relates to the field of optical character recognition on data in drop-out forms.
  • This invention allows for automated forms processing even when the processed image contains no permanent, fixed landmarks. Automated forms processing is possible because data that has been entered into a form is highly structured.
  • This invention allows the user to locate an expected data pattern within a scanned image and compare it to a defined data pattern on a template form. These comparisons are often made difficult by the fact that scanned images can experience rotation, stretching, offsetting, skewing, or other image transformations.
  • FIG. 1 and FIG. 2 represents a scanned data pattern 12.
  • the scanned data pattern 12 depicts data transformed from the original structure of the original data pattern 10. This invention uses the data itself in the scanned data pattern 12 and compares it to the original data pattern 10 to determine how to correct for any image transformations that occurred during scanning.
  • FIG. 3 depicts the overall system of the invention.
  • the user scans an original page 20 with a scanner 22.
  • the original page 20 has a form printed in drop-out ink (usually a red or blue color that the scanner 22 can filter out).
  • the drop-out form on the original page 20 also has entered data.
  • the scanner 22 produces a scanned image 24 that has image transformations with respect to the original page 20.
  • the scanned image 24 can then be processed by a computer and computer program 26 that allow automated forms processing.
  • the computer and computer program 26 have access to a database 28, and the database 28 contains plurality of template forms 30.
  • the computer and computer program 26 compare the scanned image 24 to each of the template forms 30, one at a time, to determine which template form 30 the scanned image 24 best fits.
  • FIG. 3 represents an overall schematic of a preferred embodiment of the system making up the invention. Other elements may be added without departing from the nature or purpose of the invention. Similarly, some elements may not be necessary to practice this invention.
  • FIG. 4 depicts a flow chart showing the process of matching data in a scanned image to a particular template. The flow chart begins with a scanning step 40 in which a page is scanned to create a digital image bearing data. At this point, the drop-out form has been filtered out and only the data remains.
  • the next step is the measuring step 42 in which the digital image data is measured to determine how well it fits an initial template form.
  • a comparison step 44 in which the fit of the digital image to the template form is compared to a reference fit value.
  • the comparison step is achieved by allowing the user to predetermine a reference fit value based on the data expected.
  • a user can define a minimal reference fit value below which it can be said that the scanned image does not correspond to the template being tested.
  • other types of reference fit values could also be used. For example, a maximum reference fit value could be used, or an average reference fit value could be used. Any reference fit value is acceptable so long as it enables the comparison step 44 to be conducted.
  • the comparison step 44 if the comparisons show that the digital image data matches the template being tested, then an end determination 46 is made that the expected data pattern had been identified. If the comparison step 44 yields an unfavorable result, and the digital image data being tested did not match the template, and the invention proceeds to a iterative step 48, in which a new template form is selected and the measuring step 42 and comparison step 44 are once again conducted. The iterative step 48 continues until the end determination 46 is achieved.
  • This invention operates by first setting up a template form to serve as a reference against which scanned images are later compared.
  • the template form is obtained as a digital image with minimal scanning noise or distortion.
  • the template form already has the drop-out form filtered out, but it contains representative data to allow the user to define a data pattern to compare against scanned images.
  • a preferred embodiment of this invention also uses Group III Compressed Tiff files for the images. Other image file formats could also be used.
  • a preferred embodiment implements Document Attribute Format Specification as a preferred standard for optical character recognition and document image decomposition data.
  • the user first locates a region of the template form in which the data will have a distinctive and predictable pattern.
  • an address section of the form would represent such a distinct and predictable pattern.
  • Another example would include information entered in a medical form, such as an insurance carrier, a policy member, and a policy expiration date.
  • This invention allows the user to select the data that would be used as a reference in the user-defined data pattern.
  • the user defines a boundary representing a reference region 50 around a defined data pattern 52.
  • the defined data pattern 52 encompasses generic data that represents entered data in a filled-out form.
  • the term "fille- out form" refers to a form in which some data has been entered. A form does not have to have all possible data fields completed to be a "filled-out form. "
  • the user divides the reference region 50 into subregions. These subregions are called dark zones 54, if they are expected to include data, and white zones 56, if they are expected to include white space on a filled-out form.
  • FIG 6 illustrates the dark zones 54 and white zones 56 inside the reference region 50. Together these dark zones 54 and white zones 56 make up the defined data pattern 52. This defined data pattern 52 is inside the reference region 50 and the reference region 50 is in the digital template form.
  • this invention delineates an identification region 60 in the scanned image of a page bearing a drop-out form that has been completed with data.
  • the identification region 60 of FIG 7 corresponds to the reference region 50 of FIG 5 and 6.
  • the dark zones 54 and white zones 56 which make up the defined data pattern 52 of FIG 5 and 6, are applied to the identification zone 60 to locate an expected data pattern 62.
  • FIG 7 represents the situation in which the digital image data of the expected data pattern 62 fits inside the dark zones 54 and the white zones 56 enclose only white space. This is the desired end result.
  • the invention must often search within the identification region 60 to find the best fit of the expected data pattern 62 to the dark zones 54 and white zones 56.
  • FIG. 8A through 8C represent a search of the identification region 60 to find the best fit of the expected data pattern 62 to the dark zones 54 and white zones 56.
  • FIG. 8A depicts a situation in which none of the data in the expected data pattern 62 is within a dark zones 54. Conversely, the data is partially present in the white zones 56.
  • FIG. 8B the dark zones 54 and white zones 56 have be translated, as a solid unit, vertically toward the top of the identification region 60. In this new location, approximately half of the data comprising the expected data pattern 62 is located in the dark zones 54 and none of the data is located in the white zones 56. Accordingly, FIG. 8B represents a better fit of the expected data pattern 62 to the dark zones 54 and white zones 56 than FIG. 8A.
  • FIG. 8C depicts the best fit of the expected data pattern 62 to the dark zones 54 and white zones 56. All of the data in the expected data pattern 62 is located inside the dark zones 54 where it is to be expected.
  • a preferred embodiment of this invention uses a weighted pixel count.
  • FIG. 9 and FIGS. 10A through 10C depict a simplified version of a weighted pixel count fit being determined.
  • FIG. 9 represents the expected data pattern 62 comprising two letters totaling 80 individual pixels. In actuality, the pixels are contiguous; however, in FIG. 9, the pixels have been illustrated with spaces between them for conceptual clarity.
  • FIG. 10A the initial location of the dark zone 54 encloses 30 pixels of the expected data pattern 62.
  • a weighted pixel count can be determined by assigning a first value to pixels located in a dark zone, and a second value to pixels located in a white zone.
  • a preferred embodiment gives a score of plus one to pixels in a dark zone and negative one to pixels in a white zone.
  • each pixel inside the dark zone 54 obtains a score of positive one. Accordingly, in FIG. 10A, the 30 pixels inside the dark zone 54 obtain a score of 30. In FIG. 10B, 65 pixels are inside the dark zone 54 at the second location, and a score of 65 is obtained. Finally, in FIG. 10C all 80 pixels of the expected data pattern 62 are inside the dark zone 54 at the third location. Because a score of 80 represents the maximum score of all the locations tested, FIG. 10C represents the location of the expected data pattern 62. Again with reference to FIG. 7. once the expected data pattern 62 has been located by determining the best fit of the digital image data within the dark zones 54 and white zones 56.
  • the invention uses the location of the expected data pattern 62 to solve the problems faced in automated processing of drop-out forms. Although there are many problems facing automated processing of drop off forms, this description of a preferred embodiment of the invention will address two in particular. These two are the drop-out form identification problem, and the drop-out form registration point location problem.
  • the location of the expected data pattern 62 is known, the user also has a registration point 70 available. In FIG. 7, the registration point 70 is defined as the top left corner of the top dark zone 56. The user can then measure the distance between registration point 70 and a fixed point such as the corner point 72 of the identification region 60.
  • the corner point 72 on the identification region 60 corresponds to the corner point 76 on the reference region 50.
  • a user can then measure the distance between the corner point 76 on the reference region 50, and the original registration point location 74 in the reference region 50.
  • the user can determine the degree to which the image underwent transformation during the scanning process. Hence, the problem of drop-out form registration point location has been solved.
  • a final confidence level can be calculated. This final confidence level quantifies the degree to which a user can be certain that the expected data pattern 62 was truly located. In a preferred embodiment of this invention, the user predefines a minimum value, below which it can be said there is zero confidence the expected data pattern 62 was found.
  • a maximum value can be set. or an expected average value could be used.
  • a preferred embodiment of this invention also uses a maximum expected confidence value above which it can be said that there is complete confidence the expected data pattern 62 was found.
  • the maximum and minimum values define the slope of a line. Confidence values in between the minimum value and maximum value can be interpreted by linear interpolation (a process well know to those skilled in the art). If the defined data pattern 52 was defined to be unique to one particular type of form, then a high confidence level at that location of the expected data pattern can only occur for one type of form. Accordingly, if a high confidence level is achieved, the form has been identified. This provides a solution to the drop-out form identification problem.
  • a preferred embodiment of this invention uses some specific optimization procedures to search for the best fit of the digital image data to the defined data pattern. While an exhaustive search of all possible translations of the defined data pattern would have the benefit of guaranteeing that the best fit has been found, it would operate far too slow to be an effective procedure. Accordingly, some straight-forward, but effective, optimizations can be made.
  • One important feature of the digital image data is its general size or spatial frequency.
  • a value of s is desired that minimizes N. Calculus determines that a minimum number N occurs when:
  • a preferred embodiment of this invention can implement further optimizations with respect to the determination of a confidence level.
  • One preferred range of confidence values is between 0 (no confidence) and 255 (highest confidence).
  • the user defines the following values: a “minimum” score, a “plenty” score, a “mini pixel fraction,” and a “plenty pixel fraction. " If “M” is the maximum score, and “D” is the number of black pixels in the identification region, then the confidence level, "C, " is defined as follows:
  • R (M/D - minimum pixel fraction)/(plenty pixel fraction - minimum pixel fraction) otherwise.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Character Input (AREA)

Abstract

The overall system of the invention. The user scans an original page (20) with a scanner (22). Usually a red or blue color that the scanner (22) can filter out. The scanner (22) produces a scanned image (24). The scanned image (24) can then be processed by a computer and computer program (26) that allow automated forms processing. And the data base (28) contains plurality of template forms (30).

Description

METHOD AND SYSTEM FOR IDENTIFYING A REFERENCE REGION
ON AN IMAGE OF A DROPPED-OUT FORM
Technical Field This invention relates to the field of automated processing of drop-out forms for use with optical character recognition technology.
Background of the Invention
Automated forms processing is possible because data that has been entered into a form is highly structured. The physical location of data and the structure of data, once located, are well specified. As used throughout this specification and the attached claims, the term "data" refers to any information entered into a form, whether numbers, text, initials, logos, shaded areas, or any other sort of marking entered into the form. Unfortunately, scanned images can be rotated, stretched, offset, or skewed.
Thus, in order to successfully read a form, software must be able to correct for any of these image transformations. By finding landmarks on the image and comparing them with the expected locations of these landmarks as exemplified by a template form, the mapping from image coordinates to template form coordinates can be determined. These landmarks are called registration points and the process of finding the image- to-template coordinate transformation is called registration. One of the difficulties of using optical character recognition (OCR) technology to automatically read data on a form is that the form itself will often occlude the data. This happens when someone filling out the forms does not properly provide the desired data within the boundaries provided on the form. If a box on a form is intended to hold data, but the person filling out the form writes too large to fit inside the box, then the lines of the box itself will strike through or obscure a portion of the data. The same results occur if data is being typed or printed into a form and the data does not fall cleanly within the boundaries of the box.
To avoid occluding data, forms can be printed in drop-out ink (usually red or blue) that the scanner can filter out to leave only data in the scanned image.
However, when a form is printed entirely of drop-out ink. all of the known landmarks are lost at the time of scanning. This leads to what is known as the drop-out form registration point location problem, which refers to the difficulties inherent in locating registration points in the absence of fixed landmarks on the form. In the case of a mixed stream of image types, the processing system must identify the particular form with which each image is associated. For standard (nondrop-out) forms, this is a relatively simple task because the form is included in each image, and the form will contain landmarks to identify the form type. However, in the case of drop-out forms, where the original form is filtered out of the digital image, the lack of known landmarks makes the problem of form identification vastly more complicated. This is what is referred to as the drop-out form identification problem.
The drop-out form identification and registration point selection problems are only two common examples of the problems encountered when processing forms that provide no fixed landmarks. Another problem is encountered when a form is being processed only to perform OCR on one particular type of data entry. If that particular data entry cannot be located, processing becomes impossible. Existing OCR systems do not provide a convenient, reliable, or efficient automated process to solve any of these problems. Solving these and other problems associated with automated drop-out form processing is the subject of the present invention. Summary of the Invention
This invention uses the patterns and structure of the actual data entered into the form to provide an identification region for use in processing the form. As used in this specification and the attached claims, the phrase "identification region" refers to an area in the digital image of a dropped-out form that corresponds to a pre-defined area on a template form. As used in this specification and the attached claims, the phrase "template form" refers to a digital image of the drop-out form that serves as a standardized reference against which later images may be compared. Also, as used in this specification and the attached claims, the pre-defined area on the template form, to which the identification regions correlates, is referred to as a "reference region. "
Once identified, the data within the identification region may be used to identify the particular form from among a mixed stream of forms, provide a registration point for use in registering the image, or solve other types of problems encountered when processing drop-out forms that lack fixed landmarks. The steps in this invention may be configured by the user to function on any form type or mixed stream of form types.
The first step in implementing this invention is to set up the template form. During the set-up phase, the user locates and defines the boundaries of a region on the form in which the entered data ideally will have a distinctive and predictable pattern.
This region on the template form is the reference region, and the corresponding region on the scanned image is called the identification region.
As used in this specification and the attached claims, the term "pattern" refers to the formation, shape, or structure represented by the data. One example of a data pattern would be that found in an address section of a filled-out form that uses a standard address format. The first horizontal line typically represents a name; the second line represents a street address: finally, there is a line for a city, state, and zip code. The type of data pattern selected will depend on the particular form and the information requested in the form. For example, a medical insurance claim form may have data fields for patient name, insurance carrier, and policy number. The size, number, distribution, and position relative to one another of data fields such as these define a particular data pattern. As used throughout this specification and in the attached claims, the term "defined data pattern" refers to the data pattern defined in a template form within the reference region, and the term "expected data pattern" refers to the data pattern found within the identification region in the digital image of a form. The expected data pattern corresponds to the defined data pattern.
As used in this specification and the attached claims, the term "distinctive" describes a data pattern that is dissimilar to other data patterns on the same form, thus reducing the probability of mistaking another data pattern for the expected data pattern. Also, as used in this specification and in the attached claims, the term
"predictable" describes a data pattern that is expected to be present on substantially all forms that are filled out and possess a fairly standard and constant structure. Because the ink comprising a drop-out form is filtered out during scanning, the data field must be filled out, or there will be nothing to use in identifying the form, locating registration points, or performing other form processing procedures.
If part of the automated drop-out form processing requires identification of the form, then the defined data pattern should also be unique to one particular type of form. As used in this specification and in the attached claims, the term "unique" describes a data pattern that is at a particular location on only one type of form. Because someone using this invention selects the data pattern such that there is a one- to-one correspondence between the type of form and the particular location of the data pattern, verifying that the data pattern exists at that location verifies the identity of the form. If form identification is not required (for example, if only one type of form is being processed, or if distinguishing the type of form is not necessary), then the selected data pattern does not have to be unique.
To delineate the defined data pattern, the user divides the reference region into sub-regions where data (i.e., dark matter) is expected and sub-regions where no data (i.e., white space) is expected. A sub-region is referred to as a "dark zone" if data is expected or a "white zone" if no data is expected. The characteristics of the dark zones and white zones, including their size, number, distribution, and position relative to each other, essentially create the defined data pattern. This defined data pattern acts as a fingerprint of the data expected to be in a form within the reference region. Applying this fingerprint to the identification region on the image of a form, and identifying entered data that matches this fingerprint within the identification region, identifies the location of the expected data pattern in that form. Identifying and knowing the location of the expected data pattern is what allows the form to be processed.
To implement this invention, the identification region and the dark and white zones are used both to locate the expected data pattern and to calculate an associated confidence level. One method of locating the expected data pattern is by generating a weighted pixel count on each of the dark zones and white zones within the identification region. For example, a pixel within a dark zone can have a weight of 1, and a pixel within a white zone can have a weight of -1. It is not necessary to use positive and negative values, but values should be selected so as to obtain a most favorable score when the data is in the dark zones where it is expected.
For a particular location of the dark zones and white zones, a score is calculated. The invention uses a sequence of searching steps to maximize this scoring function over a search domain. The search domain is defined by the enclosing boundaries of the identification region. One example of a sequence of searching steps would be to select an initial location for the dark and white zones within the search domain, and then, while keeping the position of the dark and white zones fixed with respect to one another, translating the entire set of dark and white zones to a new location within the search domain. As used in this specification and the attached claims, the term "translate" means to move within the space represented by the search domain, without turning or rotating the item being moved. The movement during translation is relative, so translating the defined data pattern with respect to the digital image data achieves the same result as translating the digital image data with respect to the defined data pattern.
A weighted pixel count is calculated for each location of the set of dark and white zones until a maximum weighted pixel count has been identified. Once found, the location of the maximum weighted pixel count gives the location of the expected data pattern within the identification region. Knowing the location of the expected data pattern provides the location of a registration point, and hence a solution to the problem of drop-out form registration point location. The value of the maximum weighted pixel count may be compared to a predetermined reference value to calculate a final confidence level. Some examples of a predetermined reference value include an average weighted pixel count, a minimal acceptable weighted pixel count, a maximum allowable weighted pixel count, or any other suitable reference value. The final confidence level quantifies the degree to which the user is certain the expected data pattern was actually located. If the user chose the defined data pattern to be unique, the expected data pattern will only be at that particular location on one form, so a high final confidence level at that location will only be achieved for that particular form. Hence, the drop-out form identification problem is also solved. Additional objects and advantages of this invention will be apparent from the following detailed description of preferred embodiments thereof which proceeds with reference to the accompanying drawings.
Brief Description of the Drawings FIG. 1 represents a typical page containing data in a pattern.
FIG. 2 represents a scanned image of the page of FIG. 1 , wherein the data pattern illustrates image transformation in the scanned image.
FIG. 3 illustrates a schematic overview of the overall system.
FIG. 4 is a flow diagram overview of the process of identifying an expected data pattern. FIG. 5 depicts a reference region that encloses a portion of a template form containing data with a distinctive and predictable pattern.
FIG. 6 further illustrates the subdivision of the reference region of FIG. 5 into dark zones where data is expected to be present, and white zones where no data is expected to be present, in order to define a data pattern.
FIG. 7 represents an expected data pattern in a drop-out that fits the dark zones and white zones defining the defined data pattern of FIG. 6.
FIGS. 8A-8C depict a translation of the dark zones and white zones making up the defined data pattern to identify the location of the expected data pattern that achieves the best fit to the dark zones and white zones.
FIG. 9 represents the individual pixels comprising two characters of an expected data pattern to be found within a dark zone.
FIGS. 10A-10C represent a search procedure to achieve a best fit of the expected data pattern, quantified as a maximum weighted pixel count within a dark zone, and FIG. 10C represents the location of the dark zone that obtains the maximum weighted pixel count.
Detailed Description of a Preferred Embodiment
This invention relates to the field of optical character recognition on data in drop-out forms. This invention allows for automated forms processing even when the processed image contains no permanent, fixed landmarks. Automated forms processing is possible because data that has been entered into a form is highly structured. This invention allows the user to locate an expected data pattern within a scanned image and compare it to a defined data pattern on a template form. These comparisons are often made difficult by the fact that scanned images can experience rotation, stretching, offsetting, skewing, or other image transformations. With reference to FIG. 1 and FIG. 2, FIG. 1 represents an original data pattern 10 and FIG. 2 represents a scanned data pattern 12. The scanned data pattern 12 depicts data transformed from the original structure of the original data pattern 10. This invention uses the data itself in the scanned data pattern 12 and compares it to the original data pattern 10 to determine how to correct for any image transformations that occurred during scanning.
FIG. 3 depicts the overall system of the invention. With reference to FIG. 3, the user scans an original page 20 with a scanner 22. The original page 20 has a form printed in drop-out ink (usually a red or blue color that the scanner 22 can filter out). The drop-out form on the original page 20 also has entered data. The scanner 22 produces a scanned image 24 that has image transformations with respect to the original page 20. The scanned image 24 can then be processed by a computer and computer program 26 that allow automated forms processing. The computer and computer program 26 have access to a database 28, and the database 28 contains plurality of template forms 30. The computer and computer program 26 compare the scanned image 24 to each of the template forms 30, one at a time, to determine which template form 30 the scanned image 24 best fits. This is an example of drop-out forms processing to allow identification of a scanned image 24. Drop-out form identification is only one potential use for this invention, as the invention applies to a much broader spectrum of drop-out form processing problems. FIG. 3 represents an overall schematic of a preferred embodiment of the system making up the invention. Other elements may be added without departing from the nature or purpose of the invention. Similarly, some elements may not be necessary to practice this invention. FIG. 4 depicts a flow chart showing the process of matching data in a scanned image to a particular template. The flow chart begins with a scanning step 40 in which a page is scanned to create a digital image bearing data. At this point, the drop-out form has been filtered out and only the data remains. The next step is the measuring step 42 in which the digital image data is measured to determine how well it fits an initial template form. Next is a comparison step 44 in which the fit of the digital image to the template form is compared to a reference fit value. In a preferred embodiment of this invention, the comparison step is achieved by allowing the user to predetermine a reference fit value based on the data expected. A user can define a minimal reference fit value below which it can be said that the scanned image does not correspond to the template being tested. However, other types of reference fit values could also be used. For example, a maximum reference fit value could be used, or an average reference fit value could be used. Any reference fit value is acceptable so long as it enables the comparison step 44 to be conducted. Once the comparison step 44 is conducted, if the comparisons show that the digital image data matches the template being tested, then an end determination 46 is made that the expected data pattern had been identified. If the comparison step 44 yields an unfavorable result, and the digital image data being tested did not match the template, and the invention proceeds to a iterative step 48, in which a new template form is selected and the measuring step 42 and comparison step 44 are once again conducted. The iterative step 48 continues until the end determination 46 is achieved.
This invention operates by first setting up a template form to serve as a reference against which scanned images are later compared. In a preferred embodiment, the template form is obtained as a digital image with minimal scanning noise or distortion. In a preferred embodiment, the template form already has the drop-out form filtered out, but it contains representative data to allow the user to define a data pattern to compare against scanned images. A preferred embodiment of this invention also uses Group III Compressed Tiff files for the images. Other image file formats could also be used. Also, a preferred embodiment implements Document Attribute Format Specification as a preferred standard for optical character recognition and document image decomposition data.
During the set up phase, the user first locates a region of the template form in which the data will have a distinctive and predictable pattern. In a preferred embodiment, an address section of the form would represent such a distinct and predictable pattern. However, other patterns could also be used. Another example would include information entered in a medical form, such as an insurance carrier, a policy member, and a policy expiration date. This invention allows the user to select the data that would be used as a reference in the user-defined data pattern. With reference to FIG 5. the user defines a boundary representing a reference region 50 around a defined data pattern 52. In this illustration of a preferred embodiment, the defined data pattern 52 encompasses generic data that represents entered data in a filled-out form. As used in this specification and the attached claims, the term "fille- out form" refers to a form in which some data has been entered. A form does not have to have all possible data fields completed to be a "filled-out form. "
In the next step, the user divides the reference region 50 into subregions. These subregions are called dark zones 54, if they are expected to include data, and white zones 56, if they are expected to include white space on a filled-out form. FIG 6 illustrates the dark zones 54 and white zones 56 inside the reference region 50. Together these dark zones 54 and white zones 56 make up the defined data pattern 52. This defined data pattern 52 is inside the reference region 50 and the reference region 50 is in the digital template form.
With reference to FIG 7 this invention delineates an identification region 60 in the scanned image of a page bearing a drop-out form that has been completed with data. The identification region 60 of FIG 7 corresponds to the reference region 50 of FIG 5 and 6. Again with reference to FIG 7. the dark zones 54 and white zones 56, which make up the defined data pattern 52 of FIG 5 and 6, are applied to the identification zone 60 to locate an expected data pattern 62. FIG 7 represents the situation in which the digital image data of the expected data pattern 62 fits inside the dark zones 54 and the white zones 56 enclose only white space. This is the desired end result. However to achieve this result, the invention must often search within the identification region 60 to find the best fit of the expected data pattern 62 to the dark zones 54 and white zones 56.
FIG. 8A through 8C represent a search of the identification region 60 to find the best fit of the expected data pattern 62 to the dark zones 54 and white zones 56. FIG. 8A depicts a situation in which none of the data in the expected data pattern 62 is within a dark zones 54. Conversely, the data is partially present in the white zones 56. In FIG. 8B the dark zones 54 and white zones 56 have be translated, as a solid unit, vertically toward the top of the identification region 60. In this new location, approximately half of the data comprising the expected data pattern 62 is located in the dark zones 54 and none of the data is located in the white zones 56. Accordingly, FIG. 8B represents a better fit of the expected data pattern 62 to the dark zones 54 and white zones 56 than FIG. 8A. FIG. 8C depicts the best fit of the expected data pattern 62 to the dark zones 54 and white zones 56. All of the data in the expected data pattern 62 is located inside the dark zones 54 where it is to be expected.
Although there are various ways to determine the fit of the expected data pattern 62. a preferred embodiment of this invention uses a weighted pixel count.
FIG. 9 and FIGS. 10A through 10C depict a simplified version of a weighted pixel count fit being determined. FIG. 9 represents the expected data pattern 62 comprising two letters totaling 80 individual pixels. In actuality, the pixels are contiguous; however, in FIG. 9, the pixels have been illustrated with spaces between them for conceptual clarity. In FIG. 10A, the initial location of the dark zone 54 encloses 30 pixels of the expected data pattern 62. A weighted pixel count can be determined by assigning a first value to pixels located in a dark zone, and a second value to pixels located in a white zone. A preferred embodiment gives a score of plus one to pixels in a dark zone and negative one to pixels in a white zone. Although FIGS. 10A through 10C do not depict a white zone 56, each pixel inside the dark zone 54 obtains a score of positive one. Accordingly, in FIG. 10A, the 30 pixels inside the dark zone 54 obtain a score of 30. In FIG. 10B, 65 pixels are inside the dark zone 54 at the second location, and a score of 65 is obtained. Finally, in FIG. 10C all 80 pixels of the expected data pattern 62 are inside the dark zone 54 at the third location. Because a score of 80 represents the maximum score of all the locations tested, FIG. 10C represents the location of the expected data pattern 62. Again with reference to FIG. 7. once the expected data pattern 62 has been located by determining the best fit of the digital image data within the dark zones 54 and white zones 56. The invention uses the location of the expected data pattern 62 to solve the problems faced in automated processing of drop-out forms. Although there are many problems facing automated processing of drop off forms, this description of a preferred embodiment of the invention will address two in particular. These two are the drop-out form identification problem, and the drop-out form registration point location problem. Once the location of the expected data pattern 62 is known, the user also has a registration point 70 available. In FIG. 7, the registration point 70 is defined as the top left corner of the top dark zone 56. The user can then measure the distance between registration point 70 and a fixed point such as the corner point 72 of the identification region 60. With reference to FIGS. 6 and 7, because the identification region 60 corresponds to the reference region 50, the corner point 72 on the identification region 60 corresponds to the corner point 76 on the reference region 50. A user can then measure the distance between the corner point 76 on the reference region 50, and the original registration point location 74 in the reference region 50. By comparing the difference in distance between the corner point 76 in the reference region 50 and the original registration point location 74 in the reference region 50 to the difference in distance between the corner point 72 in the identification region 60 and the registration point 70 in the identification region 60, the user can determine the degree to which the image underwent transformation during the scanning process. Hence, the problem of drop-out form registration point location has been solved. Using the weighted pixel average calculated by determining the best fit of the expected data pattern 62 to the dark zones 54 and white zones 56, a final confidence level can be calculated. This final confidence level quantifies the degree to which a user can be certain that the expected data pattern 62 was truly located. In a preferred embodiment of this invention, the user predefines a minimum value, below which it can be said there is zero confidence the expected data pattern 62 was found.
Alternatively, a maximum value can be set. or an expected average value could be used. In addition to the minimum confidence value, a preferred embodiment of this invention also uses a maximum expected confidence value above which it can be said that there is complete confidence the expected data pattern 62 was found. The maximum and minimum values define the slope of a line. Confidence values in between the minimum value and maximum value can be interpreted by linear interpolation (a process well know to those skilled in the art). If the defined data pattern 52 was defined to be unique to one particular type of form, then a high confidence level at that location of the expected data pattern can only occur for one type of form. Accordingly, if a high confidence level is achieved, the form has been identified. This provides a solution to the drop-out form identification problem. In order to allow automated drop-out forms processing to occur in a more efficient and effective manner, a preferred embodiment of this invention uses some specific optimization procedures to search for the best fit of the digital image data to the defined data pattern. While an exhaustive search of all possible translations of the defined data pattern would have the benefit of guaranteeing that the best fit has been found, it would operate far too slow to be an effective procedure. Accordingly, some straight-forward, but effective, optimizations can be made. One important feature of the digital image data is its general size or spatial frequency. Borrowing concepts from the "Nyquist Sampling Theorem," one could translate the translating entity in increments of one half the spatial frequency, and then fine-tune the placement of the entity by translations amounting to an exhaustive search only in the neighborhood of the best point on the "Nyquist Grid. " The user configuring the search area will know what type of data to expect and can set properties that specify the horizontal and vertical spatial frequencies. When the search procedure is operating, the invention does an initial search over the grid of points whose separation is determined by these properties. Then, an exhaustive search will be performed in the four quadrants bordering the best grid point.
Given this method, an optimal grid spacing can be computed based on the size of the search area. To simplify the computations, it can be assumed that the horizontal and vertical spatial frequencies are equal. If the area being searched has a width "W, " a height "H, " and a grid spacing "s, " then the number "N" of positions the search procedure will try, using the method just described, is approximated by the following equation: W H
N = — * — + 4s2
S S
A value of s is desired that minimizes N. Calculus determines that a minimum number N occurs when:
Figure imgf000016_0001
This value of s should be used only if it is smaller than one half the spatial frequency of the expected data pattern. In practice, this a priori optimal grid size is often acceptable. This results in a search procedure that operates in a time quantity:
Figure imgf000016_0002
In the case of registering data, it is important to note that lines of text have a very strong horizontal nature. This affords another extreme optimization. Due to the strong horizontal nature, virtually any vertical path in the search area will result in a best y-value that is very close to the actual optimal y-position. A subsequent horizontal sweep at that y-value will result in a point (x.y) that is very close to the optimal pint. Finally, an exhaustive search in the neighborhood of this point can be performed to find the optimal registration point location. In a preferred embodiment, each of these sweeps can use a 1 -dimensional grid technique analogous to the 2- dimensional grid technique described above. This results in a search procedure that operates in a time quantity:
Figure imgf000016_0003
A preferred embodiment of this invention can implement further optimizations with respect to the determination of a confidence level. One preferred range of confidence values is between 0 (no confidence) and 255 (highest confidence). Once the search procedure has located a best fit, which occurs when the defined digital image data falls within the dark zones of the defined data pattern, it must then convert the fit to a confidence value between 0 and 255. Due to the wide variety of applications of this procedure, there is no a priori way of associating a confidence value with a best fit. such as that determined by a raw pixel count. Thus, some user- defined properties are used. The confidence level is a function of two things: 1) the maximum pixel count, and 2) the ratio of the maximum score to the total number of black pixels contained in the identification region. The confidence level measures not only a raw score, but also the percentage of pixels in the identification region that are present in the desired location.
The user defines the following values: a "minimum" score, a "plenty" score, a "mini pixel fraction," and a "plenty pixel fraction. " If "M" is the maximum score, and "D" is the number of black pixels in the identification region, then the confidence level, "C, " is defined as follows:
= 255*S* R where
S= 0 if M < or = the minimum score; S = 1 if M > or = the plenty score: and S = (M - minimum score)/(plenty score - minimum score) otherwise and
R = 0 if M/D < or = minimum pixel fraction; R = 1 if M/D > or = plenty pixel fraction; and
R = (M/D - minimum pixel fraction)/(plenty pixel fraction - minimum pixel fraction) otherwise.
Thus, S grows linearly with M from a value of 0 when M is the minimum score, to a value of 1 when M is the plenty score. Similarly. R grows linearly with M/D from a value of 0 when the ratio is the minimum pixel fraction, to a value of 1 when the ratio is the plenty pixel fraction It will be obvious to those having skill in the art that many changes may be made to the details of the above-described embodiment of this invention without departing from the underlying principles thereof. The scope of the present invention should, therefore, be determined only by the following claims.

Claims

Claims
1. A method of identifying an expected data pattern for use in facilitating automated drop-out forms processing, the expected data pattern being in a digital image of a page, the digital image being created by scanning the page, and the page bearing data arranged in a predetermined configuration generally conforming to a dropped-out form represented by a digital template form, the method comprising the steps of: obtaining the digital template form; defining, in the digital template form, a reference region having a boundary; identifying, within the reference region boundary, first areas where data is expected to be present, and second areas where no data is expected to be present, so as to define a defined data pattern; obtaining the digital image of the page bearing data; comparing the defined data pattern to the digital image data at an initial location within an identification region in the digital image, the identification region corresponding to the reference region in the template form; determining a fit of the defined data pattern to the digital image data at the initial location; and repeatedly translating the defined data pattern relative to the digital image data, and redetermining the fit, until a location is identified that provides a best fit of the defined data pattern to the digital image data.
2. The method of claim 1, wherein the step of repeatedly translating the defined data pattern translates the defined data pattern within the identification region.
3. The method of claim 2, wherein the step of repeatedly translating the defined data pattern includes: dividing the identification region into a grid having predefined spacing; and translating the defined data pattern to each grid point in the grid.
4. The method of claim 1 , wherein the step of determining a fit includes: counting digital image data pixels within the first areas and the second areas; assigning a first value to the digital image data pixels within the first areas and a second value, different from the first value, to digital image data pixels within the second areas; calculating a total of first values and a total of second values; and calculating a total value for the defined data pattern by combining the total of first values and the total of second values.
5. A method of identifying an expected data pattern for use in facilitating automated drop-out forms processing by allowing identification of a dropped-out form, the expected data pattern being in a digital image of a page, the digital image being created by scanning the page, and the page bearing data arranged in a predetermined configuration generally conforming to a dropped-out form represented by a digital template form, the method comprising the steps of: obtaining the digital template form; defining, in the digital template form, a reference region having a boundary; identifying, within the reference region boundary, first areas where data is expected to be present, and second areas where no data is expected to be present, so as to define a defined data pattern; obtaining the digital image of the page bearing data; comparing the defined data pattern to the digital image data at an initial location within an identification region in the digital image, the identification region corresponding to the reference region in the template form; determining a fit of the defined data pattern to the digital image data at the initial location; repeatedly translating the defined data pattern relative to the digital image data, and redetermining the fit, until a location is identified that provides a best fit of the defined data pattern to the digital image data: and determining a final confidence level by comparing the best fit to a predetermined reference fit.
6. The method of claim 5. wherein the defined data pattern is unique to a particular form.
7. The method of claim 5, wherein the step of repeatedly translating the defined data pattern translates the defined data pattern within the identification region.
8. The method of claim 7, wherein the step of repeatedly translating the defined data pattern includes: dividing the identification region into a grid having predefined spacing; and translating the defined data pattern to each grid point in the grid.
9. The method of claim 5, wherein the step of determining a fit includes: counting digital image data pixels within the first areas and the second areas; assigning a first value to the digital image data pixels within the first areas and a second value, different from the first value, to digital image data pixels within the second areas; calculating a total of first values and a total of second values; and calculating a total value for the defined data pattern by combining the total of first values and the total of second values.
10. A method of identifying an expected data pattern for use in facilitating automated drop-out forms processing by allowing identification of a registration point location, the expected data pattern being in a digital image of a page, the digital image being created by scanning the page, and the page bearing data arranged in a predetermined configuration generally conforming to a dropped-out form represented by a digital template form, the method comprising the steps of: obtaining the digital template form: defining, in the digital template form, a reference region having a boundary; identifying, within the reference region boundary, first areas where data is expected to be present, and second areas where no data is expected to be present, so as to define a defined data pattern; defining a registration point location with respect to the defined data pattern, the defined data pattern being inside the reference region boundary; obtaining the digital image of the page bearing data; comparing the defined data pattern to the digital image data at an initial location within an identification region in the digital image, the identification region corresponding to the reference region in the digital template form; determining a fit of the defined data pattern to the digital image data at the initial location; repeatedly translating the defined data pattern relative to the digital image data, and redetermining the fit, until a location is identified that provides a best fit of the defined data pattern to the digital image data; and identifying the registration point location with respect to the expected data pattern, the expected data pattern being at the location where the defined data pattern provides the best fit to the digital image data within the identification region.
11. The method of claim 10, wherein the step of repeatedly translating the defined data pattern translates the defined data pattern within the identification region.
12. The method of claim 11, wherein the step of repeatedly translating the defined data pattern includes: dividing the identification region into a grid having predefined spacing; and translating the defined data pattern to each grid point in the grid.
13. The method of claim 10, wherein the step of determining a fit includes: counting digital image data pixels within the first areas and the second areas; assigning a first value to the digital image data pixels within the first areas and a second value, different from the first value, to digital image data pixels within the second areas; calculating a total of first values and a total of second values; and calculating a total value for the defined data pattern by combining the total of first values and the total of second values.
14. A method of identifying a drop-out form comprising the steps of: scanning a page; measuring a fit of data in the scanned image of the page to a first drop-out form template; comparing the fit to a predefined reference fit; responsive to a favorable fit comparison, ending the identification process; and responsive to an unfavorable fit comparison, repeatedly selecting a second drop-out form template, measuring a fit to the second drop-out form template, and comparing the new fit to the predefined reference fit until a favorable fit comparison is achieved.
15. A system for automated drop-out forms processing to identify a particular dropped-out form, comprising: a scanner for scanning a page bearing data to create a digital image of the page bearing data; a database containing a plurality of digital template forms, each digital template form representing a particular dropped-out form having a unique data pattern; and a computer program stored on a machine-readable media, the computer program being able to compare the digital image data to, one at a time, the plurality of digital template forms to determine to which particular digital template form the digital image data correlates.
PCT/US2000/005222 1999-02-26 2000-02-28 Method and system for identifying a reference region on an image of a dropped-out form WO2000051065A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU37126/00A AU3712600A (en) 1999-02-26 2000-02-28 Method and system for identifying a reference region on an image of a dropped-out form

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US12174799P 1999-02-26 1999-02-26
US60/121,747 1999-02-26

Publications (1)

Publication Number Publication Date
WO2000051065A1 true WO2000051065A1 (en) 2000-08-31

Family

ID=22398540

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2000/005222 WO2000051065A1 (en) 1999-02-26 2000-02-28 Method and system for identifying a reference region on an image of a dropped-out form

Country Status (3)

Country Link
US (1) US6539112B1 (en)
AU (1) AU3712600A (en)
WO (1) WO2000051065A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2317471A1 (en) * 2009-10-16 2011-05-04 Vestel Elektronik Sanayi ve Ticaret A.S. An automated test method including transparent black regions

Families Citing this family (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6950553B1 (en) * 2000-03-23 2005-09-27 Cardiff Software, Inc. Method and system for searching form features for form identification
US6904182B1 (en) * 2000-04-19 2005-06-07 Microsoft Corporation Whiteboard imaging system
US6778703B1 (en) * 2000-04-19 2004-08-17 International Business Machines Corporation Form recognition using reference areas
US6898317B2 (en) * 2001-05-07 2005-05-24 Hewlett-Packard Development Company, L.P. Method and system for fit-to-form scanning with a scanning device
SG161229A1 (en) * 2005-04-07 2010-05-27 Hermes Precisa Pty Ltd A method and system for managing information
JP2007241356A (en) * 2006-03-06 2007-09-20 Fuji Xerox Co Ltd Image processor and image processing program
US8255570B2 (en) * 2007-03-12 2012-08-28 Citrix Systems, Inc. Systems and methods of compression history expiration and synchronization
US7453379B2 (en) * 2007-03-12 2008-11-18 Citrix Systems, Inc. Systems and methods for identifying long matches of data in a compression history
US7532134B2 (en) * 2007-03-12 2009-05-12 Citrix Systems, Inc. Systems and methods for sharing compression histories between multiple devices
US7619545B2 (en) 2007-03-12 2009-11-17 Citrix Systems, Inc. Systems and methods of using application and protocol specific parsing for compression
US7865585B2 (en) 2007-03-12 2011-01-04 Citrix Systems, Inc. Systems and methods for providing dynamic ad hoc proxy-cache hierarchies
US7460038B2 (en) * 2007-03-12 2008-12-02 Citrix Systems, Inc. Systems and methods of clustered sharing of compression histories
US7827237B2 (en) * 2007-03-12 2010-11-02 Citrix Systems, Inc. Systems and methods for identifying long matches of data in a compression history
US7712150B2 (en) * 2007-10-01 2010-05-11 Pardillo Joseph M Protective pad assembly magnetically attachable to garment
EP2253121B1 (en) * 2008-01-11 2012-07-04 Telefonaktiebolaget L M Ericsson (publ) Message handling in an ip multimedia subsystem
US8195003B2 (en) * 2008-06-30 2012-06-05 International Business Machines Corporation Method of correcting digital image distortion caused by a sheet-fed scanner
US10095946B2 (en) 2016-07-07 2018-10-09 Lockheed Martin Corporation Systems and methods for strike through detection
CN115497114B (en) * 2022-11-18 2024-03-12 中国烟草总公司四川省公司 Structured information extraction method for cigarette logistics receiving bill

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5016040A (en) * 1987-05-26 1991-05-14 Silhouette Technology, Inc. Method and apparatus for forming a recording on a recording medium
US5542007A (en) * 1993-10-27 1996-07-30 International Business Machines Corporation Form dropout compression method which handles form white-out and writing in shaded and white-out areas of the form

Family Cites Families (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4741045A (en) 1983-09-23 1988-04-26 Dest Corporation Optical character isolation system, apparatus and method
US5038381A (en) 1988-07-11 1991-08-06 New Dest Corporation Image/text filtering system and method
US5657395A (en) 1989-03-02 1997-08-12 Minolta Camera Kabushiki Kaisha Image processing device providing improved image data processing by converting original image into a plurality of pixels and analyzing density data of each pixel
JP3105967B2 (en) * 1991-11-14 2000-11-06 キヤノン株式会社 Character recognition method and device
US5680479A (en) 1992-04-24 1997-10-21 Canon Kabushiki Kaisha Method and apparatus for character recognition
US5321768A (en) 1992-09-22 1994-06-14 The Research Foundation, State University Of New York At Buffalo System for recognizing handwritten character strings containing overlapping and/or broken characters
US6005976A (en) 1993-02-25 1999-12-21 Fujitsu Limited Image extraction system for extracting patterns such as characters, graphics and symbols from image having frame formed by straight line portions
JP3442466B2 (en) 1993-04-12 2003-09-02 株式会社リコー Image processing apparatus and image processing method
JP3547025B2 (en) 1994-08-23 2004-07-28 株式会社リコー Information extraction method
JP3345224B2 (en) 1995-03-06 2002-11-18 富士通株式会社 Pattern extraction device, pattern re-recognition table creation device, and pattern recognition device
JPH09168083A (en) 1995-12-15 1997-06-24 Murata Mach Ltd Image processing unit
JP3830998B2 (en) 1995-12-28 2006-10-11 株式会社東芝 Ruled line removal method and character recognition apparatus using the same
JP2973944B2 (en) * 1996-06-26 1999-11-08 富士ゼロックス株式会社 Document processing apparatus and document processing method
JP3189870B2 (en) 1996-12-24 2001-07-16 シャープ株式会社 Image processing device

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5016040A (en) * 1987-05-26 1991-05-14 Silhouette Technology, Inc. Method and apparatus for forming a recording on a recording medium
US5542007A (en) * 1993-10-27 1996-07-30 International Business Machines Corporation Form dropout compression method which handles form white-out and writing in shaded and white-out areas of the form

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2317471A1 (en) * 2009-10-16 2011-05-04 Vestel Elektronik Sanayi ve Ticaret A.S. An automated test method including transparent black regions

Also Published As

Publication number Publication date
AU3712600A (en) 2000-09-14
US6539112B1 (en) 2003-03-25

Similar Documents

Publication Publication Date Title
US6539112B1 (en) Methods and system for identifying a reference region on an image of a dropped-out form
US5410611A (en) Method for identifying word bounding boxes in text
US5822454A (en) System and method for automatic page registration and automatic zone detection during forms processing
US6778703B1 (en) Form recognition using reference areas
US5245676A (en) Determination of image skew angle from data including data in compressed form
US5539841A (en) Method for comparing image sections to determine similarity therebetween
US5506918A (en) Document skew detection/control system for printed document images containing a mixture of pure text lines and non-text portions
JP3308032B2 (en) Skew correction method, skew angle detection method, skew correction device, and skew angle detection device
JP5620244B2 (en) Image registration method and program for image comparison and document authentication
US4748678A (en) Method of storing and retrieving image data
US5465304A (en) Segmentation of text, picture and lines of a document image
US5748809A (en) Active area identification on a machine readable form using form landmarks
US7567708B2 (en) Apparatus and method for image processing
Nakano et al. An algorithm for the skew normalization of document image
KR100315318B1 (en) Apparatus and method of bitmap image processing, storage medium storing an image processing program
JPH05250409A (en) High-speed detection for page direction
CN109215016B (en) Identification and positioning method for coding mark
JPH0778818B2 (en) Binary image processing method for decoding self-clocking symbol shape code
EP0776512B1 (en) Method and apparatus for automatically specifying a portion of text from a bitmap image of the text
US20090167884A1 (en) Self-Similar Capture Systems
CN115619656A (en) Digital file deviation rectifying method and system
CN114970590A (en) Bar code detection method
US6968501B2 (en) Document format identification apparatus and method
US7961941B2 (en) Color form dropout using dynamic geometric solid thresholding
CN115953399B (en) Industrial part structural defect detection method based on contour features and SVDD

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AL AM AT AU AZ BA BB BG BR BY CA CH CN CR CU CZ DE DK DM EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

122 Ep: pct application non-entry in european phase