A method and system for performing image mark recognition for a document is disclosed. A document is scanned into a digital image. Reference image marks are sensed from the digital image. The reference image marks may include trigger row marks and/or corner/crop marks. Coordinates denoting the location of cells within the digital image are determined base on the locations of the reference image marks. Response marks are evaluated for darkness, opacity, and/or grayness on a pixel-by-pixel basis. The response marks are each assigned a percentage value based on the ratio between a total color value for a cell and a maximum color value for a cell.

Patent
   7574047
Priority
Jan 20 2004
Filed
Oct 09 2007
Issued
Aug 11 2009
Expiry
Feb 16 2024

TERM.DISCL.
Extension
27 days
Assg.orig
Entity
Small
1
66
all paid
1. A computer-implemented method of performing image mark recognition, comprising:
using a computer, accessing a scanned image of a document containing a plurality of image marks;
performing reference image mark recognition on the scanned image with the computer;
using the computer, performing coordinate mapping on the scanned image to determine coordinates for a plurality of cells within the scanned image, wherein each cell comprises a plurality of pixels;
analyzing one or more pixels in each cell with the computer; and
determining a cell color value for each cell with the computer based on values assigned to the one or more pixels.
8. A system for performing image mark recognition, comprising:
a processor; and
one or more computer readable mediums operatively coupled to the processor, wherein at least one of the one or more computer readable mediums contains a scanned image of a document,
wherein at least one of the one or more computer readable mediums contains computer program instructions for performing a method of performing image mark recognition, the method comprising:
accessing a scanned image of a document containing a plurality of image marks,
performing reference image mark recognition on the scanned image,
performing coordinate mapping on the scanned image to determine coordinates for a plurality of cells within the scanned image, wherein each cell comprises a plurality of pixels,
analyzing one or more pixels in each cell, and
determining a cell color value for each cell based on values assigned to the one or more pixels.
2. The method of claim 1 wherein performing reference image mark recognition comprises determining a location for each of one or more alignment marks.
3. The method of claim 2, further comprising verifying image mark alignment and spacing based on a resolution of the scanned image and the locations of the one or more alignment marks.
4. The method of claim 2 wherein determining the location of one or more alignment marks comprises examining each corner of the scanned image.
5. The method of claim 2 wherein determining the location of one or more alignment marks comprises searching for trigger rows on at least one margin of the scanned image.
6. The method of claim 1 wherein analyzing one or more pixels comprises incrementing a total color value for the cell by a color value for each pixel.
7. The method of claim 6 wherein determining a cell color value comprises setting the cell color value to the ratio of the total color value to a maximum color value.
9. The system of claim 8 wherein the computer program instructions for the step of performing reference image mark recognition comprise determining a location for each of one or more alignment marks.
10. The system of claim 9 wherein the computer program instructions for the step of performing reference image mark recognition further comprise verifying image mark alignment and spacing based on a resolution of the scanned image and the locations of the one or more alignment marks.
11. The system of claim 9 wherein the computer program instructions for the step of determining the location of one or more alignment marks comprise examining each corner of the scanned image.
12. The system of claim 9 wherein the computer program instructions for the step of determining the location of one or more alignment marks comprise searching for trigger rows on at least one margin of the scanned image.
13. The system of claim 8 wherein the computer program instructions for the step of analyzing one or more pixels comprise incrementing a total color value for the cell by a color value for each pixel.
14. The system of claim 13 wherein the computer program instructions for the step of determining a cell color value comprise setting the cell color value to the ratio of the total color value to a maximum color value.

This is a continuation application of U.S. Non-Provisional Patent Application No. 10/760,850, filed Jan. 20, 2004, titled “Method and System for Image Mark Recognition,” which was pending at the time the present application was filed and was issued as U.S. Pat. No. 7,298,902B2 on Nov. 20, 2007 and is hereby incorporated by reference.

The present invention generally relates to the field of mark recognition. The present invention particularly relates to scanning an image of a document and performing mark recognition on the scanned image. In an embodiment, a response mark may be indicative of an intended response to a question and/or may be representative of coded information. In a preferred embodiment, the present invention provides a hardware independent methodology for evaluating the presence or absence of one or more marks on a document, which may be configured to higher or lower resolutions as needed.

Conventional mark reading systems have been employed to extract information, such as pencil marks, from a document. These systems typically include a scanner configured to read marks from the document. The scanner creates a digitized representation of the document. The digitized document is then processed by a separate application to evaluate the document. In an assessment environment, this system can be used to score examination responses. For example, many students and other examinees are familiar with the typical response format for answering multiple choice questions which requires the student to fill-in a space (e.g. a circle, square, etc.) associated with the question. Mark reading systems can grade such answer sheets significantly faster and more accurately than a human grader. However, conventional mark reading systems are limited by the accuracy with which the document is digitized. Increased accuracy is available with more sophisticated and more expensive systems. When a high degree of accuracy is required, expensive scanning equipment can be purchased which will provide a digitized representation of the document within most accepted error rates. However, in addition to the expense associated with such systems, they may not be easily configured to alternate environments where differing accuracy levels are required. When accuracy in the digitizing process is sacrificed, such systems often improperly grade test answers. For example, the systems may not recognize properly marked target areas that are incompletely filled or in which the pencil marking is lighter than a threshold value. Moreover, the systems may recognize stray marks and erasures as incorrect answers when a test taker did not intend such an answer.

A particularly egregious error may result when a test taker properly marks a correct answer for a question, but a mark reading system recognizes a stray mark or erasure as an incorrect answer for the same question. In this case, although the test taker correctly answered the question, the test taker would not receive credit because the system would report that the test taker provided two answers. As such, the test taker may be penalized because of the medium in which the test is presented rather than the substantive material on which the test is based.

What is needed is an improved image mark recognition system that can accurately evaluate the marks on a document to be digitized which is both cost effective and independent of the hardware being used to scan the document.

Before the present methods, systems, and materials are described, it is to be understood that this invention is not limited to the particular methodologies, systems and materials described, as these may vary. It is also to be understood that the terminology used in the description is for the purpose of describing the particular versions or embodiments only, and is not intended to limit the scope of the present invention which will be limited only by the appended claims.

It must also be noted that as used herein and in the appended claims, the singular forms “a,” “an,” and “the” include plural references unless the context clearly dictates otherwise. Thus, for example, reference to a “mark” is a reference to one or more marks and equivalents thereof known to those skilled in the art, and so forth. Unless defined otherwise, all technical and scientific terms used herein have the same meanings as commonly understood by one of ordinary skill in the art. Although any methods, materials, and devices similar or equivalent to those described herein can be used in the practice or testing of embodiments of the present invention, the preferred methods, materials, and devices are now described. All publications mentioned herein are incorporated by reference. Nothing herein is to be construed as an admission that the invention is not entitled to antedate such disclosure by virtue of prior invention.

The present invention relates to enhancing a scanned image of a document or more accurately performing mark recognition of the scanned image to determine the presence or absence of a mark. The methodology of the present invention begins after a document has been digitized. The digitized image is then mapped to a coordinate system. The mapping is performed on the digitized image by using trigger rows, which are determined using existing marks on one margin of the document. The trigger rows are then logically extended across the document. Next, the present invention completes the mapping process by logically creating columns orthogonal to the extended trigger rows. The number of columns created can vary and may be configured to the degree of accuracy required. Alternately, the mapping process may be performed by using corner marks (also known as anchor marks or crop marks), which are marks printed at each corner of the document. A coordinate system is then created by determining the number of rows desired, which may vary, and equally dividing the space into horizontal rows, and by determining the number of columns desired, which may vary, and equally dividing the space into vertical columns. Document deskew may also be performed. Those skilled in the art will understand that other methods for mapping the digitized image to a coordinate system may also be used.

Next, a resolver engine evaluates each cell on a pixel-by-pixel basis. The pixel-by-pixel analysis results in a value for each cell. The value may represent the opacity of a response mark present in the cell. The result is that the digitized image has been converted into a series of numeric values each representing a potential response mark on the original document.

In a preferred embodiment, a method of performing image mark recognition includes accessing a scanned image of a document containing a plurality of image marks, performing reference image mark recognition on the scanned image, performing coordinate mapping on the scanned image to determine coordinates for a plurality of cells within the scanned image, analyzing one or more pixels in each cell, and determining a cell color value for each cell based on values assigned to the one or more pixels. In an embodiment, performing reference image mark recognition includes determining a location for each of one or more alignment marks and may further include verifying image mark alignment and spacing based on a resolution of the scanned image and the locations of the one or more alignment marks. In an embodiment, determining the location of one or more alignment marks includes examining each corner of the scanned image. In an embodiment, determining the location of one or more alignment marks includes searching for trigger rows on at least one margin of the scanned image.

In an embodiment, performing coordinate mapping includes computing logical row coordinates for a number of rows in the scanned image, computing logical column coordinates for a number of columns in the scanned image, creating a mapping within the scanned image, and resolving the mapping into coordinate values for each of a plurality of cells. In an embodiment, computing logical row coordinates and logical column coordinates includes using one or more trigger marks as reference points. In an embodiment, computing logical row coordinates and logical column coordinates includes using one or more corner marks as reference points. The number of rows and the number of columns may be automatically determined based on the step of performing reference image mark recognition. Alternately, the number of rows and the number of columns may be pre-determined. In an embodiment, creating a mapping includes determining absolute pixel positions for each cell within the scanned document as defined by an intersection of a row and a column.

In an embodiment, analyzing one or more pixels includes incrementing a total color value by a color value for each pixel in a cell. Determining a cell color value may include setting the cell color value to the ratio of the total color value to a maximum color value.

In a preferred embodiment, a system for performing image mark recognition includes a processor and one or more computer readable mediums operatively coupled to the processor. At least one of the one or more computer readable mediums contains a scanned image of a document. At least one of the one or more computer readable mediums contains computer program instructions for performing a method of performing image mark recognition.

The accompanying drawings, which are incorporated in and form a part of the specification, illustrate preferred embodiments of the present invention and, together with the description serve to explain the principles of the invention. The embodiments illustrated in the drawings should not be read to constitute limiting requirements, but instead are intended to assist the reader in understanding the invention.

FIG. 1 depicts an exemplary process flow for the image mark recognition process according to an embodiment of the present invention.

FIG. 2 depicts an exemplary cell evaluation ordering according to an embodiment of the present invention.

FIGS. 3A and 3B depict exemplary computations of cell color values according to an embodiment of the present invention.

FIG. 4 is a block diagram of exemplary internal hardware that may be used to contain or implement the program instructions of a system embodiment of the present invention.

The present invention relates to resolving a scanner image of a document to perform mark recognition on the scanned image. The present invention provides a system and method for resolving a digitized image of a document, independent of the hardware used to create the digitized image. The digitized image is enlisted by the present invention and a sequence of values representing the presence or absence of response marks on the document is created. This sequence of values is then available for use by any application which may be created to enhance the document, such as a scoring application which may be used to return a score for an examination.

An aspect of the present invention involves mapping the digitized image of a document to a coordinate system. This coordinate mapping may be performed by detecting trigger row marks and/or corner/crop marks to create a logical document mapping. A trigger row mark is a mark on an edge of a document that denotes boundaries of a row logically extending across the document. In an assessment environment, responses may be entered within the logical rows. Each row for which a test taker can potentially enter one or more responses may have a separate trigger row mark. Alternatively, a trigger row mark may be used to denote every second row, third row, or any other pattern that may be used to logically subdivide a document. An alternate method of coordinate mapping includes use of corner/crop marks placed in each corner of the document. The crop marks are located and used to define the coordinate system by sub-dividing the space between the crop marks into a logical coordinate system. For example, adjacent crop marks along the vertical margin of the document may be subdivided into a predetermined number of equally sized rows. Similarly, adjacent crop marks along the horizontal margin of the document may be subdivided into a predetermined number of equally sized columns intersecting with the rows. The pre-determined number of rows and columns may be verified by analyzing the location of the response marks. Alternatively, a user may enter information regarding the document, such as a document number, from which the row and column placement is determined.

As shown in FIG. 1, the process begins by accessing the digital image of a previously scanned document 105. The digital image may have been created by any means available and is not dependent on any particular hardware, format or resolution. A logical coordinate system is then mapped over the digital image 110 using the trigger row technique, the corner/crop mark techniques, or any other technique. The result of the coordinate mapping process 110 is that the digital image of the document has been logically subdivided into a plurality of cells that, in combination, comprise the entire digital image. The number of cells is configurable during the mapping process, and more or fewer cells may preferably be generated depending on the type of document being scanned and the processing required. The process then begins to iteratively analyze the image on a cell-by-cell basis to generate machine-readable data representative of the response marks on the document. This iterative process begins by obtaining the image 115 in the first cell. For example, as shown in FIG. 2, the mapping process may create an (N×M) coordinate system requiring evaluation of N*M cells. This process may, for example, begin by evaluating cell (1,1) 205, proceed to cell (1,2) 210 and continue in the same manner to cell (1,M) 215. Next, cell (2,1) 220 through (2,M) 225 and so on to cell (N,M) 230 may be evaluated. Alternate cell evaluation orderings and numberings are considered to be within the scope of this invention.

Returning to FIG. 1, the process may resolve the response marks in each cell into computer readable values. For each cell, a cell color value from, for example, 0 to 100, inclusive, representing the amount of darkness/lightness of a cell's coloration may be assigned 120. The cell color value may be assigned 120 based on a ratio of the sum of color values for each pixel in a cell having a non-white color (collectively, a total color value) to a maximum color value for the cell. The maximum color value may be equal to the number of pixels in a cell multiplied by a maximum pixel color value.

Two exemplary computations of a cell color value are shown in FIGS. 3A and 3B. In the exemplary computations, each pixel within a cell may be assigned a pixel color value ranging from 0 to 15, inclusive. In FIG. 3A, the cell has 4 pixels having a pixel color value equal to 0, 8 pixels having a pixel color value equal to 4, 12 pixels having a pixel color value equal to 9, and 12 pixels having a pixel color value equal to 15. The total color value for the cell in FIG. 3A is tc=(4*0)+(8*4)+(12*9)+(12*15)=32+108+180=320. The maximum color value for the cell in FIG. 3A is pc=(6 rows*6 columns*15)=540. The cell color value for the cell in FIG. 3A is v=tc/pc≈59%. Thus, the cell color value for the cell in FIG. 3A is 59.

Similar computations are used to compute the total color value, maximum color value and cell color value in FIG. 3B. Here, tc=172, pc=540, and v≈32%. Hence, the cell color value for the cell in FIG. 3B is 32.

Preferably, a value of 0 may represent the lightest possible mark (e.g., white), and a value of 100 may represent the darkest possible mark (e.g., black). The machine-readable data values for each cell may optionally be stored 125 to an output file.

FIG. 4 is a block diagram of exemplary internal hardware that may be used to contain or implement the program instructions of a system embodiment of the present invention. Referring to FIG. 4, a bus 428 serves as the main information highway interconnecting the other illustrated components of the hardware. CPU 402 is the central processing unit of the system, performing calculations and logic operations required to execute a program. Read only memory (ROM) 418 and random access memory (RAM) 420 constitute exemplary memory devices.

A disk controller 404 interfaces with one or more optional disk drives to the system bus 428. These disk drives may be external or internal floppy disk drives such as 410, CD ROM drives 406, or external or internal hard drives 408. As indicated previously, these various disk drives and disk controllers are optional devices.

Program instructions may be stored in the ROM 418 and/or the RAM 420. Optionally, program instructions may be stored on a computer readable medium such as a floppy disk or a digital disk or other recording medium, a communications signal or a carrier wave.

An optional display interface 422 may permit information from the bus 428 to be displayed on the display 424 in audio, graphic or alphanumeric format. Communication with external devices may optionally occur using various communication ports 426. An exemplary communication port 426 may be attached to a scanner 430 used to scan an image of a document into one or of the memory devices or disk drives.

In addition to the standard computer-type components, the hardware may also include an interface 412 which allows for receipt of data from input devices such as a keyboard 414 or other input device 416 such as a remote control, pointer and/or joystick.

An embedded system may optionally be used to perform one, some or all of the operations of the present invention. Likewise, a multiprocessor system may optionally be used to perform one, some or all of the operations of the present invention.

Although the invention has been described with reference to the preferred embodiments, it will be apparent to one skilled in the art that variations and modifications are contemplated within the spirit and scope of the invention. The drawings and description of the preferred embodiments are made by way of example rather than to limit the scope of the invention, and it is intended to cover within the spirit and scope of the invention all such changes and modifications.

Cichielo, Robert, Hoffman, Jeffrey

Patent Priority Assignee Title
8233200, Mar 14 2005 GTech Corporation Curvature correction and image processing
Patent Priority Assignee Title
3601906,
3870865,
4153895, Aug 01 1978 BANTEC, INC , A CORP, OF DELAWARE Mark sense reader
4217487, May 26 1978 National Computer Systems, Inc. Self-calibrating apparatus for photoelectrically scanning answer documents
4300123, Jan 02 1979 NATIONAL COMPUTER SYSTEMS, INC Optical reading system
4817179, Dec 29 1986 SCAN-OPTICS, LLC Digital image enhancement methods and apparatus
4979136, Mar 01 1988 TRANSITIONS RESEARCH CORPORATION, A CORP OF CT Processing system and method for enhancing image data
5038393, Jun 30 1987 Kabushiki Kaisha Toshiba Method of effectively reading data written on data sheet, and data reading apparatus therefor
5101448, Aug 24 1988 Hitachi, Ltd. Method and apparatus for processing a document by utilizing an image
5102341, May 05 1989 Touchstone Applied Science Associates, Inc. Test answer and score sheet device
5103490, Jun 13 1990 NATIONAL COMPUTER SYSTEMS, INC Method and apparatus for storing and merging multiple optically scanned images
5129014, Dec 08 1989 Xerox Corporation; XEROX CORPORATION, A CORP OF NY Image registration
5134669, Jun 13 1990 DATA MANAGEMENT I LLC Image processing system for documentary data
5253304, Nov 27 1991 American Telephone and Telegraph Company Method and apparatus for image segmentation
5452379, Sep 13 1991 NCS PEARSON, INC Image capture and storage techniques in association with optical mark reading
5568571, Dec 14 1992 PROQUEST INFORMATION AND LEARNING, LLC Image enhancement system
5572601, Mar 20 1992 Xerox Corporation Mark sensing on a form
5600732, Dec 08 1994 BANCTEC, INC Document image analysis method
5625770, Apr 20 1991 Ricoh Company, Ltd. File system for filing documents, using a catalog card having a plurality of marks thereon
5649026, Nov 21 1994 Opex Corporation Apparatus for detecting marks on documents
5672060, Jul 08 1992 NCS PEARSON, INC Apparatus and method for scoring nonobjective assessment materials through the application and use of captured images
5675671, Sep 06 1991 System for orienting documents in the automated processing of bulk mail and the like
5754674, Dec 08 1994 BancTec, Inc. Document image analysis method
5818976, Oct 25 1993 SOQUE HOLDINGS, LTD Method and apparatus for document skew and size/shape detection
5825947, Sep 01 1995 Olympus Optical Co., Ltd. Optical reproducing system for multimedia information recorded with code data having function for correcting image reading distortion
5870488, May 07 1996 Fortrend Engineering Corporation Method and apparatus for prealigning wafers in a wafer sorting system
5956422, Oct 20 1995 BCL Computers, Inc. Processor based method for extracting tablets from printed documents
6079624, Dec 08 1997 William C., Apperson Data processing form using a scanning apparatus
6129278, May 19 1994 Symbol Technologies, LLC Detecting image cell position with subpixel accuracy
6173154, Jul 31 1997 HARCOURT ASSESSMENT, INC System and method for imaging test answer sheets having open-ended questions
6188787, Apr 05 1996 HANGER SOLUTIONS, LLC Image recognition method and device and copier and scanner employing same
6213664, Jun 15 2000 Card marking device
6259814, Oct 17 1997 Canon Kabushiki Kaisha Image recognition through localized interpretation
6299066, Apr 16 1997 Delaware Capital Formation, Inc. Marked card reader employing afocal spatial filtering
6311040, Jul 31 1997 HARCOURT ASSESSMENT, INC System and method for scoring test answer sheets having open-ended questions
6325286, Apr 16 1997 Delaware Capital Formation, Inc Chemical dispensing system using keyboardless data entry
6408106, Jul 27 1998 Olympus Optical Co., Ltd. Information reading system
6542629, Jul 22 1999 Xerox Corporation Digital imaging method and apparatus for detection of document security marks
6575367, Nov 05 1998 Welch Allyn Data Collection, Inc Image data binarization methods enabling optical reader to read fine print indicia
6580820, Jun 09 1999 Xerox Corporation Digital imaging method and apparatus for detection of document security marks
6684052, Jul 31 1997 HARCOURT ASSESSMENT, INC Scanning system for imaging and storing the images of test answer sheets having open-ended questions
6741738, Mar 13 2000 Accusoft Corporation Method of optical mark recognition
6927872, Jul 25 2001 HEWLETT-PACKARD DEVELOPMENT COMPANY L P Data acquisition system and method using answer forms
6970267, Jan 12 2001 SCAN-OPTICS, LLC Gray scale optical mark reader
7004393, Jul 26 1996 Opex Corporation Method and apparatus for processing and determining the orientation of documents
7164797, Apr 25 2002 Microsoft Technology Licensing, LLC Clustering
7236653, Mar 27 2003 Sharp Kabushiki Kaisha System and method for locating document areas using markup symbols
7242806, Jan 23 2002 Honeywell International Inc. Methods, functional Data, and Systems for image feature translation
20010033688,
20020168090,
20020181805,
20030072019,
20030173404,
20040139391,
20040185424,
20040202992,
20050157930,
20050201639,
20060062453,
20060250660,
20060252023,
20060255145,
20070116358,
20070192176,
JP6176192,
JP62263584,
///
Executed onAssignorAssigneeConveyanceFrameReelDoc
Jan 14 2004CICHIELO, ROBERTEducational Testing ServiceASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0213320464 pdf
Jan 16 2004HOFFMAN, JEFFREYEducational Testing ServiceASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0213320464 pdf
Oct 09 2007Educational Testing Service(assignment on the face of the patent)
Date Maintenance Fee Events
Jan 25 2013M2551: Payment of Maintenance Fee, 4th Yr, Small Entity.
Jan 26 2017M2552: Payment of Maintenance Fee, 8th Yr, Small Entity.
Jan 15 2021M2553: Payment of Maintenance Fee, 12th Yr, Small Entity.


Date Maintenance Schedule
Aug 11 20124 years fee payment window open
Feb 11 20136 months grace period start (w surcharge)
Aug 11 2013patent expiry (for year 4)
Aug 11 20152 years to revive unintentionally abandoned end. (for year 4)
Aug 11 20168 years fee payment window open
Feb 11 20176 months grace period start (w surcharge)
Aug 11 2017patent expiry (for year 8)
Aug 11 20192 years to revive unintentionally abandoned end. (for year 8)
Aug 11 202012 years fee payment window open
Feb 11 20216 months grace period start (w surcharge)
Aug 11 2021patent expiry (for year 12)
Aug 11 20232 years to revive unintentionally abandoned end. (for year 12)