systems and methods for image modification to increase contrast between text and non-text pixels within the image. In one embodiment, an original document image is scaled to a predetermined size for processing by a convolutional neural network. The convolutional neural network identifies a probability that each pixel in the scaled is text and generates a heat map of these probabilities. The heat map is then scaled back to the size of the original document image, and the probabilities in the heat map are used to adjust the intensities of the text and non-text pixels. For positive text, intensities of text pixels are reduced and intensities of non-text pixels are increased in order to increase the contrast of the text against the background of the image. Optical character recognition may then be performed on the contrast-adjusted image.

Patent
   10776903
Priority
Jul 17 2017
Filed
Dec 21 2018
Issued
Sep 15 2020
Expiry
Oct 02 2038
Extension
81 days
Assg.orig
Entity
Large
2
11
currently ok
6. A method for modifying images, the method comprising:
receiving an input image from an input/output device communicatively connected to an image recognition system;
if an original size of the input image does not match a predetermined image size which the image recognition system is configured to process, scaling the input image to the predetermined image size;
generating, using a convolutional neural network of the image recognition system, at least a first heat map, wherein the first heat map indicates, for each pixel in the input image, a probability that the pixel forms a portion of a text character in the input image;
if the predetermined image size which the image recognition system is configured to process does not match the original size of the input image, scaling the first heat map to the original size of the input image;
modifying intensities of at least a subset of the pixels in the input image according to the first heat map to produce a modified image, wherein the intensities of a set of text pixels are adjusted in a first direction and a set of non-text pixels neighboring the set of text pixels are adjusted in a second direction opposite the first direction, thereby increasing a contrast between the set of text pixels and the set of neighboring non-text pixels;
storing the modified image in one or more memories coupled to the image recognition system;
retrieving the modified image from the one or more memories;
performing character recognition on the modified image and thereby extracting one or more text characters from the modified image; and
storing the one or more extracted text characters in the one or more memories.
11. A computer program product comprising a non-transitory computer-readable medium storing instructions executable by one or more processors to cause the one or more processors to perform:
receiving an input image from an input/output device communicatively connected to an image recognition system;
if an original size of the input image does not match a predetermined image size which the image recognition system is configured to process, scaling the input image to the predetermined image size;
generating, using a convolutional neural network of the image recognition system, at least a first heat map, wherein the first heat map indicates, for each pixel in the input image, a probability that the pixel forms a portion of a text character in the input image;
if the predetermined image size which the image recognition system is configured to process does not match the original size of the input image, scaling the first heat map to the original size of the input image;
modifying intensities of at least a subset of the pixels in the input image according to the first heat map to produce a modified image, wherein the intensities of a set of text pixels are adjusted in a first direction and a set of non-text pixels neighboring the set of text pixels are adjusted in a second direction opposite the first direction, thereby increasing a contrast between the set of text pixels and the set of neighboring non-text pixels;
storing the modified image in one or more memories coupled to the image recognition system;
retrieving the modified image from the one or more memories;
performing character recognition on the modified image and thereby extracting one or more text characters from the modified image; and
storing the one or more extracted text characters in the one or more memories.
1. An image modification system comprising:
at least one processor; and
at least one non-transitory computer readable medium storing instructions that are executable by the at least one processor to cause the at least one processor to perform:
implementing a fully convolutional neural network;
receiving an input image from an input/output device communicatively connected to the image recognition system over a network;
if an original size of the input image does not match a predetermined image size which the convolutional neural network is configured to process, scaling the input image to the predetermined image size;
performing, using the fully convolutional neural network, a set of convolutional operations on the input image so as to produce at least a first heat map, wherein the first heat map indicates, for each pixel in the input image, a probability that the pixel forms a portion of a text character in the input image;
if the predetermined image size which the convolutional neural network is configured to process does not match the original size of the input image, scaling the first heat map to the original size of the input image;
modifying intensities of at least a subset of the pixels in the input image according to the first heat map to produce a modified image, wherein the intensities of a set of text pixels are adjusted in a first direction and a set of non-text pixels neighboring the set of text pixels are adjusted in a second direction opposite the first direction, thereby increasing a contrast between the set of text pixels and the set of neighboring non-text pixels;
storing the modified image in one or more memories coupled to the at least one processor;
retrieving the modified image from the one or more memories;
performing character recognition on the modified image and thereby extracting one or more text characters from the modified image; and
storing the one or more extracted text characters in the one or more memories.
2. The image modification system of claim 1, wherein for each pixel in the input image, a magnitude of an intensity adjustment for the pixel is determined in dependence upon an initial intensity of the pixel in the input image and a probability value in the first heat map corresponding to the pixel.
3. The image modification system of claim 1, wherein for each pixel in the input image, the pixel has a corresponding RGB (Red, Green, Blue) representation, wherein modifying the intensity of the pixel comprises converting the RGB representation of the pixel to an HSL (Hue, Saturation, Lightness) representation of the pixel, adjusting a lightness value of the HSL representation of the pixel, and converting the adjusted HSL representation of the pixel to an adjusted RGB representation of the pixel.
4. The image modification system of claim 1, wherein for at least a portion of the pixels in the input image, the corresponding modified intensities of the pixels are less than a maximum intensity and greater than a minimum intensity.
5. The image modification system of claim 1, wherein modifying the intensities of the pixels in the input image comprises:
for text pixels representing positive text characters, reducing the corresponding intensities of the pixels;
for text pixels representing inverted text characters, increasing the corresponding intensities of the pixels;
for non-text pixels neighboring positive text characters, increasing the corresponding intensities of the pixels; and
for non-text pixels neighboring inverted text characters, reducing the corresponding intensities of the pixels.
7. The method of claim 6, wherein for each pixel in the input image, a magnitude of an intensity adjustment for the pixel is determined in dependence upon an initial intensity of the pixel in the input image and a probability value in the first heat map corresponding to the pixel.
8. The method of claim 6, wherein for each pixel in the input image, the pixel has a corresponding RGB (Red, Green, Blue) representation, wherein modifying the intensity of the pixel comprises converting the RGB representation of the pixel to an HSL (Hue, Saturation, Lightness) representation of the pixel, adjusting a lightness value of the HSL representation of the pixel, and converting the adjusted HSL representation of the pixel to an adjusted RGB representation of the pixel.
9. The method of claim 6, wherein for at least a portion of the pixels in the input image, the corresponding modified intensities of the pixels are less than a maximum intensity and greater than a minimum intensity.
10. The method of claim 6, wherein modifying the intensities of the pixels in the input image comprises:
for text pixels representing positive text characters, reducing the corresponding intensities of the pixels;
for text pixels representing inverted text characters, increasing the corresponding intensities of the pixels;
for non-text pixels neighboring positive text characters, increasing the corresponding intensities of the pixels; and
for non-text pixels neighboring inverted text characters, reducing the corresponding intensities of the pixels.
12. The computer program product of claim 11, wherein for each pixel in the input image, a magnitude of an intensity adjustment for the pixel is determined in dependence upon an initial intensity of the pixel in the input image and a probability value in the first heat map corresponding to the pixel.
13. The computer program product of claim 11, wherein for each pixel in the input image, the pixel has a corresponding RGB (Red, Green, Blue) representation, wherein modifying the intensity of the pixel comprises converting the RGB representation of the pixel to an HSL (Hue, Saturation, Lightness) representation of the pixel, adjusting a lightness value of the HSL representation of the pixel, and converting the adjusted HSL representation of the pixel to an adjusted RGB representation of the pixel.
14. The computer program product of claim 11, wherein for at least a portion of the pixels in the input image, the corresponding modified intensities of the pixels are less than a maximum intensity and greater than a minimum intensity.
15. The computer program product of claim 11, wherein modifying the intensities of the pixels in the input image comprises:
for text pixels representing positive text characters, reducing the corresponding intensities of the pixels;
for text pixels representing inverted text characters, increasing the corresponding intensities of the pixels;
for non-text pixels neighboring positive text characters, increasing the corresponding intensities of the pixels; and
for non-text pixels neighboring inverted text characters, reducing the corresponding intensities of the pixels.

This application is a continuation in part of and a benefit of priority under 35 U.S.C. § 120 from U.S. patent application Ser. No. 16/035,307, filed Jul. 13, 2018, entitled “SYSTEMS AND METHODS FOR IMAGE BASED CONTENT CAPTURE AND EXTRACTION UTILIZING DEEP LEARNING NEUTRAL NETWORK AND BOUNDING BOX DETECTION TRAINING TECHNIQUES,” which claims a benefit of priority under 35 U.S.C. § 119(e) from U.S. Provisional Application No. 62/533,576 filed Jul. 17, 2017 entitled “SYSTEMS AND METHODS FOR IMAGE BASED CONTENT CAPTURE AND EXTRACTION UTILIZING DEEP LEARNING NEURAL NETWORK AND BOUNDING BOX DETECTION TRAINING TECHNIQUES”. Both of these applications are incorporated by reference herein in their entireties.

This disclosure relates generally to the field of image recognition systems, and more particularly to the field of fully convolutional neural networks that implement image enhancement to enable improved optical character recognition.

Increasingly, businesses require the ability to capture information that is contained in conventional, tangible documents and to transfer this information into their own computer systems. This process typically involves capturing images of the pre-existing documents, identifying areas of the document that contain the desired information (which may include text or other types of information within the document), and interpreting these areas of the document image (e.g., through optical character recognition, image extraction, etc.) to obtain the information.

Document images often have “noisy” backgrounds that make it more difficult to distinguish text in an image from other elements of the image. These other background elements may include, for example, photographic images, security elements (e.g., fine lines), watermarks, and the like. The document image may also have lighting variations, skewing, or other characteristics which increase the difficulty of distinguishing text in the image from the various background elements. The background elements and other document image characteristics may cause optical character recognition (OCR) algorithms to be confused, and to mis-identify textual and non-textual pixels. For instance, a line which is a security element may be identified as a text element, or a thin or lightened part of a text element may be identified as background or noise.

In some cases, conventional empirical techniques such as thin-line removal are used in an effort to compensate for non-textual features, effectively attempting to remove the background. Once the background has been removed, traditional OCR may be more effective at correctly extracting the text from the document image. Line removal is not a complete solution, however, as it is sensitive to such factors as the size of the text and the thickness of the lines. Furthermore, line removal is ineffective with respect to other types of backgrounds, such as photographic images.

Another problem with conventional techniques for improving the effectiveness of OCR is that they commonly focus on removal of the background. When a background region is identified, it is typically “removed” by making the pixels white. This may create a sharp change in pixel intensities at the boundaries of the background region (the original pixel intensities outside the region may be relatively light, but may still contrast with the white (maximum intensity) pixels in the background region. This sudden change in intensity at the edge of the background region may be misinterpreted by OCR algorithms as an identifiable feature in the image. This can “confuse” the algorithms and reduce the effectiveness of the algorithms.

It would therefore be desirable to have systems and methods for improving the effectiveness of identifying the portions of a document image that correspond to text and the portions that correspond to non-textual background elements, so that OCR algorithms can more effectively recognize and extract text from the document image.

Embodiments disclosed herein make use of deep learning techniques to identify text pixels and non-text pixels in an image, and to adjust the intensities of the pixels to increase the contrast between the text pixels and the non-text pixels. Rather than simply setting the non-text pixels to white, the intensities of the pixels are adjusted by variable amounts to create smoother transitions at the boundaries between text pixels and non-text pixels, thereby reducing or preventing sudden transitions that can confuse or be misinterpreted by OCR algorithms. The increased contrast and reduction of abrupt transitions in the background enables more efficient and effective character recognition.

These benefits may be provided in a system that uses an advanced recognition platform which implements a convolutional neural network to identify probabilities that pixels in an image are text or non-text (background). This information is then used to increase the contrast between text and non-text portions of the image.

One embodiment comprises an image modification system that has a processor and a computer readable medium which stores instructions that are executed by the processor to implement a convolutional neural network. The system is configured to receive an input image (e.g., from an input/output device that is communicatively connected to the system via a network). If the original size of the input image does not match a predetermined image size which the convolutional neural network is configured to process, the system scales the input image to the predetermined image size. The system then performs a set of convolutional operations on the input image to produce a first heat map which indicates, for each pixel in the input image, a probability that the pixel forms a portion of a text character in the input image. If the original image was scaled to the predetermined image size, the first heat map is then scaled back to the original size of the input image. The intensities of the pixels in the input image are then modified according to the first heat map to produce a modified image that has increased contrast between the text pixels and non-text pixels. This includes adjusting the intensities of a set of text pixels in a first direction and a set of neighboring non-text pixels in a second, opposite direction, thereby increasing a contrast between the set of text pixels and the set of neighboring non-text pixels. The system then stores the modified image in a memory coupled to the at least one processor. The system may retrieve the modified image from the memory, perform optical character recognition on the modified image and thereby extract text characters from the modified image, and store the extracted text characters in the memory.

As used herein, “direction” refers to the increase or decrease of the pixel intensities. For example, if the text pixels are adjusted downward (if their intensities are decreased), the neighboring non-text pixels are adjusted upward (their intensities are increased). Because a document image may include both positive and inverted text, some portions of the image may be handled one way (e.g., text pixels are darkened and non-text pixels are lightened), while other portions of the image may be handled a different way (e.g., non-text pixels are darkened and text pixels are lightened). In one embodiment, modifying the intensities of the pixels in the input image includes: for text pixels representing positive text characters, reducing the corresponding intensities of the pixels; for text pixels representing inverted text characters, increasing the corresponding intensities of the pixels; for non-text pixels neighboring positive text characters, increasing the corresponding intensities of the pixels; and for non-text pixels neighboring inverted text characters, reducing the corresponding intensities of the pixels.

In one embodiment, for each pixel in the input image, the magnitude of the intensity adjustment for the pixel is determined in dependence upon an initial intensity of the pixel in the input image and a probability value in the first heat map corresponding to the pixel. For instance, if a pixel is already very dark, its intensity may be reduced only slightly, but if it is lighter, its intensity may be reduced to a greater degree to make it darker. Similarly, if a given pixel has a corresponding text probability of 90%, it may be adjusted (e.g., darkened) more than if its text probability was 60%. In one embodiment, the modified intensities of the pixels are constrained to be less than the maximum possible intensity and greater than the minimum possible intensity.

In one embodiment, each pixel in the input image may have a corresponding RGB (Red, Green, Blue) representation, and modifying the intensity of the pixel may comprise converting the RGB representation of the pixel to an HSL (Hue, Saturation, Lightness) representation, adjusting a lightness value of the HSL representation of the pixel, and converting the adjusted HSL representation of the pixel back to an adjusted RGB representation. By converting to the HSL representation, the intensity adjustment involves only the lightness value in the pixel representation, so the contrast of the image is increased while the colors remain the same.

Alternative embodiments may include methods for modifying images generally as described above, as well as computer program products which implement these functions, where these products comprise a non-transitory computer-readable medium that stores instructions which are executable by a processor to cause the processors to perform the described functions. Numerous other alternative embodiments are also possible.

These, and other, aspects of the disclosure will be better appreciated and understood when considered in conjunction with the following description and the accompanying drawings. It should be understood, however, that the following description, while indicating various embodiments of the disclosure and numerous specific details thereof, is given by way of illustration and not of limitation. Many substitutions, modifications, additions, and/or rearrangements may be made within the scope of the disclosure without departing from the spirit thereof, and the disclosure includes all such substitutions, modifications, additions, and/or rearrangements.

The drawings accompanying and forming part of this specification are included to depict certain aspects of the invention. A clearer impression of the invention, and of the components and operation of systems provided with the invention, will become more readily apparent by referring to the exemplary, and therefore non-limiting, embodiments illustrated in the drawings, wherein identical reference numerals designate the same components. Note that the features illustrated in the drawings are not necessarily drawn to scale.

FIG. 1 depicts a high-level diagrammatic representation of an exemplary system for extracting information from images and providing this information to a customer according to some embodiments.

FIG. 2 depicts a diagrammatic representation of a distributed network computing environment where embodiments disclosed can be implemented

FIG. 3 depicts a diagrammatic representation of an architecture of a data recognition and extraction server according to some embodiments.

FIG. 4 depicts a flow chart illustrating a method for processing images in a convolutional neural network according to some embodiments.

FIGS. 5A-5C depict diagrammatic representations of an input image such as a driver's license and corresponding image maps showing areas in which features such as faces and text are found according to some embodiments.

FIG. 6 is a flow diagram illustrating a method for increasing the contrast of an image according to some embodiments.

FIG. 7 is a diagrammatic representation of the topology of a convolutional neural network according to some embodiments.

FIG. 8 depicts a flow chart illustrating a method for generating a training data set and training a convolutional neural network according to some embodiments.

FIG. 9 depicts a diagrammatic representation of the generation of simulated documents and corresponding simulated document images based on a set of document components according to some embodiments.

FIGS. 10A-10D depict several examples of image variations that may be incorporated into simulated document images that are generated according to some embodiments.

The invention and the various features and advantageous details thereof are explained more fully with reference to the non-limiting embodiments that are illustrated in the accompanying drawings and detailed in the following description. Descriptions of well-known starting materials, processing techniques, components and equipment are omitted so as not to unnecessarily obscure the invention in detail. It should be understood, however, that the detailed description and the specific examples, while indicating some embodiments of the invention, are given by way of illustration only and not by way of limitation. Various substitutions, modifications, additions and/or rearrangements within the spirit and/or scope of the underlying inventive concept will become apparent to those skilled in the art from this disclosure.

Referring to FIG. 1, a high-level diagrammatic representation of an exemplary system for modifying images, extracting information from images and providing this information to a customer is shown. In this system, an input/output (I/O) device 110 is provided to capture images from which data is to be extracted. I/O device 110 may, for example, be a mobile phone, a camera, a document scanner, a digital document storage, or any other device suitable for inputting images to the system. I/O device 110 provides images (e.g., photographs, scanned documents, faxes, etc.) to a server 120. Server 120 can be configured for receiving and processing the images from I/O device 110 to extract information from the images. Server 120 utilizes a processor 122 that receives program instructions from a memory 126 which implement a convolutional neural network 124.

It should be noted that references herein to modification of images or pixels of images should be construed as modifications to image files and/or pixel values contained in the files. The image files and pixel values (including RGB, HSL and other representations of the pixels) may be stored temporarily or persistently in the various memories and other data storage devices that are available to the system.

The program instructions may be configured in some embodiments to enable processor 122 to dynamically generate simulated documents, and to generate images of these simulated documents from document components which are stored in data storage 128. The images of the simulated documents, as well as the information contained in the document components, are used to train convolutional neural network 124. The trained convolutional neural network can then identify areas within the input images received from I/O device 110 which contain text, images, signatures and/or other information, and extract the corresponding information (e.g., using optical character recognition to recognize text within the text areas). The extracted information is then provided to a customer or consumer of the information, such as backend server 130. The consumer may be a data storage device, a backend server, a user, or any other consumer of the data.

These documents and images that are dynamically generated by processor 122 may alternatively be referred to by such terms as “dynamically generated”, “artificial”, “simulated”, or “synthetic” to indicate that the documents and images are created by the system using the document components, rather than being copies of pre-existing documents or images of these pre-existing documents. For example, a dynamically generated document to simulate a driver's license may be constructed using randomly selected components such as a name, address, photograph, driver's license number, license format, etc. The simulated document is not associated with a real person, but instead represents a fictitious person. The components individually may contain either real or fictitious information (e.g., an actual, physical address or a fictitious address). It should be noted that, for purposes of this disclosure, a “simulated” image is a real image, but it is an image of a simulated document that has been created from the document components.

FIG. 2 depicts a diagrammatic representation of an exemplary distributed network computing environment where embodiments disclosed herein can be implemented. The structure of this computing environment is illustrative, rather than limiting, and embodiments of the invention can be implemented in environment having many other, alternative structures. In the example illustrated in FIG. 2, network computing environment 200 includes network 214 that can be bi-directionally coupled to computer 212, computer 215, and computer 216. Computer 216 can be bi-directionally coupled to data store 218. Network 214 may represent a combination of wired and wireless networks that network computing environment 200 may utilize for various types of network communications known to those skilled in the art. In one embodiment, computer 212 may capture images that are provided via network 214 to computer 216, which recognizes and extracts information from the images as disclosed in more detail below. The information extracted from the images may be classified and otherwise interpreted and provided to backend systems implemented on computer 215.

For the purpose of illustration, a single system is shown for each of computer 212, computer 215, and computer 216. However, with each of computer 212, computer 215, and computer 216, a plurality of computers (not shown) may be interconnected to each other over network 214. For example, a plurality of computers 212 and a plurality of computers 215 may be coupled to network 214. Computers 212 may include data processing systems for communicating with computer 216. Computers 215 may include data processing systems for individuals whose jobs may require them to configure services used by computers 212 in network computing environment 200.

First enterprise computer 212 can include central processing unit (“CPU”) 220, read-only memory (“ROM”) 222, random access memory (“RAM”) 224, hard drive (“HD”) or storage memory 226, and input/output device(s) (“I/O”) 228. I/O 229 can include a keyboard, monitor, printer, electronic pointing device (e.g., mouse, trackball, stylus, etc.), or the like. Computer 212 can include a desktop computer, a laptop computer, a personal digital assistant, a cellular phone, or nearly any device capable of communicating over a network. Computer 215 may be similar to computer 212 and can comprise CPU 250, ROM 252, RAM 254, HD 256, and I/O 258.

Likewise, computer 216 may include CPU 260, ROM 262, RAM 264, HD 266, and I/O 268. Computer 216 may include one or more backend systems configured for providing a variety of services to computers 212 over network 214. These services may utilize data stored in data store 218. Many other alternative configurations are possible and known to skilled artisans.

Each of the computers in FIG. 2 may have more than one CPU, ROM, RAM, HD, I/O, or other hardware components. For the sake of brevity, each computer is illustrated as having one of each of the hardware components, even if more than one is used. Each of computers 212, 215, and 216 is an example of a data processing system. ROM 222, 252, and 262; RAM 224, 254, and 264; HD 226, 256, and 266; and data store 218 can include media that can be read by CPU 220, 250, or 260. Therefore, these types of memories include non-transitory computer-readable storage media. These memories may be internal or external to computers 212, 215, or 216.

Portions of the methods described herein may be implemented in suitable software code that may reside within ROM 222, 252, or 262; RAM 224, 254, or 264; or HD 226, 256, or 266. In addition to those types of memories, the instructions in an embodiment disclosed herein may be contained on a data storage device with a different computer-readable storage medium, such as a hard disk. Alternatively, the instructions may be stored as software code elements on a data storage array, magnetic tape, floppy diskette, optical storage device, or other appropriate data processing system readable medium or storage device.

FIG. 3 is a diagrammatic representation of the architecture of a data recognition and extraction server in an exemplary embodiment. Extraction service 320 may be configured as described herein to modify received images as described herein to adjust their pixel intensities, increasing the contrast between text and non-text pixels, which in turn increases the effectiveness of the system's character recognition algorithms. Extraction service 320 uses a convolutional neural network 322 to identify regions within input images where the desired types of information (e.g., text, images, signatures, etc.) are found. The information can then be extracted from the identified regions of the images. Convolutional neural network 322 is initially trained in one embodiment by a training service 350. In this embodiment, the convolutional neural network is trained offline, and the already-trained network can be downloaded or otherwise provided to the extraction service 320. Because the convolutional neural network has already been trained offline, it can be used immediately by extraction service 320. While neural network 322 has been trained prior to being provided to extraction service 320, it may be further trained as it processes input images during normal operation of extraction service 320.

As will be described in more detail below, the offline training of the convolutional neural network involves providing document components such as faces, signatures, names addresses, and the like to a simulated document generator, which then generates simulated documents and simulated document images from the document components. The simulated document images and the information from the document components are then used by a neural network learner 352 which implements a learning service to train a convolutional neural network. While the training process implemented by the neural network learner is known in the field, this process conventionally requires vast quantities of actual data. For example, in order to train a convolutional neural network to recognize driver's licenses, it would be necessary to provide large numbers of images of actual driver's licenses to the system. The convolutional neural network would process the images and extract information from the images. This extracted information would then be compared to information that was expected to be extracted, and corrections would be made to the weights of the filters used in the convolution layers of the convolutional neural network in order to improve the network's recognition accuracy moving forward. In many cases, the large amount of pre-existing data (images and expected recognition output) often makes training expensive and impractical (or even impossible). The use of a simulated data generator eliminates the need for a data source that can supply the large numbers of actual document images by generating simulated document images from document components and providing them to the neural network learner.

An application program interface (API) 310 is provided to enable access to the server on which extraction service 320 executes. API 310 allows images to be input to the server so that the server can recognize and extract information from the images. The images that are input through API 310 are provided to extraction service 320, which performs the recognition of relevant portions of the images and extraction of text or other information (images, signatures, etc.) from the relevant portions of the images. Information extracted from the images by service 320 is provided to learning service 330. Correction information is also provided to learning service 330 from API 310. The extracted information and correction information are used by learning service 330 to further train extraction service 320. Information extracted from images provided to service 320 can also be communicated to content transformation service 340, which may transform or interpret the raw data from the extraction service (e.g., identifying text as a name or identifying numbers as a birthdate).

As noted above, the extraction of information from images using template-based techniques and the maintenance of systems that use these techniques can be very inefficient and ineffective. Extraction service 320 is therefore implemented in one embodiment using a convolutional neural network. This convolutional neural network is designed in this embodiment to process images generally according to the process illustrated in FIG. 4. As depicted in this figure, the process begins when an image is received (402). The received image is initially pre-processed (404).

In one embodiment, the pre-processing includes linearizing and centering the image, converting color in the image to gray scale and normalizing the grayscale to a range from −0.5 to 0.5. Some embodiments may also implement an edge-preserving noise filter to improve quality. Normally, each of the colors of the image (red, green, blue) are processed as separate layers, so converting the image to a gray scale (typically ranging from 0-255) reduces these three layers to a single layer, which increases the processing speed of the system. The gray scale is normalized from the range of 0 to 255 to a range of −0.5 to 0.5 in order to improve the learning quality of the system. In alternative embodiments, it may be more desirable to improve the recognition accuracy of the system than to increase the processing speed. In these cases, the original colors of the image (typically red, green, blue) may be retained and the image may be processed with a separate pass for each of the colors. It may even be desirable to perform color separation into additional layers (e.g., defining layers for six different colors) and processing the image for each of these layers to provide improved recognition at the expense of processing speed (which would be ⅙ as fast as when the colors were converted to a gray scale).

After the pre-processing of the image is complete, the image is input to the convolutional neural network (406). The convolutional neural network uses multiple convolutional layers to detect increasingly complex features in the image. These layers produce a set of feature maps which are upscaled and combined with skip layers to produce a corresponding set of maps that are the same size as the input image. Each of the image maps shows a “heat map” that roughly defines the boundaries of a corresponding type of feature within the original image. For instance, one of the maps shows where in the original image text has been detected, while another may show where a face, a signature, document background, image background, or another type of feature has been detected. In one embodiment, a text heat map identifies the boundaries of the individual text characters (i.e., the edges of the lines that form the characters) rather than identifying bounding boxes that surround entire words or sentences. This text heat map may comprise probability values for each pixel (where the value indicates the probability that the pixel is text), where the probability values are used to adjust the intensities of the pixels before OCR is performed by the convolutional neural network.

Referring to FIG. 5A, a diagram is provided to illustrate an input image 500 of a document such as a driver's license and the corresponding image maps showing areas in which faces and text are found. In this figure, the original image of the driver's license includes, for example, a picture of the driver (505), the state that issued the license (510), information associated with the driver (515), and the driver's signature (520).

Two exemplary heat maps that are produced by one embodiment of the convolutional neural network are depicted in the figure. A first one of the heat maps (525) shows the area (527) of the driver's license in which the driver's face is found. In this case, there is a larger photograph of the driver on the left side of the image, and a smaller photograph on the right side of the image. A second one of the heat maps (530) shows the areas or “bounding boxes” (532) of the license in which text is found. A third one of the heat maps (535) shows the specific boundaries of individual text characters. Other heat maps may also be generated to indicate the locations of other types of features, such as signatures, logos, security features, etc.

In one embodiment, the convolutional neural network is configured to generate a heat map that identifies the boundaries of individual characters, rather than identifying bounding boxes that enclose entire words or groups of words (see heat map 535 in FIG. 5A). This heat map is illustrated in more detail in FIG. 5B. FIG. 5B depicts the same driver's license image 500 shown in FIG. 5A, but only a portion of the heat map 535 is depicted on the right side of the figure, and this portion of the heat map is enlarged to show more detail. The enlarged portion of the heat map shows the boundaries of the individual characters. The pixels within the boundaries of the individual characters (e.g., 536) have higher probabilities, while the pixels outside the boundaries of the individual characters (e.g., 537) have lower probabilities. A heat map indicating probabilities that pixels are non-text (background) may also be generated. This heat map may be generally complementary to the text heat map. The character-level text heat map of FIG. 5B (and/or non-text heat map) may be generated in addition to the other heat maps, or individually.

Although the boundaries are clearly delineated in FIG. 5B for purposes of simplicity, it should be noted that the probabilities need not be binary representations. In other words, each pixel may have a probability that is in a range from 0 (0%) to 1 (100%). This is illustrated in FIG. 5C. FIG. 5C depicts, for a single row of the pixels on the image, the probability that each of these pixels is within the boundaries of a text character in the document image. Thus, the pixels in groups 540, 542 and 544 have higher probabilities of being text pixels, while the pixels in groups 550, 552, 554 and 556 have lower probabilities of being text pixels (i.e., higher probabilities of being non-text pixels).

The heat maps generated by the convolutional neural network may use colors, shades of gray, or other means to indicate a range of likelihoods of finding a particular feature type at specific locations within the image (e.g., blues and greens may represent lower likelihoods, while oranges and reds represent higher likelihoods). Alternatively, the heat maps may indicate areas that are bounded by distinct lines (bounding boxes), where the portion of the image within a bounding box has an above-threshold likelihood of having the particular feature type and the portion of the image outside the bounding box has a below-threshold likelihood of having the particular feature type. The threshold may be set by default to a predetermined value (e.g., 0.5 on a scale from 0 to 1), or it may be user-configurable. As noted above, while the heat maps depicted in FIGS. 5A and 5B show the image regions as simple lines (e.g., bounding boxes 532 or text area 536) dividing the relevant areas (e.g., text pixels) from other areas (e.g., non-text pixels), the heat maps may more typically indicate (e.g., via colors or shades of gray) the probability that each pixel in the map corresponds to the desired feature. As a non-limiting example, image regions indicated by the bounding boxes (e.g., 532) shown in FIG. 5A illustrate where these probabilities may exceed a threshold probability. It should also be noted that the heat maps need not be displayable, so the maps may simply comprise stored probability values.

Referring again to FIG. 4, the heat map for text in the image is used in some embodiments to determine the boundaries (e.g., bounding boxes) of the text in the image, and OCR is used to detect and/or identify the text within these boundaries (408). In one embodiment, the images are processed to increase the contrast between text pixels and non-text pixels before OCR processing is performed. The processing to increase the text contrast is performed after the convolutional neural network identifies which pixels form text components of the image, and which pixels form non-text (background) components of the image. The identification of text pixels and non-text pixels may comprise identifying, for each pixel, a corresponding probability that the pixel is text. Then, the intensity of each pixel may be adjusted by a variable amount that depends on the associated probability and the original intensity of the pixel. In one embodiment, the intensity-adjusted image may be processed to identify bounding boxes for text prior to performing OCR on the portions of the image within the bounding boxes.

Similarly, a heat map for facial images can be used to detect and/or identify a face in the image, a heat map for signatures can be used to detect and/or identify a signature in the image, and so on. In some embodiments, the detected/identified information may be used to classify the imaged document (e.g., as a driver's license, invoice, etc.).

The recognized characters in the text areas, or text bounding boxes, can then be processed to identify, derive, interpret, extract, and/or infer meaningful information within the text (410). For example, if the characters include “DOB” followed by a set of numbers, the numbers may be interpreted as a birthdate, or if the characters include “HAIR: BRN”, the characters may be interpreted to indicate that the associated person's hair is brown. This part of the process may be guided by the context corresponding to the classification of the document. For instance, if the image is classified as a driver's license, a set of numbers may be interpreted as a license number associated with a particular driver, rather than as an invoice number or some other type of number.

In some embodiments, the system may be configured to detect security features in the processed image. Many common documents have security features that are designed to allow verification of the authenticity of the document, prevent unauthorized copying of the document, or otherwise secure the document. For example, a driver's license will commonly have such features as holograms or background images which are visible when the license is copied and when viewed by the human eye. Checks commonly have both watermarks that are visible to the human eye and security backgrounds which may not visible to the human eye but, when copied, the word “VOID” appears in the image. It may be useful to recognize such security features in images that are processed by the present systems and methods.

For instance, in some embodiments, it may be desirable to detect and identify security features in an image in order to verify that the image represents an authentic document. In this case, the convolutional neural network may generate a heat map for the security feature so that the feature can be extracted, identified or otherwise verified by the system. In other embodiments, it may be helpful to identify a security feature or other type of background in order to be able to remove, minimize or otherwise account for this background so that text or other features in the document are more easily recognizable. In the example of the “VOID” security background of a check, the background image commonly overlaps with other features on the check, such as text indicating a payor, a payee, an amount, a routing number, etc. The overlapping security image can make it more difficult to identify and recognize the text with which it overlaps. Some embodiments may therefore be configured to generate a heat map for the background image, identify the background image within the area(s) shown in the heat map, and account for the background image when extracting other types of features from the areas that are overlapped by the background image. The system may thereby increase the accuracy of recognition of the features that are overlapped by background features in an input image.

In some embodiments, it may be desirable to minimize the impact of security features, images or other background elements in a document image rather than detecting and identifying these elements. Diminishing the impact of these features may increase the effectiveness of text recognition algorithms that are applied to the image. This may be achieved, for example, by identifying both text and non-text elements in the image, and then increasing the contrast between the text and non-text elements. In one embodiment, a text heat map is generated, so that each pixel in the image has an associated probability that the pixel is text. Pixels that form parts of textual characters have higher probabilities, and pixels that form part of the background have lower probabilities. Based on the values in this heat map, the intensities (lightness and darkness) of the pixels are adjusted to increase the contrast of the text against the background. OCR can then be performed on the increased-contrast document image with increased efficiency, and data can be extracted from the recognized characters.

An exemplary method for increasing the contrast of the document image is illustrated in FIG. 6, which are explained in detail below. This method begins with receipt of an input document image (605). The image will be processed by the convolutional neural network, but the original image may not have the exact dimensions that are required by the convolutional neural network. For instance, the convolutional neural network may be configured to process images that are 1000 pixels by 1000 pixels, but the image may be 3000 pixels by 3000 pixels. If the input image is larger or smaller than the required size (610), the image is scaled to the required size (615). In this example, the 3000-by-3000 image would be scaled down to 1000-by-1000. If the input image is not square, it may be scaled by different factors in the horizontal and vertical directions (which may alternatively be referred to as length and width). (It should be noted that alternative embodiments may handle the size mismatch in different ways, such as, for example, processing the original 3000-by-3000 image in nine 1000-by-1000 chunks.)

After the input image has been scaled to the required input size (if necessary), the scaled image is processed by the convolutional neural network to identify the text and non-text portions of the image (620). In this embodiment, the text and non-text portions of the image are identified on a pixel-by-pixel basis. That is, each pixel is separately identified as being a text pixel or a non-text pixel. More specifically, each pixel is examined, and a value is assigned to the pixel, where the value identifies a probability that the pixel is a text pixel. Alternatively, the value may identify a probability that the pixel is a non-text pixel. For instance, if a pixel has a probability value of 0.8, this may indicate that there is an 80% probability that the pixel is a text pixel. The probability that the pixel is a non-text pixel may be the complement of the text probability (i.e., for a text probability pt, the non-text probability pnt may be 1−pt), so there is a 20% probability that the pixel is a non-text pixel. The probability values for all of the pixels in the scaled input image collectively form a probability map (or heat map) for text in the image.

If the original input image had to be scaled before being input to the convolutional neural network (625), the probability map is then scaled back to the size of the original input image (630). In the example above, the generated text probability map would be 1000 pixels by 1000 pixels, and this map would be scaled up to 3000 pixels by 3000 pixels to match the original input image size. The pixels of the scaled probability map may be interpolated or generated in any suitable manner to achieve the desired scaling. Alternatively, the individual pixel values of the generated probability map may be used to produce a substantially continuous probability function which can then be scaled as needed to match the size of the original input image.

The scaled probability map is then used to apply intensity corrections to the original input image, producing a modified, intensity-corrected image (635). In one embodiment, the degree to which the pixels are lightened or darkened is dependent upon the degree of certainty with which the pixels are determined to be text or non-text. In other words, if a pixel is very likely to be text (e.g., 90% probability), it will be darkened to a greater degree (for dark text on a light background), but if the pixel is closer to the text/non-text boundary (e.g., it has a 50% probability of being text) it may remain largely unchanged. Similarly, if a pixel is very likely to be non-text (e.g., 10% probability of being text), it will be lightened to a greater degree (for dark text on a light background), but if the pixel is closer to the text/non-text boundary, it may be subject to little or no adjustment.

In one embodiment, the process of adjusting the intensities of the pixels involves a conversion of the pixel from an RGB representation to an HSL representation. In an RBG (Red, Green, Blue) system for representing color, each pixel has a separate intensity value for red, green and blue. In an HSL (Hue, Saturation, Lightness) system, the color of the pixel includes a hue value (a degree on the color wheel) a saturation percentage, and a lightness percentage. There are known methods for making conversions between these systems. In this case, the advantage of the HSL system is that only one of the values—lightness—needs to be modified to change the intensity of the pixel. After the intensity is adjusted, the pixel can be converted from HSL back to RGB. Alternatively, if the input image originally uses the HSL representation, no conversion is necessary. RGB representation of pixel colors can also be modified to adjust their intensities without necessarily converting to an HSL representation.

Many different functions and/or algorithms may be implemented to determine the specific adjustments to the pixels. In one example, the adjustment includes a factor corresponding to the probability and a factor corresponding to the intensity of the pixel. This can be expressed (for dark text on a light background) as:
if pt≥0.5, Δi=(0.5−pt)*(i−imin)  (1)
if pt≤0.5, Δi=(0.5−pt)*(imax−i)  (2)

where pt is the probability that a pixel is text (0<pt<1), Δi is the intensity adjustment for the pixel, i is the initial intensity of the pixel, imin is the minimum intensity value, and imax is the maximum intensity value. Thus, pixels that have a higher than 50% probability of being text (pt>0.5) have their intensities reduced (Δi<0) to darken them, and pixels that have a probability of less than 50% of being text (pt<0.5) have their intensities increased (Δi>0) to lighten them. The magnitude of the change for a given pixel is linearly dependent on the probability associated with that pixel, and the initial intensity, i, of the pixel. Equations (1) and (2) adjust the intensity of a 100%-probability text pixel to half of its original intensity, and adjust the intensity of a 100%-probability non-text pixel half-way from its original intensity to the maximum intensity. An additional factor can included to increase or decrease the magnitude of the adjustment.

It should be noted that equations (1) and (2) apply to dark text on a light background. It is therefore desirable to darken the text pixels and lighten the non-text pixels. The input document image, however, may also (or alternatively) include light text on a dark background. For the purposes of this disclosure, darker text on a lighter background may be referred to as “normal” or “positive” text, and lighter text on a darker background may be referred to as “inverted” or “negative” text. In the case of inverted text, the adjustment to the intensity of the pixel can be expressed as:
if pt≥0.5, Δi=(pt−0.5)*(imax−i)  (3)
if pt≤0.5, Δi=(pt−0.5)*(i−imin)  (4)

where pt is the probability that a pixel is text (0<pt<1), Δi is the intensity adjustment for the pixel, i is the initial intensity of the pixel, imin is the minimum intensity value, and imax is the maximum intensity value. Because algorithms for recognition of text characters typically detect the boundaries of the characters against the background, they can generally distinguish between text and background, regardless of whether the text is normal or inverted.

The adjustment of the pixel intensities may be more effective than prior art image processing in improving the effectiveness of OCR processes for several reasons. For instance, because the intensities of the text and non-text pixels are adjusted rather than removing background pixels (setting them to a maximum intensity) and leaving text pixels, the contrast of the text relative to the background is increased without creating the drastic changes in intensity that can confuse OCR algorithms. Because all the pixels remain in the image, even if lightened, a text pixel that was incorrectly assessed as a background pixel can still be used by the OCR engine, improving the OCR quality. Further, rather than considering each pixel in a binary fashion to be either background or text, the probability that the pixel is text is calculated as a continuous value in the range from 0 (definitely background) to 1 (definitely text). Since this probability value for each pixel is used to adjust the degree of lightening or darkening of the pixel, pixels near the text-background boundary (those likely to have a text probability of around 0.5) will be largely unchanged. This produces a smoother transition from likely-background to likely-text pixels, which further improving the OCR quality.

As noted above, the document image may include positive text (darker text on a lighter background) and/or inverted text (lighter text on a darker background). These two different types of text may be handled in several ways. In one embodiment, the convolutional neural network may generate four heat maps: dark text; light text; dark background; and light background. Then, pixels that are identified as dark text or dark background can be darkened (their intensities reduced), and pixels identified as light text or light background can be lightened. In another embodiment, a single heat map identifying text pixels can be generated. Then, pixels identified as text pixels can be either darkened or lightened, depending upon whether they are darker or lighter than the neighboring non-text pixels. If the text pixels are darker than the neighboring non-text pixels, the text pixels are darkened and the neighboring non-text pixels are lightened. If the text pixels are lighter than the neighboring non-text pixels, the text pixels are lightened and the neighboring non-text pixels are darkened. In yet another embodiment, heat maps identifying dark text and inverted text are generated, and the text identified by the inverted text heat map is itself inverted so that it is positive text. The pixels identified as text pixels can then all be darkened, and all of the pixels identified as non-text (background) pixels can be lightened. Still other alternatives can also be implemented.

It should be noted that “neighboring” pixels, as used herein, refers to pixels that are in a region adjacent to a region containing the pixels being considered. For example, if the pixels being considered form a portion of a part text character, the neighboring pixels would generally be the pixels in a background region immediately adjacent to the formed portion of the text character. The size of the region of the neighboring pixels may vary, but will typically be in the same order of magnitude as the region of the text character.

As noted above, the main part of the extraction service is implemented in a convolutional neural network. In one embodiment, the lightening and/or darkening of the input image are implemented using the text probability heat map generated by the convolutional neural network. Conventional convolutional neural networks that are used to recognize information within images have a number of convolutional layers that are followed by a fully connected layer (see Y. LeCun, et al., “Gradient-Based Learning Applied to Document Recognition”, Proc. Of the IEEE, November 1998). The convolutional layers identify features within the image, and the fully connected layer identifies which of a set of possibilities is most likely represented by the image. In the embodiments disclosed herein, on the other hand, the convolutional neural network is fully convolutional—it has no fully connected layer. Consequently, it does not produce a unique probability for each classification of an image, but instead produces a set of maps that indicate the image regions in which different types of information are found.

Referring to FIG. 7, a diagrammatic representation of the topology of a non-limiting example of a convolutional neural network according to one embodiment is shown. In the example of FIG. 7, each operation is represented by a resulting feature map. The image 602 that is input to the convolutional neural network is shown at the left side of the figure. The image is processed from left to right in the figure. In this example, the convolutional neural network has three parts.

The first part of the convolutional neural network can include a series of layers that perform operations including convolution operations, average pooling operations, max pooling operations layers, and the like to identify relevant features in the image. These convolutional layers produce a set of feature maps that are smaller than the original input image, which in this example can be 1300 by 800 pixels. Characteristics of each of the feature maps in the first part of the convolutional neural network are shown in Table 1 below. These characteristics include height, width, kernel size and number of features.

TABLE 1
Layers 702 703 704-710 712 714-720 722 724-726 728 730-732 734 736-738
Width 1300 650 650 325 325 163 163 82 82 41 41
Height 800 400 400 200 200 100 100 50 50 25 25
Kernel 5x5 5x5 3x3 1x1 1x1
Features 1 1 N N 2xN 2xN 4xN 4xN 8xN 8xN 16xN

In the second part of the convolutional neural network, these reduced-size feature maps are used to generate a set of heat maps, each of which corresponds to the likelihood that a particular type of feature (e.g., text, image, etc.) is present in a specific part of the image. These heat maps can then be processed by a third part of the convolutional neural network via a series of convolution transpose layers and skip layers to upscale the heat maps to the size of the input image. Ultimately, the convolutional neural network produces a final set of heat maps (760) that are the same size as the original input image, where each of the heat maps shows the probability that a feature of a particular type (e.g., text or an image) is found at each pixel.

In the example of FIG. 7, the first part of the convolutional neural network can have five convolutional layers. The first operation performed on the input image is an average pool (703), where each set of 2×2 pixels is averaged, thereby producing an image that is half the size of the input image (650×400 in this example). The reduced-size image then undergoes a first layer of convolutions that has a stack of four operations (704-710), where each of the operations includes a convolution, a batch normalization (see https://www.tensorflow.org/api_docs/python/tf/nn/batch_normalization) and a rectified linear unit (RELU). The convolutions may, in one embodiment, use a 5×5 kernel to produce feature maps that are 650×400 in size and have a number, N, of features.

After the stack of operations in the first layer has been completed, the resulting feature maps are processed by a second convolutional layer. In this layer, a first max pool operation (712) is performed. The max pool operation results in a set of feature maps that are 325×200 pixels in size and have 2N (2*N) features. These feature maps again undergo a stack of four operations (714-720), where each of the operations includes a convolution, a batch normalization and a rectified linear unit. In this embodiment, the convolutions again use a 5×5 kernel. The operations produce feature maps that are 325×200 in size and have 2N features.

After the operations in the second layer has been performed, the resulting feature maps are processed by a third convolutional layer that includes a second max pool operation (722) that results in a set of feature maps which are 163×100 pixels in size and have 4N features. The resulting feature maps are then processed by a stack of two operations (724-726), where each of the operations again includes a convolution, a batch normalization and a rectified linear unit. In this layer, the convolutions use a 3×3 kernel. The operations produce feature maps that are 163×100 in size and have 4N features.

Following completion of the third of convolutions, the resulting feature maps are processed by a fourth convolutional layer that begins with a third max pool operation (728). This max pool operation produces a set of feature maps which are 82×50 in size and have 8N features. This set of feature maps is then processed by a stack of two operations 730-732), where each of the operations includes a convolution, a batch normalization and a rectified linear unit. In this layer, the convolutions use a 1×1 kernel, and the resulting feature maps are 82×50 in size, with 8N features.

After the fourth layer of convolutions is completed, the feature maps are processed by a fifth convolutional layer. The first operation in this layer is a fourth max pool (734). This max pool operation produces a set of feature maps which are 41×25 in size and have 16N features. These feature maps undergo a stack of two operations (736-738), where each of the operations includes a convolution, a batch normalization and a rectified linear unit. In this layer, the convolutions use a 1×1 kernel. The resulting feature maps are 41×25 in size, and have 16N features.

The second part of the convolutional neural network is a single convolution operation which produces output layer 740. This operation includes a convolution and a batch normalization. The convolution is performed with a 1×1 kernel and produces a set of heat maps from the feature maps. In one embodiment, the 16N features of the feature maps are used to identify five types of features—text, face, signature, document background and image background.

Because it is more accurate to have heat maps that are the same size as the input image, the 41×25-pixel heat maps can be upscaled in the third part of the convolutional neural network (as optimized by the network) to the size of the input image. This process can be accomplished through the use of four skip layers, and a final upscale layer. Each of the skip layers can include a transposed convolution of the previous feature maps, which is upscaled by a factor of 2. The feature maps from a previous layer go through a convolution layer with a 1×1×5 kernel, and are summed with the output of the upscale layer. (It should be noted that, in the example of FIG. 7, each of the operations in the third part of the convolutional neural network are performed on the images for each of the 5 features types.) For instance, in the first skip layer, a convolution transpose of 41×25 heat map 740 produces 82×50 map 742. This map (742) can then be added to map 732 from the fourth convolution layer to produce 82×50 map 744. This map includes the heat map information of map 742 and the finer-grained map information contained in map 732.

In the second skip layer, a convolution transpose 746 of output map 744 of the first skip layer, upscaled by 2, can be added to map 726 from the third convolution layer to produce map 748. This produces a map having a size of 163×100, which again has the heat map information of map 746 and the finer-grained map information contained in map 726. Similarly, in the third skip layer, a convolution transpose 750 of the output map of the second skip layer, upscaled by 2, can be added to map 720 from the second convolution layer to produce map 752, and in the fourth skip layer, a convolution transpose 754 of the output map of the third skip layer, upscaled by 2, can be added to map 710 from the first convolution layer to produce map 756.

The output of the fourth skip layer (map 756) is still smaller than the input image. (In the example of FIG. 7, it is 650×400 pixels, as compared to 1300×800 pixels for the input image). The convolutional neural network of FIG. 7 can therefore include an upscale operation to produce output layer 758 which is effectively the reverse of the initial average pool operation. Finally, a softmax operation can be performed to produce output layer 760 so that the final heat maps will show probability distributions of the respective features, rather than raw data.

While the fully convolutional neural network depicted in FIG. 7 is unique, it is still necessary to train the network. Training of convolutional neural networks can be performed by processing images and comparing the actual results of the processing (e.g., identification of text in the image) with expected outputs. Differences between the actual and expected outputs are then used to adjust the weights used in the convolution layers. The quality of the output of the convolutional neural network can improve with more training.

Traditionally, training has been performed using input images for which the expected output is determined through means other than the convolutional neural network. For instance, a person might view an image, identify text in the image, and then record the text as the expected output of processing by the network. Due in part to the need to determine the expected output for each image, it has traditionally been difficult to provide large amounts of training data. Embodiments disclosed herein may include methods for providing much larger sets of training data than were traditionally available.

Referring to FIG. 8, a flow chart illustrating an example of a method for generating a training data set according to some embodiments is shown. As depicted in this figure, the method begins with the generation of a simulated document (802). In a convolutional neural network that is designed to recognize features of a driver's license, an input image may be expected to have corresponding image elements. For instance, the image may include a photo of the driver's face, the driver's signature, text corresponding to some of the driver's identifying characteristics (e.g., height, weight, hair color, birthdate, etc.), text corresponding to the license itself (e.g., issuing state, license number, etc.), document background (e.g., seals, logos, security features, etc.), and various other features. These elements may be arranged in any of a set of predetermined formats that are used by the respective issuing states. Generation of a simulated document can entail selecting a set of the document components (each of which may be fictitious and/or randomly generated) and plugging the components into one of the predetermined formats. The resulting dynamically generated simulated document therefore can have the same characteristics (photo, signature, text, etc.) as a genuine license, but does not include the private information of any real individual. (It should be noted that “dynamically generated” is used herein to indicate that the simulated document did not exist until it was created from the selected document components.)

Based on the simulated document, a simulated document image is generated (804). This image is a representation of the simulated document, which in this embodiment doesn't actually exist—it was instead generated from the document components and exists in the system's memory. The simulated document image is processed in this embodiment to include one or more of the images variations that are found in typical input images (706). For instance, the simulated image may be skewed so that the document (text) is angled with respect to horizontal, it may be tilted so that the scale of the document changes across the image (e.g., top to bottom or side to side), the document may be offset from the center of the image, the lighting may be made uneven across the image, or various other filters may be applied to the image to mimic effects that are found in typical input images.

It should be noted that, in addition to changing the simulated document image with image variations such as tilt, skew, centering, lighting, etc., variations may be introduced into the simulated document prior to the generation of the simulated document image through the generation of the simulated document using different fonts, colors and sizes for the document components. The use of different fonts, colors, sizes, etc. in the generation of the simulated documents will provide more variation in the corresponding images so that, when the convolutional neural network is trained with these images, it will help the system to perform even more robust detection.

Referring to FIG. 9, a diagram is provided to illustrate the creation of a simulated document image in accordance with one embodiment. As depicted in this figure, a set of document components 902 are initially stored in a memory or other data storage device, although they may be provided in some other manner in alternative embodiments. The set of document components includes components of various different types, such as names, addresses, photographs, document formats, signatures, and the like. There are multiple components within each of these component types. For example, within the category of photographs, there may be many different individual photograph document components. Similarly, there may be many different name document components, many different address document components, etc.

In step 802 of FIG. 8, a set of the document components is selected and used to generate a simulated document. The simulated document is shown as element 904 in FIG. 9. It can be seen that simulated document 904 includes one of the photograph document components (“photoi”), one of the name document components (“namej”), one of the address document components (“addressk”), and so on. Each of the selected document components is selected from the set of document components within the corresponding document component type. The particular document component from each type may be selected randomly, or in some ordered fashion. The simulated document will normally have format associated with it. For example, the simulated document may have the format of a California driver's license, or a New York driver's license, or some other type of document. The format may be one of the document components, and may consequently vary from one simulated document to the next, or the simulated documents may be generated according to one or more predetermined formats. Because the simulated document is not a tangible physical document, it may exist in various forms that allow it to be stored (in transitory or persistent form) and processed. For example, in one embodiment the simulated document may simply be the defined list of document components that have been selected to form the document (e.g., photographi, namej, addressk, etc.).

Referring again to FIG. 9, the simulated document is then used to generate a simulated document image 906. Simulated document image may have any suitable image format, such as a JPEG, PNG, BMP, GIF, PDF, etc. The image format of the simulated document image will typically have the same format as the input images that are expected to be processed by the convolutional neural network after it has been trained, and the convolutional neural network will process the simulated document image and the expected input images in the same manner. In the example of FIG. 9, simulated document image 906 depicts the selected document components according to the selected format, but does not include image variations of the type that are commonly found in images that are input to the system for recognition. Simulated document image 906 is processed to incorporate some of these variations and thereby produce simulated document image 908. In simulated document image 908, the simulated document is slightly rotated and is offset from the center of the image. Although simulated document image 906 may be used in the training of the convolutional neural network, the trained convolutional neural network will normally be more robust if there is more variation in the training data set, so use of the simulated document image with the variations is preferred. It should be noted that it is not necessary to generate an unaltered image (e.g., image 906) prior to incorporating the image variations—the variations may be incorporated into the image as it is originally generated, producing the final image depicted as 908.

Referring to FIGS. 10A-10D, several examples of the types of variations that may be incorporated into the simulated document images are shown. Each of these figures depicts an image of a driver's license.

In FIG. 10A, the portion of the image depicting the driver's license is distorted as if the license were tilted when photographed, so that the opposing edges of the license are not parallel to each other. Consequently, the lines of text on the license will not be parallel, and the images and characters will be slightly distorted. In FIG. 10B, the lighting is uneven across the image (similar to having a shadow on part of the image). These figures illustrate just a few of the types of images variations that may be present in input images, and which may be incorporated into the simulated document images. In FIG. 10C, the portion of the image depicting the driver's license is slightly skewed or rotated, so that horizontal lines of text on the license are not horizontal in the generated image. In FIG. 10D, the portion of the image depicting the driver's license is offset from the center of the generated image.

Referring again to FIG. 8, as the simulated images are generated, they may be stored for use as a training data set. The information corresponding to the document components from which the simulated documents and simulated document images are generated may also be stored as part of the training data set. This data set is used to train the convolutional neural network (808). As an alternative to storing the simulated document images and the associated data, this information may be provided directly to the convolutional neural network to train it. During the training process, each simulated document image is processed by the convolutional neural network to recognize areas of the image in which there is a likelihood that particular features are found. The convolutional neural network may produce “heat maps” showing the relative likelihood (e.g., expressed as a percentage) of finding the particular features in specific areas of the image. Alternatively, the convolutional neural network may identify bounding boxes for the respective features, where the likelihood of finding a feature within a bounding box is greater than or equal to a defined threshold value, and the likelihood of finding the features outside the bounding box is less than the defined threshold value. The “boxes” may be irregularly shaped, and may include multiple enclosed areas.

The convolutional neural network may be configured simply to generate the heat maps or bounding boxes, or it may be further configured to extract information from the identified areas within the images. For instance, the convolutional neural network may extract text from areas that are identified as containing text, or it may extract sub-images (e.g. a photograph of a license owner) from the processed simulated document image. In either case, the output produced by the convolutional neural network is compared to a corresponding expected output, and the parameters of the convolutional neural network are adjusted as needed based on the comparison. For example, a generated heat map may be compared to a corresponding expected feature map, or a generated address may be compared to an expected address. If the output produced by the convolutional neural network is very close to the expected output, the adjustment to the parameters may be very small. If, on the other hand, the difference is greater, the adjustment to the parameters may likewise be greater. The adjustment of the parameters may be configurable by a user, as is known in the art.

There are a variety of methodologies and corresponding parameters that can be used to train the network. In one embodiment, the batch size is one image, the loss uses a weighted average (see J. Long, et al., “Fully Convolutional Networks for Semantic Segmentation,”) and the learning rate is 0.001. An Adam optimizer (see D. Kingma, et al., “Adam: A Method for Stochastic Optimization,” ICLR 2015) is used in the training of the convolutional neural network.

Because of the variables in the information that is used in the simulated documents, it is possible to generate a large number of documents from a very small amount of information. For instance, if only ten names, ten photos, and ten license formats are used, 1000 different simulated documents can be generated. With more variables (e.g., height, weight, hair color, birthdate, etc.) and more values for each of these variables, the number of different synthetic documents that can be generated for the training set rapidly increases. This number increases even further when the image variations such as varying fonts, colors, sizes, lighting, skew, tilt, etc. (as opposed to document information variations) are added to the images of the synthetic documents. The present embodiments therefore enable robust training of a convolutional neural network by someone who does not have access to large amounts of actual data (images of pre-existing documents and corresponding document information).

It should be noted that, although many variations of the simulated documents and images are possible, it may be preferable not to generate all of the variations that are possible from a given set of document components. This is because some of the variations may be small enough that the resulting images form a less effective training set than a comparable number of images that have greater variations between them. In other words, a training data set that has greater variations will train a neural network to recognize input images with greater variations, so the neural network will be more robust. For example, consider a first training data set that includes ten simulated images which are identical except for the names of the fictitious people associated with the images, and a second training data set that includes ten simulated images, each of which has a unique name, address, driver's license number, photograph and image variation (tilt, skew, etc.). Training a neural network with the second set will, in general, enable the neural network to recognize a wider variety of input images than if the neural network were trained with the first data set.

The methods disclosed herein provide other benefits in addition to the advantage of being able to generate large numbers of training images. One of these advantages is that it is not necessary to provide any means to determine the expected outputs which result from recognizing these documents—the expected outputs (information such as name, photo, signature, etc.) are known even before the images are generated, since this is the information that is used to generate the simulated images. By comparison, training a neural network with images of real, pre-existing documents requires that the information which is contained in the images be determined and provided to the neural network as expected outputs. Determining this information to be input to the system for training the convolutional neural network can itself be a time consuming and costly process (apart from obtaining and processing the images which are separately input to the convolutional neural network as part of the training process). It is therefore less costly to produce the training data set based on the dynamically generated, simulated documents and images as disclosed herein. Further, because the expected outputs are actually used to generate the images, the expected outputs are known with certainty, which may not be true of information extracted from actual, pre-existing documents.

The images generated using the methodology described above provide a training data set that includes a wide variety of image variations that might not be found in a set of collected images of actual documents. When the convolutional neural network is trained using this synthetic training data set, the network is more robust than networks that are trained using a captured training data set, and is better able to handle a wide range of document types and environments, and to handle unexpected variations in input images. It should be noted that, even though the convolutional neural network may have robust training through the use of this initial synthetic training data set, the network may continue to be trained as it processes actual images that are not part of a training data set.

Those skilled in the relevant art will appreciate that the invention can be implemented or practiced with other computer system configurations, including without limitation multi-processor systems, network devices, mini-computers, mainframe computers, data processors, and the like. The invention can be embodied in a computer or data processor that is specifically programmed, configured, or constructed to perform the functions described in detail herein. The invention can also be employed in distributed computing environments, where tasks or modules are performed by remote processing devices, which are linked through a communications network such as a local area network (LAN), wide area network (WAN), and/or the Internet. In a distributed computing environment, program modules or subroutines may be located in both local and remote memory storage devices. These program modules or subroutines may, for example, be stored or distributed on computer-readable media, including magnetic and optically readable and removable computer discs, stored as firmware in chips, as well as distributed electronically over the Internet or over other networks (including wireless networks). Example chips may include Electrically Erasable Programmable Read-Only Memory (EEPROM) chips. Embodiments discussed herein can be implemented in suitable instructions that may reside on a non-transitory computer readable medium, hardware circuitry or the like, or any combination and that may be translatable by one or more server machines. Examples of a non-transitory computer readable medium are provided below in this disclosure.

ROM, RAM, and HD are computer memories for storing computer-executable instructions executable by the CPU or capable of being compiled or interpreted to be executable by the CPU. Suitable computer-executable instructions may reside on a computer readable medium (e.g., ROM, RAM, and/or HD), hardware circuitry or the like, or any combination thereof. Within this disclosure, the term “computer readable medium” is not limited to ROM, RAM, and HD and can include any type of data storage medium that can be read by a processor. Examples of computer-readable storage media can include, but are not limited to, volatile and non-volatile computer memories and storage devices such as random access memories, read-only memories, hard drives, data cartridges, direct access storage device arrays, magnetic tapes, floppy diskettes, flash memory drives, optical data storage devices, compact-disc read-only memories, and other appropriate computer memories and data storage devices. Thus, a computer-readable medium may refer to a data cartridge, a data backup magnetic tape, a floppy diskette, a flash memory drive, an optical data storage drive, a CD-ROM, ROM, RAM, HD, or the like.

The processes described herein may be implemented in suitable computer-executable instructions that may reside on a computer readable medium (for example, a disk, CD-ROM, a memory, etc.). Alternatively, the computer-executable instructions may be stored as software code components on a direct access storage device array, magnetic tape, floppy diskette, optical storage device, or other appropriate computer-readable medium or storage device.

Any suitable programming language can be used to implement the routines, methods or programs of embodiments of the invention described herein, including C, C++, Java, JavaScript, HTML, or any other programming or scripting code, etc. Other software/hardware/network architectures may be used. For example, the functions of the disclosed embodiments may be implemented on one computer or shared/distributed among two or more computers in or across a network. Communications between computers implementing embodiments can be accomplished using any electronic, optical, radio frequency signals, or other suitable methods and tools of communication in compliance with known network protocols.

Different programming techniques can be employed such as procedural or object oriented. Any particular routine can execute on a single computer processing device or multiple computer processing devices, a single computer processor or multiple computer processors. Data may be stored in a single storage medium or distributed through multiple storage media, and may reside in a single database or multiple databases (or other data storage techniques). Although the steps, operations, or computations may be presented in a specific order, this order may be changed in different embodiments. In some embodiments, to the extent multiple steps are shown as sequential in this specification, some combination of such steps in alternative embodiments may be performed at the same time. The sequence of operations described herein can be interrupted, suspended, or otherwise controlled by another process, such as an operating system, kernel, etc. The routines can operate in an operating system environment or as stand-alone routines. Functions, routines, methods, steps and operations described herein can be performed in hardware, software, firmware or any combination thereof.

Embodiments described herein can be implemented in the form of control logic in software or hardware or a combination of both. The control logic may be stored in an information storage medium, such as a computer-readable medium, as a plurality of instructions adapted to direct an information processing device to perform a set of steps disclosed in the various embodiments. Based on the disclosure and teachings provided herein, a person of ordinary skill in the art will appreciate other ways and/or methods to implement the invention.

It is also within the spirit and scope of the invention to implement in software programming or code any of the steps, operations, methods, routines or portions thereof described herein, where such software programming or code can be stored in a computer-readable medium and can be operated on by a processor to permit a computer to perform any of the steps, operations, methods, routines or portions thereof described herein. The invention may be implemented by using software programming or code in one or more digital computers, by using application specific integrated circuits, programmable logic devices, field programmable gate arrays, optical, chemical, biological, quantum or nanoengineered systems, components and mechanisms may be used. The functions of the invention can be achieved by distributed or networked systems. Communication or transfer (or otherwise moving from one place to another) of data may be wired, wireless, or by any other means.

A “computer-readable medium” may be any medium that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, system or device. The computer readable medium can be, by way of example only but not by limitation, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, system, device, propagation medium, or computer memory. Such computer-readable medium shall generally be machine readable and include software programming or code that can be human readable (e.g., source code) or machine readable (e.g., object code). Examples of non-transitory computer-readable media can include random access memories, read-only memories, hard drives, data cartridges, magnetic tapes, floppy diskettes, flash memory drives, optical data storage devices, compact-disc read-only memories, and other appropriate computer memories and data storage devices. In an illustrative embodiment, some or all of the software components may reside on a single server computer or on any combination of separate server computers. As one skilled in the art can appreciate, a computer program product implementing an embodiment disclosed herein may comprise one or more non-transitory computer readable media storing computer instructions translatable by one or more processors in a computing environment.

A “processor” includes any, hardware system, mechanism or component that processes data, signals or other information. A processor can include a system with a central processing unit, multiple processing units, dedicated circuitry for achieving functionality, or other systems. Processing need not be limited to a geographic location, or have temporal limitations. For example, a processor can perform its functions in “real-time,” “offline,” in a “batch mode,” etc. Portions of processing can be performed at different times and at different locations, by different (or the same) processing systems.

As used herein, the terms “comprises,” “comprising,” “includes,” “including,” “has,” “having,” or any other variation thereof, are intended to cover a non-exclusive inclusion. For example, a process, product, article, or apparatus that comprises a list of elements is not necessarily limited only those elements but may include other elements not expressly listed or inherent to such process, product, article, or apparatus.

Furthermore, the term “or” as used herein is generally intended to mean “and/or” unless otherwise indicated. For example, a condition A or B is satisfied by any one of the following: A is true (or present) and B is false (or not present), A is false (or not present) and B is true (or present), and both A and B are true (or present). As used herein, a term preceded by “a” or “an” (and “the” when antecedent basis is “a” or “an”) includes both singular and plural of such term, unless clearly indicated otherwise (i.e., that the reference “a” or “an” clearly indicates only the singular or only the plural). Also, as used in the description herein, the meaning of “in” includes “in” and “on” unless the context clearly dictates otherwise.

It will also be appreciated that one or more of the elements depicted in the drawings/figures can also be implemented in a more separated or integrated manner, or even removed or rendered as inoperable in certain cases, as is useful in accordance with a particular application. Additionally, any signal arrows in the drawings/figures should be considered only as exemplary, and not limiting, unless otherwise specifically noted. The scope of the present disclosure should be determined by the following claims and their legal equivalents.

Lund, Christopher Dale, Samala, Sreelatha

Patent Priority Assignee Title
11625810, Jul 17 2017 Open Text Corporation Systems and methods for image modification and image based content capture and extraction in neural networks
11893777, Jul 17 2017 Open Text Corporation Systems and methods for image based content capture and extraction utilizing deep learning neural network and bounding box detection training techniques
Patent Priority Assignee Title
10007863, Jun 05 2015 CITIBANK, N A Logo recognition in images and videos
10445569, Aug 30 2016 A9.COM, INC.; A9 COM, INC Combination of heterogeneous recognizer for image-based character recognition
20170140236,
20170345140,
20180005082,
20180177461,
20180182086,
20180307942,
20190019020,
20190019021,
20190228529,
///
Executed onAssignorAssigneeConveyanceFrameReelDoc
Dec 17 2018SAMALA, SREELATHAOpen Text CorporationASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0482630547 pdf
Dec 19 2018LUND, CHRISTOPHER DALEOpen Text CorporationASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0482630547 pdf
Dec 21 2018Open Text Corporation(assignment on the face of the patent)
Date Maintenance Fee Events
Dec 21 2018BIG: Entity status set to Undiscounted (note the period is included in the code).
Feb 28 2024M1551: Payment of Maintenance Fee, 4th Year, Large Entity.


Date Maintenance Schedule
Sep 15 20234 years fee payment window open
Mar 15 20246 months grace period start (w surcharge)
Sep 15 2024patent expiry (for year 4)
Sep 15 20262 years to revive unintentionally abandoned end. (for year 4)
Sep 15 20278 years fee payment window open
Mar 15 20286 months grace period start (w surcharge)
Sep 15 2028patent expiry (for year 8)
Sep 15 20302 years to revive unintentionally abandoned end. (for year 8)
Sep 15 203112 years fee payment window open
Mar 15 20326 months grace period start (w surcharge)
Sep 15 2032patent expiry (for year 12)
Sep 15 20342 years to revive unintentionally abandoned end. (for year 12)