An image processing apparatus includes an image information acquisition device, an embedment information acquisition device, an embedment mode information acquisition device, an embeddability determination device, and an embeddability information output device. The image information acquisition device acquires image information on a target image into which supplemental information is to be embedded. The embedment information acquisition device acquires embedment information on the supplemental information. The embedment mode information acquisition device acquires embedment mode information on an embedment mode in which the supplemental information is embedded in the image information. The embeddability determination device determines embeddability of the supplemental information into the image information based on the embedment mode information, the embedment information, and the image information. The embeddability information output device outputs determination-result information on a determination result of the embeddability determined by the embeddability determination device.
|
13. A method of image processing, comprising:
acquiring image information on a target image into which supplemental information is to be embedded;
acquiring embedment information on the supplemental information;
acquiring mode information on an embedment mode in which the supplemental information is embedded in the image information;
determining embeddability of the supplemental information into the image information based on the acquired embedment mode information, the acquired embedment information, and the acquired image information;
acquiring embeddable volume information on a volume of information embeddable in the image information based on at least a total information volume designated as embedment information; and
outputting determination-result information on a determination result of the embeddability,
wherein the determining includes determining the embeddability of the supplemental information into the image information based on a comparison between a volume of the supplemental information to be embedded in the image information and the volume of information embeddable in the image information.
1. An image processing apparatus, comprising:
an image information acquisition device to acquire image information on a target image into which supplemental information is to be embedded;
an embedment information acquisition device to acquire embedment information on the supplemental information;
an embedment mode information acquisition device to acquire embedment mode information on an embedment mode in which the supplemental information is embedded in the image information;
an embeddability determination device to determine embeddability of the supplemental information into the image information based on the embedment mode information acquired by the embedment mode information acquisition device, the embedment information acquired by the embedment information acquisition device, and the image information acquired by the image information acquisition device; and
an embeddability information output device to output determination-result information on a determination result of the embeddability determined by the embeddability determination device,
wherein the embeddability determination device includes an embeddable volume acquisition unit to acquire embeddable volume information on a volume of information embeddable in the image information based on at least a total information volume designated as embedment information, and
wherein, based on a comparison between a volume of the supplemental information to be embedded in the image information and the volume of information embeddable in the image information, the embeddability determination device determines the embeddability of the supplemental information into the image information.
8. An image processing system comprising an image processing apparatus,
the image processing apparatus including
an image information acquisition device to acquire image information on a target image into which supplemental information is to be embedded,
an embedment information acquisition device to acquire embedment information on the supplemental information,
an embedment mode information acquisition device to acquire embedment mode information on an embedment mode in which the supplemental information is embedded in the image information,
an embeddability determination device to determine embeddability of the supplemental information into the image information based on the embedment mode information acquired by the embedment mode information acquisition device, the embedment information acquired by the embedment information acquisition device, and the image information acquired by the image information acquisition device, and
an embeddability information output device to output determination-result information on a determination result of the embeddability determined by the embeddability determination device,
wherein the embeddability determination device includes an embeddable volume acquisition unit to acquire embeddable volume information on a volume of information embeddable in the image information based on at least a total information volume designated as embedment information, and
wherein, based on a comparison between a volume of the supplemental information to be embedded in the image information and the volume of information embeddable in the image information, the embeddability determination device determines the embeddability of the supplemental information into the image information.
2. The image processing apparatus according to
wherein the embeddability determination device determines the embeddability of the supplemental information into the image information based on the extraction of the embedment information.
3. The image processing apparatus according to
4. The image processing apparatus according to
5. The image processing apparatus according to
wherein, when the supplemental information is not embeddable in the image information, the information embedment device changes the embedment mode to embed the supplemental information into the image information.
6. The image processing apparatus according to
an area designation information acquisition device to acquire area designation information on designation of an area in which the supplemental information is to be embedded of the image information on the target image; and
an information embedment device to embed the supplemental information into the image information, wherein the information embedment device applies different embedment modes to the designated area and to one or more areas other than the designated area of the image information.
7. The image processing apparatus according to
an area designation information acquisition device to acquire area designation information on designation of an area in which the supplemental information is to be embedded of the image information on the target image; and
an information embedment device to embed the supplemental information into the image information,
wherein the embedment mode information includes parameter information on parameters referenced when the supplemental information is embedded in the image information, and wherein the information embedment device refers to different parameters for the designated area and for one or more areas other than the designated area of the image information.
9. The image processing system according to
wherein the embeddability determination device determines the embeddability of the supplemental information into the image information based on the extraction of the embedment information.
10. The image processing system according to
wherein, when the supplemental information is not embeddable in the image information, the embeddability report device prompts a user to change the embedment mode.
11. The image processing system according to
wherein, when the supplemental information is not embeddable in the image information, the information embedment device changes the embedment mode to embed the supplemental information into the image information.
12. The image processing system according to
wherein the information embedment device applies different embedment modes to the designated area and to one or more areas other than the designated area of the image information.
14. The method according to
embedding the embedment information on the supplemental information into the image information; and
extracting the embedment information from the image information,
wherein the determining includes determining the embeddability of the supplemental information into the image information based on the information extraction.
15. The method according to
wherein a user is prompted to change the embedment mode when the supplemental information is not embeddable in the image information.
16. The method according to
wherein the embedment mode is changed to embed the supplemental information into the image information when the supplemental information is not embeddable in the image information.
17. The method according to
acquiring area designation information on designation of an area in which the supplemental information is to be embedded of the image information on the target image; and
embedding the supplemental information into the image information,
wherein, in the embedding, different embedment modes are applied to the designated area and to one or more areas other than the designated area of the image information.
|
The present patent application claims priority pursuant to 35 U.S.C. §119 from Japanese Patent Application No. 2008-018295, filed on Jan. 29, 2008 in the Japan Patent Office, the entire contents of which are hereby incorporated herein by reference.
1. Field of the Invention
The present invention relates to an apparatus, system, and method of image processing, and more specifically, to an apparatus, system, and method of embedding information into an image.
2. Description of the Background
As electronic information has become widespread, image processing apparatuses, such as printers and facsimile machines, have become indispensable to output such electronic information. Such an image processing apparatus may be configured as a multi-functional device having the several capabilities of a printer, facsimile machine, scanner, and copier. Such an image processing apparatus may also be capable of embedding supplemental information into a target image for various purposes, such as copy protection of digital content or integration of digital data and printed data.
In conventional information embedment technology, the volume of embeddable information varies depending on the information embedment mode or the target image in which information is to be embedded. In fact, the information embedment mode is preferably determined based on the target image in which information is to be embedded.
At the same time, information embedment is subject to several requirements. For example, when a relatively large volume of information is embedded in an original image, a user may request that the embedded information affect the original image as little as possible. Further, a user may request to enhance the intensity of information embedment or the extractability of embedded information. Hence, for such information embedment into an image, a conventional image processing apparatus may prompt a user to designate a desired embedment mode or setting.
The present disclosure provides an image processing apparatus capable of easily determining whether or not information is embeddable in an image and selecting an appropriate information embedment mode or setting.
In one illustrative embodiment, an image processing apparatus includes an image information acquisition device, an embedment information acquisition device, an embedment mode information acquisition device, an embeddability determination device, and an embeddability information output device. The image information acquisition device acquires image information on a target image into which supplemental information is to be embedded. The embedment information acquisition device acquires embedment information on the supplemental information. The embedment mode information acquisition device acquires embedment-mode information on an embedment mode in which the supplemental information is embedded in the image information. The embeddability determination device determines embeddability of the supplemental information into the image information based on the embedment mode information acquired by the embedment mode information acquisition device, the embedment information acquired by the embedment information acquisition device, and the image information acquired by the image information acquisition device. The embeddability information output device outputs determination-result information on a determination result of the embeddability determined by the embeddability determination device.
In another illustrative embodiment, an image processing system includes an image processing apparatus. The image processing apparatus further includes an image information acquisition device, an embedment information acquisition device, an embedment mode information acquisition device, an embeddability determination device, and an embeddability information output device. The image information acquisition device acquires image information on a target image into which supplemental information is to be embedded. The embedment information acquisition device acquires embedment information on the supplemental information. The embedment mode information acquisition device acquires embedment mode information on an embedment mode in which the supplemental information is embedded in the image information. The embeddability determination device determines embeddability of the supplemental information into the image information based on the embedment mode information acquired by the embedment mode information acquisition device, the embedment information acquired by the embedment information acquisition device, and the image information acquired by the image information acquisition device. The embeddability information output device outputs determination-result information on a determination result of the embeddability determined by the embeddability determination device.
In still another illustrative embodiment, a method of image processing includes acquiring image information on a target image into which supplemental information is to be embedded, acquiring mode information on an embedment mode in which the supplemental information is embedded in the image information, determining embeddability of the supplemental information into the image information based on the acquired embedment mode information, the acquired embedment information, and the acquired image information, and outputting determination-result information on a determination result of the embeddability.
In still another illustrative embodiment, an information-processing-apparatus-readable storage medium storing program codes to cause an information processing apparatus to execute a method including acquiring image information on a target image into which supplemental information is to be embedded, acquiring mode information on an embedment mode in which the supplemental information is embedded in the image information, determining embeddability of the supplemental information into the image information based on the acquired embedment mode information, the acquired embedment information, and the acquired image information, and outputting determination-result information on a determination result of the embeddability.
A more complete appreciation of the disclosure and many of the attendant advantages thereof will be readily acquired as the same becomes better understood by reference to the following detailed description when considered in connection with the accompanying drawings, wherein:
The accompanying drawings are intended to depict illustrative embodiments of the present disclosure and should not be interpreted to limit the scope thereof. The accompanying drawings are not to be considered as drawn to scale unless explicitly noted.
In describing embodiments illustrated in the drawings, specific terminology is employed for the sake of clarity. However, the disclosure of this patent specification is not intended to be limited to the specific terminology so selected and it is to be understood that each specific element includes all technical equivalents that operate in a similar manner and achieve similar results.
Below, illustrative embodiments according to the present disclosure are described with reference to the attached drawings.
As one illustrative embodiment, a description is given below of an image processing apparatus 1 capable of receiving a print job from a client machine via a network and embedding information into an image in executing image formation.
Next, the image processing apparatus 1 according to the present illustrative embodiment is described.
The network I/F 140 is an interface through which the image processing apparatus 1 communicates other devices, such as the client machines 2 to 4, via the network 5. Print jobs sent from the client machines 2 to 4 are input to the control unit 100 via the network I/F 104, and image formation based on the print jobs is executed in accordance with control of the control unit 100. The display panel 105 is an output interface visually displaying a status of the image processing apparatus 1 and an input interface serving as, for example, a touch panel with which a user operates the image processing apparatus 1.
The control unit 100 is configured as a combination of software and hardware to control the image processing apparatus 1. Specifically, control programs, such as firmware, stored in a non-volatile storage medium, such as a ROM (read only memory), an EEPROM (electronically erasable and programmable ROM), the HDD 110, or an optical disk are loaded on a volatile memory (hereinafter “memory”), such as a DRAM (dynamic random access memory) and constructed as a software controller in accordance with control of a CPU (central processing unit). The control unit 100 is constructed from the software controller and hardware components, such as integrated circuits.
The main controller 111 instructs to control components of the control unit 100. The engine controller 112 controls or drives, for example, the print engine 107 and the scanner unit 102. The input-and-output controller 113 inputs, to the main controller 111, a print job input via the network I/F 104 or operation information input from a display panel 105 by a user. The input-and-output controller 113 also controls the display panel 105 to display information based on the instruction of the main controller 111 or transmits information to a host machine via the network I/F 104.
The image processing device 114 generates drawing information based on information contained in a print job, a document stored in the HDD 110, or any other suitable source. The drawing information is used for the print engine 107 to draw an image in image formation. The image processing device 114 processes imaging data input from the scanner unit 102 to generate image data. The image data is stored as a product (scanned information) of the scanning operation onto the HDD 110 via the HDD controller 115 or transmitted to a host machine via the network I/F 104.
The HDD controller 115 controls storing of information onto the HDD 110 and readout of information from the HDD 110 in accordance with the control of the main controller 111. The HDD 110 stores, for example, image information input as a print job, image information read from the scanner unit 102, network address information of client machines 2 to 4 connected via the network I/F 104, an address book including destination information used when the image processing apparatus 1 operates as a facsimile machine, and setting values referred in operations of the image processing apparatus 1.
The information embedment device 116 embeds information into an image that is input as a print job and output to a paper sheet using the print engine 107 or an image read using the scanner engine 102. The information embedment device 116 is capable of embedding (adding) information in accordance with a plurality of information embedment modes (information adding modes). When a user instructs to embed information into a target image, the embeddability determination device 117 determines whether or not the information is embeddable in the target image. A further description of the embeddability determination device 117 is given in detail later.
When the image processing apparatus 1 operates as a printer, the input-and-output controller 113 receives a print job via a USB (universal serial bus) or LAN (local area network) connected to the network I/F 104. Receiving the print job from the input-and-output controller 113, the main controller 111 controls the image processing device 114 to execute image processing. In accordance with an instruction from the main controller 111, the image processing device 114 generates drawing information based on image information contained in the print job or an accumulated document stored in the HDD 110. The drawing information generated with the image processing device 114 is sequentially spooled in the HDD 110.
Controlled by the main controller 111, the engine controller 112 drives the sheet feed table 106 to transport a print sheet to the print engine 107. The engine controller 112 also reads drawing information spooled in the HDD 110 to input the drawing information to the print engine 107. Based on the drawing information received from the engine controller 112, the print engine 107 forms an image on the sheet transported from the sheet feed table 106. Thus, the print engine 107 operates as an image forming device, and is applicable to an image forming mechanism of the inkjet method, the electrophotographic method, or any other suitable method. After the image formation, the sheet having the image is discharged to the discharge tray 108.
When the image processing apparatus 1 operates as a scanner, the input-and-output controller 113 sends a scan execution signal to the main controller 111 in response to a scan execution instruction input from the display panel 105 by a user or from one of the clients 2 to 4 via the network I/F 104. The main controller 111 controls the engine controller 112 based on the scan execution signal received from the input-and-output signal 113. The engine controller 112 drives the ADF 101 to transport a target document set on the ADF 101 to the scanner unit 102. The engine controller 112 also drives the scanner unit 102 to pick up an image of the document transported from the ADF 101. Alternatively, if a target document is set not on the ADF 101 but directly on the scanner unit 102, the scanner unit 102 picks up an image in the document in accordance with the control of the engine controller 112. Thus, the scanner unit 102 operates as an image pickup device.
In the image pickup operation, an image pickup element, such as a CCD (charge-coupled device), in the scanner unit 102 optically scans a target document to generate image pickup information based on optical information. The engine controller 112 transmits the image pickup information generated by the scanner unit 102 to the image processing device 114. Controlled by the main controller 111, the image processing device 114 generates image information based on the image pickup information received from the engine controller 112. The image information generated by the image processing device 114 is sequentially spooled in the HDD 110. In response to an instruction from a user, the image information spooled in the HDD 110 is stored in the HDD 110 or transmitted to an external host machine via the input-and-output controller 113 and the host I/F 104.
When the image processing apparatus 1 operates as a copier, the image processing device 114 generates drawing information based on the image pickup information received by the engine controller 112 from the scanner unit 102, or the image information generated by the image processing device 114 and stored in the HDD 110. As in the case with the printing operation, the engine controller 112 drives the print engine 107 based on the drawing information.
According to the present illustrative embodiment, in a print job accompanying information embedment, the image processing apparatus 1 determines whether or not information subjected to an embedment instruction is embeddable in a target image, and reports a determination result to a user.
Below, operation of the image processing system 1000 according to the present illustrative embodiment is described with reference to
When transmitting a print job, a user operates the client machine 2 to display a GUI (graphical user interface) of a setting screen for setting an information embedment mode on a display device of the client machine 2 (at S301 of
As illustrated in
In the embedment mode designation section 202 are displayed a mode-selection portion 202a for selecting an information embedment mode available in the image processing apparatus 1, a setting input portion 202b for inputting settings associated with the mode selected in the mode selection portion 202a, and an embedment intensity selection portion 202c for selecting an intensity of information embedment. The user selects a desired embedment mode in the mode selection portion 202a, inputs setting information associated with the selected mode in the setting input portion 202b, and selects an intensity of information embedment in the embedment intensity selection portion 202c. Here, with reference to
The embedment intensity selection portion 202c sets a post-embedment extractability of information to be embedded in an image. Setting a higher embedment intensity increases a possibility with which embedded information may be extracted from the image. The intensity of information embedment may be adjusted by inputting identical information in a multiplexing manner. In such a case, setting a relatively higher embedment intensity may reduce the volume of information embeddable in the image.
When a user inputs information in the setting screen 200, the information is transmitted along with a print job to the image processing apparatus 1. For example, such input information may be stored in a header portion of a target image, and transmitted at S302 along with the print job. Receiving the information input to the image processing apparatus 1 via the network I/F 104, at S303 the input-and-output controller 113 transmits the information to the main controller 111. Receiving the information from input-and-output controller 113, at S304 the main controller 111 acquires image information on a target image in which information is to be embedded, the information to be embedded in the target image, information on an embedment mode, settings associated with the embedment mode, and embedment intensity. Thus, according to the present illustrative embodiment, the main controller 111 operates as an image information acquisition device, an embedment information acquisition device; and an embedment mode information acquisition device. Based on the plurality of pieces of information acquired at S304, at S305 the main controller 111 transmits, to the embeddability determination device 117, an instruction for determining whether or not the information to be embedded is embeddable in the target image.
Receiving the instruction, at S306 the embeddability determination device 117 determines whether or not the information to be embedded is embeddable in the target image based on the information acquired from the main controller 111. At S306 the embeddability determination device 117 executes different types of processing depending on designated embedment modes. Below, the embeddability determination processing at S306 is described for the each of the embedment modes described with reference to
Finishing the extraction of frequency components above the threshold frequency, at S603 the embeddability determination device 117 calculates the volume of information embeddable in the target image based on the extracted frequency components. At S603 the information on embedment intensity designated at the intensity selection portion 202c illustrated in
To verify the embeddability of information into a target image, the embeddability determination device 117 may include an embedment-information embedding unit to embed embedment information (verification information) on the information to be embedded in image information on the target image and an embedment-information extracting unit to extract the embedment information from the image information.
For example, as illustrated in
Through the above-described processing, at S307 of
Next, with reference to drawings, a description is given of such a determination result displayed in the display device of the client machine 2.
As illustrated in
Displaying the GUI as illustrated in
In the above-described example illustrated in
In the above-described example, even when the embeddability determination device 117 determines that information is embeddable, the image processing apparatus 1 reports a determination result to the client machine 2 and executes print output in response to the user's operation of the “execution” button. Such a configuration allows a user to, for example, add embedment information, increase embedment intensity, or change the information embedment mode between a perceptible embedment mode and an non-perceptible embedment mode based on the determination result. Meanwhile, it is to be noted that, when the embeddability determination device 117 determines that information is embeddable, the image processing apparatus 1 may continue print output without reporting the determination result to a user, thereby reducing user's efforts.
In the above-described example, when the embeddability determination device 117 determines that information is not embeddable, the GUIs illustrated in
In still another example, when it is determined that information is not embeddable, the embeddability determination device 117 determines the embeddability in other information embedment modes. As a result, an information embedment mode determined as an embeddable mode may be displayed on the GUI illustrated in
In
In the above-described example, a print job is input from the client machine 2 to the image processing apparatus 1 is described as one example. Meanwhile, the present illustrative embodiment is also applicable to an operation of copying a document scanned with the scanner unit 102. For example, a user may input information from the setting screen 200 illustrated in
In the above-described examples, the operations assuming image formation based on a print job or copying instruction are described. Meanwhile, the present illustrative embodiment is applicable to when image information including embedded information is also stored into the HDD 110 of the image processing apparatus 1 or the storage devices of the client machines 2 to 4. Such storage operation is performed, for example, during scan operation using the image processing apparatus 1 or when information embedded in existing image information or user-generated image information is stored. Below, the case in which information is embedded in existing image information is described with reference to
In the above-described example, the client machine 2 transmits information to be embedded to the image processing apparatus 1 at S302 in
In the above-described example, the GUIs illustrated in
In the above-described example of
For such a reduction in image resolution, the resolutions of both the dot pattern and the target image may be reduced before a dot pattern may be superimposed on a target image, or the resolution of a target image in which a dot pattern has been embedded may be reduced. For example, when image resolution is reduced to one half, a dot of 2 by 2 pixels is reduced to a dot of 1 by 1 pixel. Such a reduction in image resolution might affect the detection accuracy of a dot pattern. However, for example, such a one-half reduction of image reduction does not affect conditions of white pixels around a dot, thus not causing a significant reduction in detection accuracy.
Further, in the above-described example, the frequency transformation mode, the dot pattern mode, and the outline modification mode are employed as the information embedment mode. The present illustrative embodiment is applicable not only to the above-described information embedment modes but also to any other suitable mode. For example, a “character spacing mode” in which information to be embedded in a target image is encoded by changing character spacing based on a certain rule, a mode in which code information, such as a bar code, is embedded in a target image, or any other suitable information embedment mode may be employed.
Next, a description is given of another illustrative embodiment according to the present disclosure.
In the above-described illustrative embodiment, an image of a print job, a scanned image, or an existing image is designated as a target for information embedment. In this illustrative embodiment, a portion of an image is designated as a target for information embedment. It is to be noted that, in this illustrative embodiment, the same reference numerals as those of the above-described illustrative embodiment indicate the same or corresponding constituent elements, and redundant descriptions thereof are omitted below.
When the frequency transformation mode described in the above-described illustrative embodiment is used, as described above, information is embedded in an area of frequency components above a certain frequency set as a threshold, thereby allowing the information to be embedded in a target image without reducing the visibility of the target image. However, for example, when encoded graphic information, such as a bar code, is in an area in which information is embedded, embedding the information may affect the reading (decoding) of encoded information using a bar-code reader even without causing any significant change in the visibility of the image by the user. Accordingly, the frequency transformation mode might be unsuitable for embedding information in such an area including encoded graphical information such as a bar code.
Further, when the dot pattern mode described in the above-described illustrative embodiment is used, the encoded graphical information might overlay the dot pattern, preventing the encoded information from being properly read (or decoded). Alternatively, superimposing a dot pattern on an area including a tiny or patchy character might prevent such a character from being properly read.
Further, when the outline modification mode is used, it might be difficult to properly embed information into or read embedded information from the outline of a tiny or patchy character. Hence, according to the present illustrative embodiment, whether or not information embedment is executed and various information embedment settings can be varied for different areas of a target image.
The configurations of an image processing system 1000, an image processing apparatus 1, and client machines 2 to 4 according to the present illustrative embodiment are substantially the same as the configurations according to the above-described illustrative embodiment. Further, the operation of determining information embeddability according to the present illustrative embodiment is substantially the same as the above-described determination operation according to the first illustrative embodiment. However, it is to be noted that, in the present illustrative embodiment, embedment-area designation processing for designating an area of a target image in which information is to be embedded is executed in addition to the processing described above with reference to
A detailed description is given below of the embedment-area designation processing.
Information on the embedment area designated in the preview section 310 is generated as coordinate information. Information on the processing designated in the processing designation section 320 is generated as area designation information, and reported along with the coordinate information and other information to the image processing apparatus 1 at S302 of
According to the present illustrative embodiment, as illustrated in
Below, settings for information embedment capable of preventing occurrences of such difficulties are described with reference to drawings. Incidentally, the following description is made based on a case in which the dot pattern mode is used.
In the examples illustrated in
Specifically, for the dummy dot pattern superimposed on the photographic area A, the small-character area B, and the hatching area C, the size of a dot pattern is set smaller than the dot pattern superimposed on other areas. Such a configuration can prevent a reduction in the visibility of a photographic image drawn in the photographic area A or a character drawn in the small-character area B. Meanwhile, in the bar-code area D, the dummy dot pattern may not be superimposed or the dot size may be set to zero. As described above, the dummy dot pattern superimposed on the photographic area A, the small-character area B, and the hatching area C may be different in dot size from the dot pattern for information embedment superimposed on the other areas. Alternatively, the dot density, e.g., the density of dots against the background, may be set close between the dummy dot pattern and the dot pattern for information embedment. Such a configuration can provide a substantially uniform dot density in the photographic area A, the small-character area B, the hatching area C, and the other area (except the bar-code area D) when a user see a resultant image in which information has been embedded, thereby preventing a reduction in the visibility or degradation in the general appearance of the resultant image.
In such a case, the difference in dot density between the dot pattern superimposed on the photographic area A, the small-character area B, and the hatching area C and the dot pattern superimposed on the other area may be set, for example, within 5% or to zero.
As described above, in the image processing system according to the present illustrative embodiment, designating an area in which information is to be embedded can prevent a reduction in visibility or degradation in appearance of a target image. Further, embedding information into respective areas of the target image using different settings (in the above-described example, different dot sizes) can prevent degradation in general appearance of the image.
In the above-described example of
Thus, the dot size of the dot pattern superimposed on the photographic area A, the small-character area B, and the hatching area C may be set sufficiently large that the dots are not recognized as noise by an image pickup device, such as the scanner unit 102, yet sufficiently small that the visibility of characters in the small-character area B is not reduced. Such a small size that the visibility of characters is not reduced is, for example, a size smaller than the width of a line constituting a character as illustrated in
As described above, in the embeddability determination of a designated area, it may be determined that information is not embeddable in the designated area due to a lack of information embeddable area. In such a case, the GUI illustrated in
In the above-described example, a user manually designates an area in the embedment-area designation screen 300 illustrated in
For the detection of the small-character area B, character information in a target image may be extracted using a character recognition technology, such as OCR (optical character recognition). The image processing apparatus 1 compares the size of an extracted character with the size of a predetermined reference character to determine the embeddability of information in the small-character area B. Such processing allows the small-character area B to be detected. Extracting an area with such a small-size character that information embedment might affect the visibility of the target image, the image processing apparatus 1 embeds the information into an area other than the small-character area or into the small-character area using a dot pattern consisting of relatively small dots.
Alternatively, extracting an area including a patchy or improperly-drawn character using a pattern matching technology, the image processing apparatus 1 may embed information into an area other than the patchy or improperly-drawn area. Alternatively, the image processing apparatus 1 may embed information into the patchy or improperly-drawn area using a dot pattern consisting of relatively small dots. Further, the image processing apparatus 1 may include an extraction device to extract a designated key word from such character information recognized using an OCR unit, and embed information into an area other than the area in which the key word has been extracted. Such processing can secure the visibility of, for example, an important key word included in the target image.
For the detection of the photographic area A or the hatching area C, when an area including a photographic image or mesh image is extracted from the target image for information embedment, the image processing apparatus 1 embeds information into an area other than the extracted area or into the extracted area using a dot pattern consisting of relatively small dots.
The image processing apparatus 1 may also detect a predetermined dot pattern or encoded information as information embedded in a target image to perform a control operation, such as copy prohibition, on the target image. In such a case, unless the detection volume of the dot pattern or encoded information is beyond a certain threshold, such a control operation might not be performed. Hence, when the size of the detected or designated area in which information is not to be embedded is not above a threshold area size, and the area in which the information is to be embedded is sufficiently secured in another area, the image processing apparatus 1 embeds the information into the other area. By contrast, when the size of the area in which the information is not to be embedded is above the threshold area size, and the area in which information is to be embedded is not sufficiently secured in the other area, the image processing apparatus 1 embeds the information over the entire target image. Such processing is described below with reference to
As illustrated in S305 of
At S1802, the embeddability determination device 117 compares a predetermined threshold size with the size of each area extracted at S1801. If the embedment restriction area is not above the threshold (“YES” at S1802), at S1803 the embeddability determination device 117 determines the embeddability of information into an area other than the embedment restriction area. By contrast, if the embedment restriction area is above the threshold size (“NO” at S1802), at S1804 the embeddability determination device 117 executes embeddability determination regardless of the embedment restriction area.
Such processing can prevent the above-described possible failure from occurring in a control operation, such as copy prohibition, due to a lack of volume of the extracted dot pattern or encoded information.
As described in the above-described illustrative embodiments, in one aspect, the invention resides in an image processing apparatus including an image information acquisition device to acquire image information on a target image into which supplemental information is to be embedded, an embedment information acquisition device to acquire embedment information on the supplemental information, an embedment mode information acquisition device to acquire embedment mode information on an embedment mode in which the supplemental information is embedded in the image information, an embeddability determination device to determine embeddability of the supplemental information into the image information based on the embedment mode information acquired by the embedment mode information acquisition device, the embedment information acquired by the embedment information acquisition device, and the image information acquired by the image information acquisition device, and an embeddability information output device to output determination-result information on a determination result of the embeddability determined by the embeddability determination device.
The image processing apparatus may further include an embeddability report device to report, to a user, the determination-result in formation on the determination result output from the embeddability information output device.
The embeddability report device may further indicate another embedment mode different from the embedment mode of the embedment mode information acquired by the embedment mode information acquisition device when the supplemental information is not embeddable in the image information.
The embedment mode information may further include parameter information on parameters referred when the supplemental information is embedded in the image information, and, when the supplemental information is not embeddable in the image information, the embeddability report device may prompt a user to change the parameters.
The image processing apparatus may further include an information embedment device to embed the supplemental information into the image information, wherein, when the supplemental information is not embeddable in the image information, the information embedment device embeds the supplemental information in the image information using another embedment mode different from the embedment mode of the embedment mode information acquired by the embedment mode information acquisition device.
The information embedment device may change the parameters to embed the supplemental information into the image information when the supplemental information is not embeddable in the image information.
The image processing apparatus may further include an area designation information acquisition device to acquire area designation information on designation of an area in which the supplemental information is to be embedded of the image information on the target image, wherein the embeddability determination device determines the embeddability of the supplemental information into the designated area of the image information.
The information embedment device may embed the supplemental information into the designated area of the image information or one or more areas other than the designated area of the image information.
The information embedment device may refer to different parameters for the designated area and for the one or more areas of the image information.
In the embedment mode, when a dot pattern generated from the supplemental information is superimposed on an image drawn based on the image information, a first character drawn in the designated area may be smaller than a second character drawn in the one or more areas of the image information, and the information embedment device may superimpose, on the designated area, a dot pattern formed of dots smaller than dots forming another dot pattern superimposed on the one or more areas.
As described above, such configurations facilitate determining whether information is embeddable in an image and allows a user to easily select an appropriate information embedment mode.
Numerous additional modifications and variations are possible in light of the above teachings. It is therefore to be understood that within the scope of the appended claims, the disclosure of the present invention may be practiced otherwise than as specifically described herein.
With some embodiments of the present invention having thus been described, it will be obvious that the same may be varied in many ways. Such variations are not to be regarded as a departure from the spirit and scope of the present invention, and all such modifications are intended to be included within the scope of the present invention.
For example, elements and/or features of different illustrative embodiments may be combined with each other and/or substituted for each other within the scope of this disclosure and appended claims.
Further, as described above, any one of the above-described and other methods of the present invention may be embodied in the form of a computer program stored in any kind of storage medium. Examples of storage mediums include, but are not limited to, flexible disk, hard disk, optical discs, magneto-optical discs, magnetic tapes, nonvolatile memory cards, ROM (read-only-memory), etc.
Alternatively, any one of the above-described and other methods of the present invention may be implemented by ASIC, prepared by interconnecting an appropriate network of conventional component circuits or by a combination thereof with one or more conventional general purpose microprocessors and/or signal processors programmed accordingly.
Yoshida, Hiroyuki, Shimura, Hiroshi, Saitoh, Takashi, Ishikawa, Masaaki, Guan, Haike, Ishizu, Taeko
Patent | Priority | Assignee | Title |
8472085, | Aug 19 2008 | Ricoh Company, LTD | Image processing device, additional information providing method, and computer-readable storage medium |
9369594, | Jun 06 2013 | Fujitsu Limited | Transmission device, relay device, recording medium and control method of transmission system |
9830114, | Mar 26 2009 | FUJIFILM Business Innovation Corp | Image processing apparatus and computer readable medium and method thereof |
9986123, | May 14 2014 | Ricoh Company, LTD | Information processing apparatus, system, and method for achieving embedding process capable of accepting various condition settings |
Patent | Priority | Assignee | Title |
7031491, | Apr 09 1999 | Canon Kabushiki Kaisha | Method for determining a partition in order to insert a watermark, and associated insertion and decoding methods |
7263203, | May 08 1995 | DIGIMARC CORPORATION AN OREGON CORPORATION | Digital watermarks |
20020066019, | |||
20040120544, | |||
20040247155, | |||
20050041263, | |||
20070064973, | |||
20070074029, | |||
20070127771, | |||
20070199992, | |||
20080166013, | |||
20080232639, | |||
20080232640, | |||
EP1220152, | |||
EP1703460, | |||
JP2000216985, | |||
JP2001197297, | |||
JP2001346032, | |||
JP2004274092, | |||
JP2005235181, | |||
JP3335265, | |||
JP7036317, | |||
JP9164739, | |||
WO2007003008, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Jan 27 2009 | ISHIKAWA, MASAAKI | Ricoh Company, LTD | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 022238 | /0420 | |
Jan 27 2009 | SHIMURA, HIROSHI | Ricoh Company, LTD | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 022238 | /0420 | |
Jan 27 2009 | GUAN, HAIKE | Ricoh Company, LTD | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 022238 | /0420 | |
Jan 27 2009 | ISHIZU, TAEKO | Ricoh Company, LTD | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 022238 | /0420 | |
Jan 27 2009 | YOSHIDA, HIROYUKI | Ricoh Company, LTD | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 022238 | /0420 | |
Jan 27 2009 | SAITOH, TAKASHI | Ricoh Company, LTD | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 022238 | /0420 | |
Jan 28 2009 | Ricoh Company, Ltd. | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Jan 09 2013 | ASPN: Payor Number Assigned. |
Dec 16 2015 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Feb 17 2020 | REM: Maintenance Fee Reminder Mailed. |
Aug 03 2020 | EXP: Patent Expired for Failure to Pay Maintenance Fees. |
Date | Maintenance Schedule |
Jun 26 2015 | 4 years fee payment window open |
Dec 26 2015 | 6 months grace period start (w surcharge) |
Jun 26 2016 | patent expiry (for year 4) |
Jun 26 2018 | 2 years to revive unintentionally abandoned end. (for year 4) |
Jun 26 2019 | 8 years fee payment window open |
Dec 26 2019 | 6 months grace period start (w surcharge) |
Jun 26 2020 | patent expiry (for year 8) |
Jun 26 2022 | 2 years to revive unintentionally abandoned end. (for year 8) |
Jun 26 2023 | 12 years fee payment window open |
Dec 26 2023 | 6 months grace period start (w surcharge) |
Jun 26 2024 | patent expiry (for year 12) |
Jun 26 2026 | 2 years to revive unintentionally abandoned end. (for year 12) |