There is provided an image encode encoding apparatus which comprises a generation means for generating a prediction error from an encode encoding target pixel value and a prediction value of the encode encoding target pixel value, a judgment means for generating a generation frequency distribution of the prediction error to judge whether or not the generation frequency distribution is discrete and an entropy encode encoding means for changing encode encoding data corresponding to the prediction error and performing entropy encoding on the obtained encode encoding data, in accordance with a judged result by the judgment means. Therefore, the encoding can be effectively performed even on such the image data having the discrete pixel value as in a CG image, a limited-color image or the like.
|
12. An image encoding method comprising the steps of:
generating a prediction value of an encoding target pixel;
correcting the prediction value;
judging whether or not each of pixels of a reference area is represented by a part of one of a plurality of total available values, wherein the encoding target pixel is not involved in the judging operation, and each pixel of the reference area is encoded before the encoding target pixel is encoded; and
controlling the correcting operation in said correcting step in accordance with a result in said judging step.
13. A computer readable storage medium that stores program codes for executing an image encoding method, said method comprising the steps of:
generating a prediction value of an encoding target pixel;
correcting the prediction value;
judging whether or not each of pixels of a reference area is represented by part of one of a plurality of total available values, wherein the encoding target pixel is not involved in the judging operation, and each pixel of the reference area is encoded before the encoding target pixel is encoded; and
controlling the correcting operation in said correcting step in accordance with a result in said judging step.
9. An image encoding apparatus comprising:
generating means for generating a prediction value of an encoding target pixel;
prediction value correcting means for correcting the prediction value;
judging means for judging whether or not each of pixels of a reference area is represented by part of one of a plurality of total available values, wherein the encoding target pixel is not involved in the judging operation, and each pixel of the reference area is encoded before the encoding target pixel is encoded; and
control means for controlling the correcting operation by said prediction value correcting means in accordance with a result by said judging means.
7. An image encoding method, comprising the steps of:
generating a first prediction error difference on the basis of a first value of at least one reference pixel and a first prediction value of the reference pixel, and generating a second prediction error difference on the basis of a second value of an encoding target pixel and a second prediction value of the encoding target pixel, the reference pixel being encoded before the encoding target pixel is encoded and the first prediction error difference being generated before the second prediction error difference is generated;
judging an appearing prediction error difference and an unappearing prediction error difference on the basis of the first prediction error difference, and for encoding the second prediction error difference on the basis of the judged appearing and unappearing prediction error differences, wherein the second prediction error difference is not used in the judging operation;
changing a first relationship between prediction error difference and encoding data to a second corresponding relationship between prediction error difference and encoding data according to a result obtained in said judging step; and
encoding the second prediction error difference on the basis of a selected the appearing and unappearing prediction error differences judged in said judging step and one of the first and second corresponding relationships to obtain corresponding encoding data.
1. An image encoding apparatus, comprising:
generating means for generating a first prediction error difference on the basis of a first value of at least one reference pixel and a first prediction value of the reference pixel, and generating a second prediction error difference on the basis of a second value of an encoding target pixel and a second prediction value of the encoding target pixel, the reference pixel being encoded before the encoding target pixel is encoded and the first prediction error difference being generated before the second prediction error difference is generated;
judging means for judging an appearing prediction error difference and an unappearing prediction error difference on the basis of the first prediction error difference, and for encoding the second prediction error difference on the basis of the judged appearing and unappearing prediction error differences, wherein the second prediction error difference is not used in the judging operation;
changing means for changing a first corresponding relationship between prediction error difference and encoding data to a second corresponding relationship between prediction error difference and encoding data according to a result obtained by said judging means; and
encoding means for encoding the second prediction error difference on the basis of a selected the appearing and unappearing prediction error differences judged by said judging means and one of the first and second corresponding relationships to obtain corresponding encoding data.
8. A computer readable storage medium that stores program codes for executing an image encoding method, said method comprising the steps of:
generating a first prediction error difference on the basis of a first value of at least one reference pixel and a first prediction value of the reference pixel, and generating a second prediction error difference on the basis of a second value of an encoding target pixel and a second prediction value of the encoding target pixel, the reference pixel being encoded before the encoding target pixel is encoded and the first prediction error difference being generated before the second prediction error difference is generated;
judging an appearing prediction error difference and an unappearing prediction error difference on the basis of the first prediction error difference, and for encoding the second prediction error difference on the basis of the judged appearing and unappearing prediction error differences, wherein the second prediction error difference is not used in the judging operation;
changing a first relationship between prediction error difference and encoding data to a second corresponding relationship between prediction error difference and encoding data according to a result obtained in said judging step; and
encoding the second prediction error difference on the basis of a selected the appearing and unappearing prediction error differences judged in said judging step and one of the first and second corresponding relationships to obtain corresponding encoding data.
2. An apparatus according to
3. An apparatus according to
6. An apparatus according to
10. An apparatus according to
11. An apparatus according to
|
encode108 109 the Huffman table which was formed on the basis of characteristics of the prediction errors generated in case of prediction-encoding several sample images.
Further, as an initial stage, the conversion table 111 has been set such that the output value M(i) for the input value i satisfies M(i)=i. Also, in the counter 105, all the values (i.e., the number of appearance) F(e) corresponding to the prediction errors e (=−255 to 255) have been set as “0”.
Subsequently, an operation of each unit in the present embodiment will be explained sequentially.
Initially, an encoding target pixel x is inputted, in the order of raster scanning, into the buffer 102 and the subtracter 104 through the signal line 101.
The buffer 102 stores the signal which is inputted from the signal line 101 and corresponds to the two lines. Then, the predictor 103 reads peripheral pixels a, b and c of the encoding target pixel x from the buffer 102, and obtains a prediction value p for the encode encoding target pixel x by calculating p=a+b−c. In this case, positions of the peripheral pixels a, b and c for the encoding target pixel x are shown in FIG. 3.
The subtracter 104 generates the prediction error e by calculating e=x−p and also performs increment of the counter value F(e). For example, if the prediction error is “2”, the subtracter 104 performs the increment of the counter value F(2) corresponding to such the a prediction error “2”.
The converter 106 outputs the intermediate output value e′ corresponding to the inputted prediction error e, by referring to the conversion table 111.
Processing operations of the update judgment circuit 112, the conversion table update circuit 107 and the conversion table 111 will be explained later.
The Huffman encoder 108 outputs encoding data corresponding to the intermediate output value e′ to the signal line 110, by referring to the Huffman table stored in the Huffman table memory 109.
The above-described serial encoding process is repeatedly performed for the 16-line pixels of the inputted image data.
Subsequently, processes of the counter 105 and the update judgment circuit 112 will be explained in detail.
When the encoding process of 16 lines is terminated in the counter 105, data representing the number of appearance of each prediction error e is generated as shown in FIG. 5.
The update judgment circuit 112 checks the number (F(−Th) to F(Th)) of appearance of the prediction error (−Th to Th) (it is assumed that Th=32 in the present embodiment), in the counter values held in the counter 105.
If there is the value “0” as the number of appearances in the counter values, the update judgment circuit 112 judges that the objective image is a specific image (i.e., a CG image, a limited-color image or the like) in which the prediction error not generated exists. Thus, the circuit 112 outputs to the signal line 113 a control signal “1” for operating the conversion table update circuit 107. On the other hand, if there is no prediction error having the value “0” as the number of appearance in the 16 lines, the objective image is encoded as an ordinary image. Therefore, the update judgment circuit 112 outputs to the signal line 113 a signal “0” not for for not operating the conversion table update circuit 107. In this case, it is judged whether or not the objective image is the specific image, on the basis of the prediction error generated in the 16 lines of the image. However, the present invention is not restricted to such an operation. That is, the prediction error can be adaptively changed within a range beginning from the several pixels to the plurality of images, if in unit of image suitable for the judging.
In the example shown in
Further, the counter 105 and the conversion table 111 are reset to an initial state for every 16 lines, and the processes of the counter 105 and the update judgment circuit 112 are repeated.
Then, an operation of the conversion table update circuit 107 will be explained in detail, hereinafter.
The conversion table update circuit 107 operates in the case where the control signal from the signal line 113 is “1”, to perform a process for rewriting the conversion table 111. If the circuit does not perform the operation, the conversion table 111 is successively used.
The circuit 107 checks the number Np of the prediction error e of which the number of appearance is not “0”, by referring to the number of appearance within the range of the prediction error e from “0” to “255” in the counter 105.
Then, the circuit 107 allocates the prediction error e of which the number of appearance is not “0” to integer values “0” to “Np−1” of an intermediate output value M(e) held in the conversion table 111, in the order of “0”, “1”, “2”, “3” to “255” (i.e., “0”, “2”, “4”, “6”, . . . in FIG. 5).
Subsequently, the circuit 107 sequentially allocates the prediction error e of which the number of appearance is “0” (i.e., “1”, “3”, “5”, . . . in FIG. 5).
In the same manner as described above, the circuit 107 checks the number Nm of the prediction error e of which the number of appearance is not “0”, for the prediction error e from “−1” to “−255”.
Further, the circuit 107 allocates the prediction error e of which the number of appearance is not “0” to the integer values “−1” to “−Nm” of the intermediate output value M(e) held in the conversion table 111, in the order of “−1”, “−2”, “−3” to “−255” (i.e., “−2”, “−4”, “−6” . . . in FIG. 5).
Subsequently, the circuit 107 sequentially allocates the prediction error of which the number of appearance is “0” (i.e., “−1”, “−3”, “−5” . . . in FIG. 5), to the integer values “−Nm−1” to “−255” of the intermediate output value M(e).
By such operations, the conversion table 111 is rewritten such that the prediction errors of which frequency of generation is high concentrate on the value close to the intermediate output value “0” of which code length is short. Therefore, in the case where the objective image is judged as the specific image by the update judgment circuit 112, the value having large absolute value is allocated to the prediction error which is never generated, whereby the effective entropy encoding can be performed.
For an image having the generation frequency of the prediction error as shown in
As described above, the encoding process for the image data of the 16 lines and the updating of the conversion table 111 are repeatedly performed, the encoding is continuously performed for the final pixel inputted from the signal line 101, and then the encoding data is outputted to the signal line 110.
(Second Embodiment)
Subsequently, the second embodiment of the present invention will be explained with reference to the accompanying drawings.
Like the first embodiment, it will be explained as an example a case where an eight-bit (i.e., values “0” to “255”) monochrome image is encoded.
It is assumed that the memory 707 has stored and holds a value of the parameter k suitable for each context which value was obtained by previously performing an encoding test of several sample images.
The prediction value correction circuit 717 has therein a memory region for holding the number N(S) of generation of the prediction error e for each context S and cumulative value E(S) of the prediction error e for each context S. In an initial state, all the values have been set as “0”.
It is assumed that an initial state of the control signal outputted to the signal line 718 has been set as “0”. Further, as the initial state, all the flags f(e) held in the inspection circuit 708 have been set as “0”, and also the conversion table 711 has been set such that the output value M(i) for the input value i satisfies the following equation.
Hereinafter, an operation of each unit in the present embodiment will be sequentially explained.
Initially, image data or signal representing an encoding target pixel x is inputted, in the order of raster scanning, into the buffer 702 and the subtracters 705 and 716 through the signal line 701.
The buffer 702 stores the signal of two lines inputted from the signal line 701. The predictor 704 reads the image data of peripheral pixels a, b and c of the encoding target pixel x from the buffer 702, and obtains prediction value p for the encoding target pixel x by calculating p=a+b−c. Positions of the peripheral pixels a, b and c for the encoding target pixel x are shown in FIG. 3. The context generator 703 firstly reads the peripheral pixels a, b and c of the encoding target pixel x from the buffer 702, and secondly obtains values (a−c) and (b−c). Then, the context generator 703 obtains values q(a−c) and q(b−c) by quantizing each of the two values (a−c) and (b−c) into three levels in accordance with a table shown in FIG. 9. By using such results, the context generator 703 calculates the context S=3×q(a−c)+q(b−c) to generate and output the context S=“0” to “8”.
The parameter k selection circuit 706 reads the value of the parameter k corresponding to the context S from the memory 707 and then outputs the read value. The subtracter 716 outputs a prediction error e1=x−p which is the difference between the prediction value p generated by the predictor 704 and the encoding target pixel value x.
In a case where a control signal from the signal line 718 is “0” and a signal from a signal line 718′ is “0”, the prediction value correction circuit 717 outputs the prediction value p outputted from the predictor 704, as it is, as a corrected prediction value p′.
On the other hand, in a case where the control signal from the signal line 718 is “1” or the signal from the signal line 718′ is “1”, the circuit 717 reads the number N(S) of generation of the prediction error e for each context S and the cumulative value E(S) of the prediction error e from its internal memory region, on the basis of the context S generated by the context generation circuit 703. Subsequently, the circuit 717 obtains a mean value merr(S) of the prediction errors in the context S by calculating E(S)/N(S). By using the mean value merr(S) and the prediction value p outputted from the predictor 704, the circuit 717 outputs the corrected prediction value p′=p+merr(S), to the subtracter 705.
The subtracter 705 obtains the prediction error e=x−p′ from the corrected prediction value p′ outputted by the prediction value correction circuit 717 and the encoding target pixel x, and outputs the obtained prediction error e.
Further, every time the prediction error e is generated, in a case where the control signal from the signal line 718 is “1” or the control signal line 718′ is “1”, the prediction value correction circuit 717 newly adds the prediction error e outputted by the subtracter 705 to the cumulative value E(S), and performs increment of the number N(S) of generation of the prediction error e for each context S to store the obtained data into the internal memory region.
The converter 712 reads an intermediate output value e′ for the input value e from the conversion table 711, and then outputs the read value. The Golomb-Rice encoder 713 generates encoding data of the intermediate output value e′ on the basis of a code allocation table shown in
Subsequently, a detailed encoding procedure in the Golomb-Rice encoder 713 will be explained hereinafter.
Firstly, the intermediate output value e′ (i.e., “0” to “510” in this case) which is the target of encoding is expressed in binary number, and then divided into a lower k-bit portion and an upper remaining portion. It is assumed that the value obtained by expressing the upper remaining portion in decimal number is n. The encoding data is obtained by adding “0” to the lower k-bit portion by n, and finally adding “1”. For example,
When a prediction error el is outputted from the subtracter 716, the inspection circuit 708 changes the flag f(e1) corresponding to the prediction error el into “1”.
The inspection circuit 708 checks, in the held flags f(−Th) to f(Th), the values of the flags f(−Th) to f(Th) corresponding to the prediction errors “−Th” to “Th” (where “Th” is an arbitrary integer within the range “0” to “255”), for each pixel clock.
In a case where the flags f(−Th) to f(Th) are all “1”, the inspection circuit 708 outputs the control signal “1” to the signal line 718′, and outputs a control signal 719 for resetting or returning the conversion table 711 to the initial state.
On the other hand, in a case where at least one of the flags f(−Th) to f(Th) is “0”, the inspection circuit 708 outputs the control signal “0” to the signal line 718′.
For example, when k=2, the intermediate value e′=5 (decimal) (binary “101”) is divided into a lower 2(=k) bit part “01” and an upper part “1”. It is determined that the lower bit part “01” is the upper 2(=k) bit of Golomb encoding data. That is, it is determined that the encoding data is “01* . . . ”. Furthermore, the code “0” of the number 1 (decimal) represented by the upper bit “1” is added on or after 3 (=k+1) bits of Golomb encoding data. That is, it is determined that the encoding data is “010* . . . ”. Sequentially, the code “1” indicating the end of Golomb encoding data is added. Accordingly, it is determined that the final encoding data is “0101”.
The above-described encoding process is repeatedly performed for the eight-line pixels of the input image data.
Subsequently, an operation of the update circuit will be explained in detail.
In the above-described processes, in the flags f(−Th) to f(Th) held in the inspection circuit 708, “1” has been allocated to the prediction error generated within the eight lines, and “0” has been allocated to the prediction error not generated. When the process of the eight lines terminates, the update judgment circuit 709 refers to the flag held in the inspection circuit 708. Then, in a case where “0” exists in the flags f(−Th) to f(Th) corresponding to the prediction errors “−Th” to “Th” (where “Th” is an arbitrary integer within the range “0” to “255”), during the process of the subsequent eight lines, the update judgment circuit 709 outputs the control signal “1” for operating the conversion table update circuit 710 to the signal line 714 and also outputs the control signal “0” for controlling the operation of the prediction value correction circuit 717 to the signal line 718.
On the contrary, in a case where the flags f(−Th) to f(Th) are all “1”, during the process of the subsequent eight lines, the update judgment circuit 709 outputs the control signal “0” not for for not operating the conversion table update circuit 710 to the signal line 714 and also outputs the control signal “1” for controlling the operation of the prediction value correction circuit 717 to the signal line 718. In addition, the inspection circuit 708 and the conversion table 711 are reset or returned to the initial state.
The conversion table update circuit 710 operates in the case where the control signal from the signal line 714 is “1”, to rewrite the conversion table 711.
In this case, firstly, the circuit 710 refers to the value of the flag held in the inspection circuit 708, so as to check the number Np of the prediction error which has been generated even once in the previously-encoded eight lines, within the range of the prediction errors “0” to “255”. Secondly, the circuit 710 checks the value of the flag in the order of the prediction error “0”, “1”, “2”, “3” to “255”. In a case where the flag f(e) is “1” (i.e., the prediction error e is generated even once), the circuit 710 allocates even-number values of “0” to “2×Np−1” to the output value M(e) held in the conversion table 711. On the other hand, in a case where the flag f(e) is “0” (i.e., the prediction error e is not generated even once), the circuit 710 sequentially allocates the even-number values of “2×Np” to “510” to the output value M(e).
Subsequently, the circuit 710 checks the number Nm of the prediction error generated within the range of the prediction error “−1” to “−255”. Like the case where the prediction error is positive, the circuit 710 checks the flag f(e) in the order of the prediction error “−1”, “−2”, “−3” to “−255”. If the flag f(e) is “1”, the circuit 710 sequentially allocates odd-number values of “1” to “2×Nm−1” to the output value M (e) held in the conversion table 711. On the other hand, if the flag f(e) is “0”, the circuit 710 sequentially allocates the odd-number values of “2×Nm+1” to “509” to the output value M(e). By such the operations, the conversion table 711 is rewritten. Therefore, the prediction error which has not been generated even once in the previously-encoded eight lines is managed, irrespective of magnitude of its value, as the large-value prediction error from the time of encoding the subsequent eight lines. For this reason, in the prediction encoding in which the short encoding code is allocated to the small prediction error, the compression efficiency can be improved. After the rewriting of the conversion table 711 is terminated, the conversion table update circuit 707 initializes the inspection circuit 708.
As described above, the encoding process of the eight-line image data and the successive updating of the conversion table 711 are repeatedly performed, the encoding is continuously performed until the final encode encoding target pixel of the image inputted is inputted from the signal line 701, whereby the encoding data is outputted to the signal line 715.
The present invention is not restricted to the above-described embodiments. For example, as a prediction method of the encoding target pixel value, pre-predicting may be simply used. On the other hand, by providing several prediction methods, these methods may be appropriately switched or exchanged.
Further, in the above-described embodiments, the Huffman encoding and the Golomb-Rice encoding are used as the encoding means, but another encoding such as arithmetic encoding or the like may be used.
As described above, according to the image encoding apparatus of the present invention, the encoding can be effectively performed even for the image data having such the a discrete pixel value as seen in the CG image, the limited-color image or the like.
Especially, in the first and second embodiments, it is detected whether or not each prediction error is generated even once every time the predetermined unit of image (i.e., eight lines) is encoded. Then, on the basis of the detected result, an interpretation of the prediction error in case of encoding next time is changed, whereby the compression efficiency in the prediction encoding (i.e., variable length encoding) can be improved.
(Third Embodiment)
In
In the present embodiment, encoding is performed based on two-path structure structures. That is, in a first path, the Huffman table is structed constructed in the Huffman table memory 407, and in a second path, the practical encoding is performed by using the structured Huffman table.
It will be explained in the present embodiment a case where an eight-bit (i.e., values “0” to “255”) monochrome image is encoded. Before the encoding is performed, all the flags f(e) held in the inspection circuit 411 are set as “0”. Further, the prediction value correction circuit 405 has therein a memory region which holds the number N(S) of generation of the prediction error e for each context S and a cumulative value E(S) of the error for each state. However, as an initial state, all the values in the circuit 405 have been set as “0”.
Firstly, image data is sequentially inputted from the signal line 401. The buffer 402 stores the image data of two lines inputted from the signal line 401. The predictor 404 reads peripheral pixels a, b and c of an encode encoding target pixel x from the buffer 402, and produces a prediction value p by calculating p=a+b−c.
Positions of the peripheral pixels a, b and c are shown in FIG. 3. Similarly, the context generator 403 fetches the peripheral pixels a, b and c to obtain values (a−c) and (b−c). Then, like the first embodiment, the context generator 403 obtains each of values Q(a−c) and q(b−c) by qunatizing each of the two values (a−c)and (b−c) into three levels in accordance with the table shown in FIG. 9. By using such the results, the context generator 403 generates the context S which represents the states of the peripheral pixels expressed by “0” and “δ”, by calculating S=3×q(a−c)+q(b−c).
The prediction value correction circuit 405 inputs the context S outputted by the context generator 403, and reads from the internal memory region in the number N(S) of generation of the prediction error e for each context S and the cumulative value E(S) of the prediction error e for each context S.
Subsequently, the prediction value correction circuit 405 obtains a mean value merr(S) of the prediction error in the context S by calculating E(S)/N(S). Then, the inspection circuit 411 checks the value of a flag f(−merr(S)) which stands the flag “1” for the already-generated value, in a prediction error e2 between the prediction value p and the encode encoding target pixel value x. If the flag f(−merr(S)) is “1”, the prediction value correction circuit 405 adds the mean value merr(S) to the prediction value p outputted by the predictor 404, and outputs the obtained value as a corrected prediction value p′=p+merr(S). In this case, the mean value merr(S) may be controlled to have positive value.
On the other hand, if the mean value f(−merr(S)) is “0”, the prediction value correction circuit 405 outputs the prediction value p as it is, as the corrected prediction value p′, The subtracter 409 obtains the prediction error e=x−p′ between the encode encoding target pixel x and the corrected prediction value p′, and then outputs the obtained prediction error e.
A subtracter 412 obtains and outputs the prediction error e2 which is the difference between the encode encoding target pixel x and the prediction value p. The inspection circuit 411 checks the flag value f(e2) corresponding to the differential value e2. Then, if the flag value f(e2) is “0”, the circuit 411 rewrites it into “1”.
The prediction value correction circuit 405 performs increment on the number N(S) of the prediction error e for each context S, newly adds the prediction error e generated by the subtracter 409 to the cumulative value E(S) and stores the obtained value. The Huffman table generator 406 counts the number of generation of the prediction error e for each context S which is derived from the context generator 403.
By repeatedly performing the above-described process for all the image data inputted from the signal line 401, nine frequency distributions each corresponding to each context S are generated in the internal memory of the Huffman table generator 406. In this case, the nine Huffman tables are structured respectively from these nine frequency distributions and then stored in the Huffman table memory 407.
In the second path, the nine Huffman tables stored in the memory 407 are appropriately switched and used for each context S, whereby the encoding and outputting are performed.
When the above-described process is terminated, the next image data beginning from the first pixel is again sequentially inputted into the signal line 401, and the prediction error e is generated for each pixel in the same operation as that in the first path. The Huffman encoder 408 performs the Huffman encoding on the prediction error e by referring to the Huffman table corresponding to each context S stored in the Huffman table memory 407, and outputs the obtained data.
The above-described process is repeatedly performed for the entire encoding target pixels, and thus the encode encoding data is outputted.
The present invention is not restricted to the above-described embodiments. For example, as the prediction method of the encoding target pixel value, the pre-predicting may be simply used. On the other hand, by providing the several prediction methods, these methods may be appropriately switched or exchanged.
Further, in the above-described embodiments, the Huffman encoding and the Golomb-Rice encoding are used as the encoding means, but another encoding such as the arithmetic encoding or the like may be used.
It will be obviously understood that the object of the present invention can be achieved by supplying a storage medium in which program codes of a software to realize the functions of the above-described embodiments are stored to a system or an apparatus, and reading and executing the program codes stored in the storage medium with a computer (CPU or MPU) in the system or the apparatus.
In such a case, the program codes themselves of the software read out of the storage medium realize the functions of the above-described embodiments, thus the storage medium in which the program codes have been stored construct the present invention.
As such the storage medium to store the program codes, e.g., it is possible to use a floppy disk, a hard disk, an optical disk, an optomagnetic disk, a CD-ROM, a CD-R, a magnetic tape, a non-volatile memory card, a ROM or the like can be used.
Further, it will be obviously understood that the present invention incorporates not only a case where the functions of the above-described embodiments are realized by executing the program codes read by the computer, but also a case where an OS (operating system) or the like operating in the computer performs a part or all of the actual processes on the basis of instructions of the read program codes and by such the processes the functions of the above-described embodiments are realized.
Furthermore, it will be obviously understood that the present invention also incorporates a case where the program codes read out of the storage medium are written into a memory provided for a function expansion board of the computer or a function expansion unit connected to the computer and, after that, the CPU or the like provided for the function expansion board or the function expansion unit executes a part or all of the actual processes on the basis of instructions of the program codes, and the functions of the above-described embodiments are realized by the processes.
As explained above, in the case where the image to be encoded consists of the discrete pixel values, the image encoding apparatus according to the present invention performs the controlling such that the correction of the prediction value by the prediction value correction means is not performed. Therefore, even in the case of encoding the data of the image such as the CG image, the limited-color image or the like having the discrete pixel value, the encoding can be effectively performed by utilizing bias of the prediction error.
Many modifications and variations are possible for the present invention, within the spirit and scope of the appended claims.
Patent | Priority | Assignee | Title |
7529417, | Jun 06 2000 | Canon Kabushiki Kaisha | Apparatus, method and storage medium for image encoding/decoding using shape-based coefficient interpolation |
7689048, | May 19 2005 | Canon Kabushiki Kaisha | Image encoding apparatus, method, and computer-readable storage medium for encoding a pixel value |
7783119, | Mar 17 2006 | Canon Kabushiki Kaisha | Image encoding apparatus, image decoding apparatus and control method therefor |
7800520, | Apr 13 2007 | Apple Inc. | Method and system for entropy coding |
7804428, | Nov 10 2008 | Apple Inc. | System and method for compressing a stream of integer-valued data |
8170113, | Dec 16 2005 | Fuji Xerox Co., Ltd. | Encoder, data file, encoding method, computer readable medium and computer data signal |
8213729, | Apr 06 2007 | Canon Kabushiki Kaisha | Multidimensional data encoding apparatus and decoding apparatus, and control method thereof |
8260072, | Aug 07 2008 | Canon Kabushiki Kaisha | Image encoding apparatus and method of controlling the same |
Patent | Priority | Assignee | Title |
4363036, | Mar 16 1981 | NCR CANADA LTD - NCR CANADA LTEE, MISSISSAUGA, ONTARIO, CANADA, A CORP OF CANADA | Method and apparatus for compressing digital data using non-adaptive predictive techniques |
4749983, | Apr 29 1986 | International Business Machines Corporation | Compression of multilevel signals |
5471207, | Feb 23 1994 | Ricoh Corporation | Compression of palettized images and binarization for bitwise coding of M-ary alphabets therefor |
5680129, | Jul 18 1995 | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | System and method for lossless image compression |
5761342, | Apr 16 1991 | Canon Kabushiki Kaisha | Image processing apparatus and method |
5764374, | Aug 30 1996 | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | System and method for lossless image compression having improved sequential determination of golomb parameter |
5764804, | Oct 14 1993 | Seiko Epson Corporation | Data encoding and decoding system |
5945930, | Nov 01 1994 | Canon Kabushiki Kaisha | Data processing apparatus |
6031938, | Apr 26 1995 | Canon Kabushiki Kaisha | Image encoding apparatus with selective Markov and predictive coding |
6101282, | Jun 22 1995 | Canon Kabushiki Kaisha | Apparatus and method for image data encoding |
6233355, | Apr 02 1997 | Canon Kabushiki Kaisha | Encoding apparatus and method and storage medium |
JP8130649, | |||
JP8298599, | |||
JP8331389, | |||
JP9009264, | |||
JP9037246, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Apr 09 2001 | Canon Kabushiki Kaisha | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Jul 21 2011 | M1553: Payment of Maintenance Fee, 12th Year, Large Entity. |
Date | Maintenance Schedule |
Jan 01 2011 | 4 years fee payment window open |
Jul 01 2011 | 6 months grace period start (w surcharge) |
Jan 01 2012 | patent expiry (for year 4) |
Jan 01 2014 | 2 years to revive unintentionally abandoned end. (for year 4) |
Jan 01 2015 | 8 years fee payment window open |
Jul 01 2015 | 6 months grace period start (w surcharge) |
Jan 01 2016 | patent expiry (for year 8) |
Jan 01 2018 | 2 years to revive unintentionally abandoned end. (for year 8) |
Jan 01 2019 | 12 years fee payment window open |
Jul 01 2019 | 6 months grace period start (w surcharge) |
Jan 01 2020 | patent expiry (for year 12) |
Jan 01 2022 | 2 years to revive unintentionally abandoned end. (for year 12) |