A method and apparatus for transforming image data by recursively interleaving the data to generate blocks of component image coefficients having form suitable for subsequent quantization, motion estimation, and/or coding. In preferred embodiments, the transformed data are in optimal form for coding by conventional circuitry in accordance with the conventional JPEG or MPEG compression algorithm. In preferred embodiments, the invention includes two memory arrays (each having capacity to store one or more N×M image data blocks), and an analyzer connected between the memory arrays. The analyzer receives horizontal vectors (such as full rows) of an image data block stored in the first memory, transforms each horizontal vector into two vectors (each comprising half as many words as the horizontal vector), interleaves the two vectors, and writes the resulting interleaved data (an orthogonal representation of the horizontal vector) into a row of the second memory. The analyzer then sequentially receives vertical vectors (such as columns) of an image data block stored in the second memory, converts each vertical vector into two vectors (each comprising half as many words as the vertical vector), interleaves the vectors, and writes the resulting interleaved data into a column of the first memory. Typically, multiple iterations are performed. After each iteration, the first memory contains a set of interleaved component image blocks. Preferably, the analyzer is a wavelet transform module including a pair of conjugate mirror filters and an interleaving circuit.

Patent
   5414780
Priority
Jan 27 1993
Filed
Jan 27 1993
Issued
May 09 1995
Expiry
Jan 27 2013
Assg.orig
Entity
Large
102
11
EXPIRED
15. A method for transforming input image data using an apparatus including a first memory having at least n rows and M columns of memory locations, where n and M are integers, and a second memory having at least n rows and M columns of memory locations, said method including the steps of:
(a) receiving horizontal vectors of data stored in the first memory, transforming each of the horizontal vectors into an interleaved orthogonal representation thereof, and writing each said interleaved orthogonal representation into a different row of the second memory; and
(b) after step (a), receiving vertical vectors of data stored in the second memory, transforming each of the vertical vectors into a second interleaved orthogonal representation thereof, and writing each said second interleaved orthogonal representation into a different column of the first memory, wherein each said interleaved orthogonal representation consists of interleaved portions of one of the horizontal vectors received from the first memory, and each said second interleaved orthogonal representation consists of interleaved portions of one of the vertical vectors received from the second memory.
8. An apparatus for transforming input image data, including:
a first memory having at least n rows and M columns of memory locations, where n and M are integers;
a second memory having at least n rows and M columns of memory locations;
a first analyzer connected between the first memory and the second memory, including means for receiving horizontal vectors of data stored in the first memory, transforming each of the horizontal vectors into an interleaved orthogonal representation thereof, and writing each said interleaved orthogonal representation into a different row of the second memory; and
a second analyzer connected between the first memory and the second memory, including means for receiving vertical vectors of data stored in the second memory after each said interleaved orthogonal representation has been written into said second memory, transforming each of the vertical vectors into a second interleaved orthogonal representation thereof, and writing each said second interleaved orthogonal representation into a different column of the first memory, wherein each said interleaved orthogonal representation consists of interleaved portions of one of the horizontal vectors received from the first memory, and each said second interleaved orthogonal representation consists of interleaved portions of one of the vertical vectors received from the second memory.
1. An apparatus for transforming input image data, including:
first memory means having at least n rows and M columns of memory locations, where n and M are integers divisible by 2K where K is a positive integer;
second memory means having at least n rows and M columns of memory locations;
analyzer means connected between the first memory means and the second memory means, including means for receiving horizontal vectors of data stored in the first memory means, transforming each of the horizontal vectors into an interleaved orthogonal representation thereof, writing each said interleaved orthogonal representation into a different row of the second memory means, receiving vertical vectors of data stored in the second memory means, transforming each of the vertical vectors into a second interleaved orthogonal representation thereof, and writing each said second interleaved orthogonal representation into a different column of the first memory means, wherein each said interleaved orthogonal representation consists of interleaved portions of one of the horizontal vectors received from the first memory means, and each said second interleaved orthogonal representation consists of interleaved portions of one of the vertical vectors received from the second memory means; and
control means for controlling the analyzer means to implement K transformation iterations, each of said iterations including transformation by the analyzer means of a set of horizontal vectors of data stored in the first memory means followed by further transformation by the analyzer means of a set of vertical vectors of data stored in the second memory means.
21. (Amended) A method for compressing input image data, using an apparatus including a first memory having at least n rows and M columns of memory locations, where n and M are integers, and a second memory having at least n rows and M columns of memory locations, said method including the steps of:
(a) loading an N×M image data block into the first memory;
(b) receiving horizontal vectors of the data block stored in the first memory, transforming each of the horizontal vectors into an interleaved orthogonal representation thereof, and writing each said interleaved orthogonal representation into a different row of the second memory;
(c) after step (b), receiving vertical vectors of data stored in the second memory, transforming each of the vertical vectors into a second interleaved orthogonal representation thereof, and writing each said second interleaved orthogonal representation into a different column of the first memory, wherein each said interleaved orthogonal representation consists of interleaved portions of one of the horizontal vectors received from the first memory, and each said second interleaved orthogonal representation consists of interleaved portions of one of the vertical vectors received from the second memory;
(d) performing K iterations of steps (b) and (c) in such a manner that each of the horizontal vectors consists of M/2K-1 words during a kth one of such iterations, each of the vertical vectors consists of n/2K-1 words during the kth one of the iterations, and following the kth one of the iterations of step (c), the first memory contains a set of interleaved component image blocks defining a pyramidal representation of the image data block; and
(e) after step (d), compressing the set of interleaved component image blocks stored in the first memory by performing quantization and coding operations thereon, thereby generating a compressed block of image data.
20. A method for transforming input image data using an apparatus including a first memory having at least n rows and M columns of memory locations, where n and M are integers, and a second memory having at least n rows and M columns of memory locations, wherein each of M and n is divisible by 2K, and K is a positive integer, said method including the steps of:
(a) loading an N×M image data block into the first memory;
(b) receiving horizontal vectors of the data block stored in the first memory, transforming each of the horizontal vectors into an interleaved orthogonal representation thereof, and writing said interleaved orthogonal representation into a different row of the second memory;
(c) after step (b), receiving vertical vectors of data stored in the second memory, transforming each of the vertical vectors into a second interleaved orthogonal representation thereof, and writing each said second interleaved orthogonal representation into a different column of the first memory, wherein each said interleaved orthogonal representation consists of interleaved portions of one of the horizontal vectors received from the first memory, and each said second interleaved orthogonal representation consists of interleaved portions of one of the vertical vectors received from the second memory;
(d) performing K iterations of steps (b) and (c) in such a manner that each of the horizontal vectors consists of M/2K-1 words during a kth one of such iterations, each of the vertical vectors consists of n/2K-1 words during the kth one of the iterations, and following the kth one of the iterations of step (c), the first memory contains a set of interleaved component image blocks defining a pyramidal representation of the image data block; and
(e) after step (d), loading a second N×M image data block into the first memory, and repeating steps (b) through (d) to process said second N×M image data block.
25. An apparatus for compressing input image data, including:
a first memory means having at least n rows and M columns of memory locations, where n and M are integers, with an N×M image data block stored in the first memory means;
a second memory means having at least n rows and M columns of memory locations;
analyzer means connected between the first memory means and the second memory means, including means for receiving horizontal vectors of data stored in the first memory means, transforming each of the horizontal vectors into an interleaved orthogonal representation thereof, writing each said interleaved orthogonal representation into a different row of the second memory means, receiving vertical vectors of data stored in the second memory means, transforming each of the vertical vectors into a second interleaved orthogonal representation thereof, and writing each said second interleaved orthogonal representation into a different column of the first memory means, wherein each said interleaved orthogonal representation consists of interleaved portions of one of the horizontal vectors received from the first memory means, and each said second interleaved orthogonal representation consists of interleaved portions of one of the vertical vectors received from the second memory means;
control means for controlling the analyzer means to implement K transformation iterations, each of said iterations including transformation by the analyzer means of a set of horizontal vectors of data stored in the first memory means followed by further transformation by the analyzer means of a set of vertical vectors of data stored in the second memory means, wherein following the kth one of the iterations the first memory means contains a set of interleaved component image blocks defining a pyramidal representation of the image data block; and
means for compressing the set of interleaved component image blocks stored in the first memory means by performing quantization and coding operations thereon, thereby generating a compressed block of image data.
2. The apparatus of claim 1, wherein the analyzer means includes:
a wavelet transform module including a first pair of conjugate mirror filters, and a first interleaving circuit connected to outputs of the mirror filters, and
wherein the control means controls the wavelet transform module to operate in a selected one of a first mode in which inputs of the mirror filters are connected to the first memory means and an output of the interleaving circuit is connected to the second memory means, and a second mode in which the inputs of the mirror filters are connected to the second memory means and the output of the interleaving circuit is connected to the first memory means.
3. The apparatus of claim 2, wherein the first memory means includes a first memory having at least n rows and M columns of memory locations and a second memory having at least n rows and M columns of memory locations, wherein the second memory means includes a third memory having at least n rows and M columns of memory locations and a fourth memory having at least n rows and M columns of memory locations, and also including:
switching means connecting the wavelet transform module with the first memory means and the second memory means, the switching means having a first state connecting the wavelet transform module between the first memory and the third memory, and a second state connecting the wavelet transform module between the second memory and the fourth memory.
4. The apparatus of claim 1, wherein the analyzer means includes:
a first wavelet transform module including a first pair of conjugate mirror filters and a first interleaving circuit connected to outputs of the first pair of conjugate mirror filters, wherein inputs of the first pair of mirror filters are connected to the first memory means and an output of the first interleaving circuit is connected to the second memory means; and
a second wavelet transform module including a second pair of conjugate mirror filters and a second interleaving circuit connected to outputs of the second pair of conjugate mirror filters, wherein inputs of the second pair of conjugate mirror filters are connected to the second memory means and an output of the second interleaving circuit is connected to the first memory means.
5. The apparatus of claim 1, wherein the first memory means contains an N×M image data block at commencement of a first one of the transformation iterations, and wherein the control means controls the analyzer means so that, following a kth one of the transformation iterations, the first memory means contains a set of interleaved component image blocks defining a pyramidal representation of the image data block arranged in a repeating pattern of 2K ×2K blocks, each of said 2K ×2K blocks consisting of interleaved words from different component image blocks.
6. The apparatus of claim 1, wherein N=M.
7. The apparatus of claim 6, wherein M=16.
9. The apparatus of claim 8, wherein the first analyzer is a wavelet transform module including a first pair of conjugate mirror filters and a first interleaving circuit, and the second analyzer is a wavelet transform module including a second pair of conjugate mirror filters and a second interleaving circuit.
10. The apparatus of claim 8, wherein each of M and n is divisible by 2K, where K is a positive integer, and also including:
control means for controlling the first analyzer and the second analyzer to implement K transformation iterations, each of said iterations including transformation by the first analyzer of a set of horizontal vectors of data stored in the first memory followed by further transformation by the second analyzer of a set of vertical vectors of data stored in the second memory.
11. The apparatus of claim 10, wherein the first memory contains an N×M image data block at commencement of a first one of the transformation iterations, and wherein the control means controls the first analyzer and the second analyzer so that, following a kth one of the transformation iterations, the first memory contains a set of interleaved component image blocks defining a pyramidal representation of the image data block arranged in a repeating pattern of 2K ×2K blocks, each of said 2K ×2K blocks consisting of interleaved words from different component image blocks.
12. The apparatus of claim 8, wherein N=M=16.
13. The apparatus of claim 12, also including:
control means for controlling the first analyzer and the second analyzer to implement three transformation iterations on a 16×16 word data block stored in the first memory, each of said iterations including transformation of a set of horizontal vectors of data stored in the first memory by the first analyzer followed by further transformation of a set of vertical vectors of data stored in the second memory by the second analyzer.
14. The apparatus of claim 13, wherein the control means controls the first analyzer and the second analyzer so that each of the horizontal vectors and the vertical vectors during a first of the iterations consists of sixteen words, each of the horizontal vectors and the vertical vectors during a second of the iterations consists of eight words, and each of the horizontal vectors and the vertical vectors during a third of the iterations consists of four words.
16. The method of claim 15, wherein each of M and n is divisible by 2K, where K is a positive integer, and also including the step of:
(c) performing K iterations of steps (a) and (b) in such a manner that each of the horizontal vectors consists of M/2K-1 words during a kth one of such iterations, and each of the vertical vectors consists of n/2K-1 words during the kth one of the iterations.
17. The method of claim 16, wherein M=N.
18. The method of claim 17, wherein M=16.
19. The method of claim 17, wherein the first memory contains an M×M image data block upon commencement of the first iteration of step (a), and wherein following a kth one of the iterations of step (b), the first memory contains a set of interleaved component image blocks defining a pyramidal representation of the image data block arranged in a repeating pattern of 2K ×2K blocks, each of said 2K ×2K blocks consisting of interleaved words from different component image blocks.
22. The method of claim 21, also including the step of:
(f) after step (e), decompressing the compressed block of image data by performing an inverse coding operation and then an inverse quantization operation thereon, thereby generating a reconstructed set of interleaved component image blocks.
23. The method of claim 21, wherein M=N.
24. The method of claim 23, wherein M=16.
26. The apparatus of claim 25, also including:
means for decompressing the compressed block of image data by performing an inverse coding operation and then an inverse quantization operation thereon, thereby generating a reconstructed set of interleaved component image blocks.
27. The apparatus of claim 25, wherein the control means controls the analyzer means to perform the K transformation iterations in such a manner that following a kth one of the transformation iterations, the first memory means contains a set of interleaved component image blocks defining a pyramidal representation of the image data block arranged in a repeating pattern of 2K ×2K blocks, each of said 2K ×2K blocks consisting of interleaved words from different component image blocks.
28. The apparatus of claim 25, wherein the analyzer means is a wavelet transform module including a pair of conjugate mirror filters and an interleaving circuit.
29. The apparatus of claim 25, wherein N=M.
30. The apparatus of claim 29, wherein M=16.

The present invention relates to methods and apparatus for transforming image data (such as video data) for subsequent quantization, motion estimation, and/or coding. More particularly, the invention pertains to recursive interleaving of image data to generate blocks of component image coefficients having form suitable for subsequent quantization, motion estimation, and/or coding.

Most image sensors and displays generate or accept image signals in color raster scan format, in which pixels comprising a first horizontal line are generated or displayed sequentially (from left to right), and pixels comprising the next line are then generated or displayed sequentially (from left to right), and so on. In many conventional color image display devices, each pixel is driven by a set of three analog color component signals (a red, a green, and a blue color component signal). Typically, each analog color component signal is generated by processing a multi-bit digital data word in a digital-to-analog conversion circuit.

If a set of analog or digital image data in color raster scan format represents a monochrome image, the data are said to be in "line-scan" format. If a set of analog or digital data in color raster scan format represents a color image, the individual color component signals are typically interleaved.

It is well known to perform image compression on digital image data to generate a reduced set of (compressed) data from which the original image can be reconstructed without loss of essential features. The compressed data can be transmitted (or stored) more efficiently than can the original image data. An inverse (decompression) transformation can be applied to the transmitted data (or the data read out from storage) to recover the original image (or a reasonable facsimile thereof).

Throughout this specification, including in the claims, "block" denotes an array of N×M samples of a given color component (N and M are integers), "word" denotes a color component sample (e.g., an analog red, green, or blue sample of an analog image representation, or an eight-bit digital word defining a red, green, or blue sample of an analog image representation), and "line length" denotes the number of words per line of an image signal (for the color component having the highest horizontal resolution, in the case of color image data in which the different color components have different resolution).

Most image compression algorithms do not process image data in line-scan format, and instead process image data in N'×M' block format. For example, the conventional image compression algorithms known as the ISO "JPEG" algorithm (for still images) and the ISO "MPEG" algorithm (for video signal compression) both process image data in 8×8 block format (M'=N =8). Examples of such input data include: a repeating sequence of an 8×8 block of red words, followed by an 8×8 block of green words, followed by an 8×8 block of blue words (image processors for processing "RGB(1:1:1)" images will expect the input data to have this format); and a repeating sequence of two 8×8 blocks of Y words, followed by an 8×8 block of U words, followed by an 8×8 block of V words (image processors for processing "YUV(2:1:1)" images will expect the input data to have this format).

Typical algorithms for performing image compression on digital image data include two steps: a transformation step which generates transformed image data (in which the correlation between adjacent pixels is reduced relative to that existing in the input image data); followed by a quantization step which replaces each pixel of the transformed image data with a quantized pixel comprising fewer bits (on the average). To reduce loss of information during the quantization step, it has been proposed to design the transformation step so that the transformed image data is a set of component image signals having different spatial frequencies (so that the transformed image data is a "pyramidal" or "multiresolution" representation of the image).

For example, U.S. Pat. No. 5,014,134, issued May 7, 1991, discloses a method and apparatus for performing image compression in which image data are transformed into a pyramidal image representation. The apparatus of U.S. Pat. No. 5,014,134 includes a first transformation circuit ("analyzer") which converts each row of an M×M block of input image data into two vectors, yL and yH, each comprising M/2 words. Vectors yL define an M row×M/2 column component image representation L (representing relatively low spatial frequency information), and vectors yH define an M×M/2 component image representation H (representing relatively high spatial frequency information). The apparatus of U.S. Pat. No. 5,014,134 includes two additional analyzers. One of the additional analyzers (the "second" analyzer) receives a sequence of column vectors of image L and converts each such column vector into two column vectors yLL and yLH (each comprising M/2 words). The other of the additional analyzers (the "third" analyzer) receives a sequence of column vectors of image H and converts each such column vector into two column vectors yHL and yHH (each comprising M/2 words). The outputs of the second analyzer determine two M/2×M/2 component images (LL and LH), and the outputs of the third analyzer determine two M/2×M/2 component images (HL and HH). Image LL represents the lowest spatial frequency information of the original image, and images LH, HL, and HH represent higher spatial frequency information of the original image. Component images LL, LH, HL, and HH (each of which is a M/2×M/2 image data block, and which together define a pyramidal representation of the original image) are then separately quantized and coded to generate compressed image data representing a compressed version of the original image.

The upper left image representation, I, in FIG. 1 represents the original image, the upper right image representation in FIG. 1 represents component images L and H, and the lower left image representation in FIG. 1 represents component images LL, LH, HL, and HH (which together define a pyramidal representation of the original image).

The apparatus of U.S. Pat. No. 5,014,134 also discloses means for reconstructing the original image, which generates one M×M block of reconstructed image data from the compressed image data. The reconstruction means includes a dequantizer and decoder for reconstructing the M/2×M/2 component image representations LL, LH, HL, and HH from the compressed image data, and three transformation circuits ("synthesizers"). The first synthesizer transforms the reconstructed (decompressed) HH and HL component images into a reconstructed image representation H (an M×M/2 component image representation). The second synthesizer transforms the reconstructed (decompressed) LH and LL component images into a reconstructed image representation L (also an M×M/2 component image representation). The third synthesizer receives the reconstructed image representations L and H, and transforms them into a reconstructed image representation (an M×M representation of the original image).

U.S. Pat. No. 5,014,134 also teaches an iterative pyramidal representation generation process, in which each iteration consists of transforming the component image representation having the lowest spatial frequency (the LL component image representation) generated during previous iteration. The lower right image representation of FIG. 1 represents the result of performing a second iteration of this type on the lower left image representation of FIG. 1. Specifically, the lower right image in FIG. 1 represents component images LLLL, LLHL, LLLH, LLHH, LH, HL, and HH (which together define a pyramidal representation of the original image, with component images LLLL, LLHL, LLLH, LLHH together defining a pyramidal representation of component image LL generated during the first iteration).

However, the methods and apparatus disclosed in U.S. Pat. No. 5,014,134 for transforming image data into pyramidal representations do not result in pyramidal representations optimal for subsequent quantization and coding (particularly for coding in accordance with the conventional ISO "JPEG" or "MPEG" image compression algorithm).

The invention is a method and apparatus for transforming image data by recursively interleaving the image data to generate blocks of component image coefficients having form suitable for subsequent quantization, motion estimation, and/or coding. In preferred embodiments, the transformed data generated in accordance with the invention are in optimal form for subsequent coding in accordance with the conventional ISO "JPEG" or "MPEG" image compression algorithm.

In a class of preferred embodiments, the apparatus of the invention includes two memory arrays (each having sufficient capacity to store one or more blocks of N×M image data words), a first analyzer circuit connected between the memory arrays, and a second analyzer circuit connected between the memory arrays. M and N are integers. Typically, each of M and N is divisible by 2k, where k is the number of recursive decomposition levels of the implemented transformation.

The first analyzer sequentially receives horizontal vectors (such as full rows) of an N×M image data block stored in the first memory, transforms each horizontal vector into two vectors (each comprising half as many words as the horizontal vector), interleaves the two vectors, and writes the resulting interleaved data (an orthogonal representation of the horizontal vector) into a row of the second memory. The second analyzer sequentially receives vertical vectors (such as columns) of the image data stored in the second memory, converts each vertical vector into two vectors (each comprising half as many words as the vertical vector), interleaves the vectors, and writes the resulting interleaved data into a column of the first memory.

Typically, multiple iterations, each comprising transformation of data in the first memory by the first analyzer followed by further transformation of the data in the second memory by the second analyzer. After each iteration, the first memory contains a set of interleaved component image blocks.

More specifically, during the first iteration, the first analyzer converts each M-word row of the first memory into two vectors φ and ψ (each comprising M/2 words or "coefficients"), interleaves the vectors φ and ψ for each row, and writes the resulting interleaved vector into a row of the second memory. Each pair of vectors φ and ψ together determines an orthogonal representation of the row processed by the first analyzer. Then (also during the first iteration), the second analyzer sequentially receives N-word columns of (partially transformed) data from the second memory, converts each column into two vectors, φ and ψ (each comprising N/2 words), interleaves the vectors φ and ψ, and writes the resulting interleaved column vector into a column of the first memory.

During the second iteration, the first analyzer converts M/2-word horizontal vectors stored in the first memory into two vectors φ and ψ (each comprising M/4 words), interleaves the vectors φ and ψ for each row, and writes the resulting interleaved vector into a subset of the memory locations which comprise a row of the second memory. Then, the second analyzer sequentially receives N/2-word vertical vectors of data from the second memory, converts each vertical vector into two vectors φ and ψ (each comprising N/4 words), interleaves the vectors φ and ψ, and writes the resulting interleaved vertical vector into a subset of the memory locations which comprise a column of the first memory.

Typically, multiple iterations are performed. If the first memory and the second memory each has an M×M array of storage locations, the content of the first memory at the end of the "k"th iteration is a set of interleaved component image blocks, which are a pyramidal representation of the original image arranged in a repeating pattern of 2k ×2k blocks, each 2k ×2k block consisting of interleaved words from different component image blocks (corresponding to different spatial frequencies). During the "k"th iteration, the first analyzer processes horizontal vectors each consisting of Mk-1 words and the second analyzer processes vertical vectors each consisting of Nk-1 words.

In preferred embodiments, each analyzer is a wavelet transform module including a pair of conjugate mirror filters and an interleaving circuit.

In other preferred embodiments, the apparatus of the invention includes a single analyzer (connected between first and second memories) and a control means for controlling the analyzer. The control means causes the analyzer to perform not only the operations performed by the first analyzer (in those embodiments of the invention which employ two analyzers) but also those performed by the second analyzer (in embodiments with two analyzers).

To increase the rate at which the inventive apparatus can process data (for example, to permit processing of video images in real time, i.e., at a rate of 60 fields per second), the apparatus includes two pairs of memories (each memory having capacity to store a field of data, where the field of data comprises an integral number of N×M data blocks), two quad port bus switches (one connected between each pair of memories), and an analyzer circuit connected between the switches. The switches are controlled so that, at any time, the analyzer reads and writes data between a first memory and a second memory to implement the inventive transformation on a N×M data block in the first memory, a previously transformed field of N×M data blocks is read out of a third memory, and a new field of N×M data blocks is written into the fourth memory .

FIG. 1 is a diagram representing a conventional transformation of an image signal into a set of component image signals having different spatial frequencies, known as a "pyramidal" representation.

FIG/ 2 is a block diagram of a first preferred embodiment of the inventive apparatus.

FIG. 3 is a diagram representing a six-step transformation of an image signal in accordance with the invention, to generate an output signal which is a pyramidal representation of the image signal.

FIG. 4 is a diagram of the output signal of FIG. 3 with its component signals reordered according to spatial frequency.

FIG. 5 is a block diagram of an image compression and transmission (or storage) system which embodies the invention.

FIG. 6 is a diagram of a sequence in which the coefficients of each 8×8 block 42A, 42B, 42C, and 42D (shown in FIG. 3) are typically reordered by coding circuit 56 (of FIG. 5) to implement coding in an efficient manner.

FIG. 7 is a block diagram of a second preferred embodiment of the inventive apparatus (which is preferably employed, as a substitute for the FIG. 3 apparatus, to implement either or both of image transformation circuit 52 and inverse image transformation circuit 64 of the FIG. 5 system).

FIG. 8 is a diagram representing the sequence in which the FIG. 7 apparatus processes data in the data compression mode.

FIG. 2 is a block diagram of a preferred embodiment of the invention. This embodiment includes first memory array 2 and second memory array 4, each having sufficient capacity to store one or more blocks of N×M image data words (where M and N are integers), first analyzer circuit 6 connected between memory arrays 2 and 4, and second analyzer circuit 8 connected between memory arrays 2 and 4. Typically, the FIG. 2 apparatus processes binary data, and each of integers M and N is divisible by 2k, where k is the number of recursive decomposition levels of the implemented transformation (i.e., the number of iterations in which data are processed first in analyzer 6 and then in analyzer 8).

In operation, an N×M block of image data is initially stored in memory array 2. During the first iteration, analyzer 6 sequentially receives rows of the stored block and converts each row (comprising M words) into two vectors, φ and ψ (each comprising M/2 words or "coefficients"). Analyzer 6 then interleaves the vector pair φ and ψ for each row, and writes the resulting interleaved vector for each row (comprising M coefficients) into a row of memory array 4.

In the preferred embodiment of FIG. 2, analyzer 6 is a wavelet transform module which includes a pair of conjugate mirror filters (10 and 12) and an interleaving circuit 14. For each horizontal vector xi from memory 2 (which can be a full row defined by M'=M words xi, where 0≦i≦M-1, or an M'-word subset of a row, where M'<M), filter 10 outputs vector φm having form φm =Σ ak x2m+k, where k ranges from 0 through 2X-1, X is the order of the wavelet basis, ak are the coefficients of the wavelet basis, m is an integer in the range from 0 through (M'/2-1), and xi+M, is defined to equal xi. Similarly, for each horizontal vector xi from memory 2, filter 12 outputs vector ψm having form ψm =Σ(-1)k a2X-1-k x2m+k, where k ranges from 0 through 2X-1, X is the order of the wavelet basis, 0≦m≦M'/2-1, and xi+M, is defined to equal xi. Each pair of vectors φ and ψ together determines an orthogonal representation of the horizontal vector processed by the first analyzer.

Interleaver 14 interleaves the words of each vector φ with the words of the corresponding vector ψ, to generate an interleaved, transformed horizontal vector 15 (comprising M' words). Each interleaved vector 15 is written into a different row of memory array 4. During the first iteration of a transformation, analyzer 6 typically processes full rows of data stored in memory 2 (each row comprising M'=M words), so that each transformed horizontal vector 15 comprises M words during the first iteration.

After analyzer 6 sequentially processes all necessary horizontal vectors xi from memory 2 (and causes the processed, interleaved data to be written into memory 4), analyzer 8 sequentially processes vertical vectors (which can be full columns defined by N'=N words yi, where 0≦i≦N-1, or an N'-word subset of a column, where N'<N) of the transformed N×M block in memory 4.

Analyzer 8 sequentially receives vertical vectors of the N×M block in memory 4, converts each vertical vector (comprising N' words) into two vectors, φ and ψ (each comprising N'/2 words), interleaves the vectors φ and ψ, and writes the resulting interleaved, transformed vertical vector into a column of memory 2.

In the preferred embodiment of FIG. 2, analyzer 8 is identical to analyzer 6, except that it is designed to process N'-word input vectors rather than M'-word vectors (however, in a preferred embodiment, M'=N'). Analyzer 8 is thus preferably a wavelet transform module which includes a pair of conjugate mirror filters (16 and 18) and an interleaving circuit 20. For each vertical vector from memory 4 (defined by N' words yi, where 0≦i≦N'-1), filter 16 outputs vector φm having form φm =Σ ak y2m+k, where k ranges from 0 through 2X-1, X is the order of the wavelet basis, m=0≦m≦N'/2-1, and yi+N ' is defined to equal yi. Similarly, for each vertical vector yi from memory 4, filter 18 outputs vector ψm having form ψm =Σ (-1)k a2X-1-k x2m+k, where k ranges from 0 through 2X-1, X is the order of the wavelet basis, 0≦m≦N' /2-1, and yi+N' is defined to equal yi. Vector φ (defined by N'/2 words) and vector ψ (defined by N'/2 words) together determine an orthogonal representation of vertical vector yi from memory 4.

Interleaver 20 interleaves the words of each vector φ with the words of the corresponding vector ψ, to generate an interleaved vertical vector 21 (comprising N' words). Each interleaved vector 21 is written into a different column of memory array 2.

Timing and control circuitry 24 supplies timing and control signals to filters 10, 12, 16, and 18, and interleaving circuits 14 and 20, to control their operation to cause them to implement the inventive method described herein. Typically, circuitry 24 includes two transform address generators (such as address generators 87 and 88 of FIG. 7) each of which provides appropriate addresses for writing vectors of data from one of the memories to one of the analyzers (or for reading interleaved vectors from one of the analyzers to one of the memories), and a transform sequence controller (such as controller 89 of FIG. 7) for controlling the transform address generators.

After a single iteration (in which data from all rows of a block stored in memory 2 are processed by analyzer 6 to write a partially transformed block into memory 4, and data from all columns of the partially transformed block are then processed by analyzer 8 to write a transformed data block back into memory 2), memory 2 contains a set of four interleaved component image blocks, which together define a pyramidal representation of the original image.

Where M=N, and M is divisible by 2k, the component image blocks are "interleaved" in the sense that, the transformed data in memory 2 consist of a repeating pattern of 2k ×2k blocks after each iteration. Typically, multiple iterations are performed (k=1 during the first iteration, and is incremented by one before each subsequent iteration). In the case that M=N (where M is divisible by 2k), after the first iteration, the interleaved component image blocks in memory 2 consist of a repeating pattern of 2 word×2 word blocks, each 2×2 block consisting of a word from a first (low spatial frequency) component image block and one word each from a second, third, and fourth component image block, where the second, third, and fourth component image blocks represent higher spatial frequency information regarding horizontal image features, vertical image features, and diagonal image features, respectively. At the end of the "kth" iteration, the first memory array (memory 2) contains a pyramidal representation of the input image defined by a repeating pattern of 2k ×2k blocks, each 2k ×2k block consisting of interleaved words (when its words are considered sequentially in line-scan format) from a number of component image blocks (corresponding to different spatial frequencies).

In preferred embodiments in which M=N=16, the FIG. 2 apparatus can perform a three-iteration transformation on each 16×16 word image data block (such as block 30 shown in FIG. 3) of an image signal, to generate a transformed data block (such as block 42 shown in FIG. 3) of an output signal which is a pyramidal representation of the image signal. Each such transformed data block consists of four, 8 ×8 word, interleaved data blocks (e.g., blocks 42A, 42B, 42C, and 42D shown in FIG. 3). Each interleaved 8×8 data block (e.g., block 42B of FIG. 3) is in form suitable for efficient processing in conventional JPEG or MPEG coding circuitry. In some preferred embodiments, each word of each 8×8 block comprises 12 bits. In alternative embodiments, each such word can comprise another number of bits (e.g., 8 or 32 bits).

Next, an example of a three-iteration transformation implemented by an embodiment of the FIG. 2 apparatus in which M=N=16 (and memory 2 and memory 4 each have at least 256-word capacity), will be described with reference to FIG. 3. The first step in this transformation is to load input image block 30 (comprising 16×16=256 data words) into memory 2. Typically, image block 30 is a portion of a digital image data stream.

The first iteration includes the following steps. For each 16-word row vector xi of block 30, filter 10 outputs vector φm (of the above-described type, where 0≦m≦15). Similarly, for each row vector xi, filter 12 outputs vector ψm (of the above-described type, where 0≦m≦15). Each pair of vectors φm and ψm together determines an orthogonal representation of row vector xi from memory 2.

Interleaver 14 interleaves the words of each vector φm with the words of the corresponding vector ψm, to generate a sixteen-word interleaved row vector which is written into a row of memory array 4. When filters 10 and 12 and interleaver 14 have processed all sixteen of the row vectors xi of block 30, and have caused the processed and interleaved data to be written into memory 4, memory 4 contains partially transformed 16×16 word block 32. This block consists of data φ1 and data ψ1 indicated in FIG. 3.

Analyzer 8 then sequentially processes each sixteen-word column yi of block 32 by converting each column into two eight-word vectors φ and ψ, interleaving the vectors φ and ψ, and writing the resulting sixteen-word transformed (and interleaved) column vector into a column of memory 2. Specifically, for each odd-numbered column of block 32 (e.g., the first column), filter 16 of analyzer 8 outputs an eight-word vector φφ and filter 18 of analyzer 8 outputs an eight-word vector φψ. For each even-numbered column of block 32 (e.g., the second column), filter 16 outputs eight-word vector ψφ and filter 18 outputs an eight-word vector ψψ. In FIG. 3, vectors φφ, φψ, ψφ, and ψψ, are denoted respectively as φ1, ψ1v, ψ1h, and ψ1d. Vectors φ1 represent relatively low spatial frequency data, and vectors ψ1v, ψ1h, and ψ1d represent relatively high spatial frequency data.

Interleaver 20 interleaves the words of each vector φ1 with the words of the corresponding vector ψ1v to generate an interleaved column vector, and interleaver 20 interleaves the words of each vector ψ1h with the words of the corresponding vector ψ1d to generate an interleaved column vector. After all the interleaved column vectors have been sequentially written into different columns of memory 2, the content of memory 2 is first transformed block 34.

Block 34 consists of sixty-four transformed data blocks, each consisting of 2×2 words, which together define a pyramidal representation of original image data 30. The transformed data blocks are stored in memory 2 as a repeating pattern of 2×1 word blocks, each 2×2 block consisting of a word representing low spatial frequency data and three words representing high spatial frequency data.

With reference again to FIG. 3, the second iteration of the transformation generates second transformed block 38 from first transformed block 34. During this second iteration, filters 10 and 12 are caused by timing and control circuitry 24 to receive and process only odd-numbered words from odd-numbered rows of block 34 (e.g., filter 10 processes the first, third, fifth, seventh, ninth, eleventh, thirteenth, and fifteenth words of the first row of block 34, but none of the words of the second row of block 34). For each of the 8-word row vectors xi written to filter 10 from an odd row of block 34, filter 10 outputs a four-word vector φ2m (of the above-described type, where 0≦m≦7). Similarly, for each of the same 8-word row vectors xi from block 34, filter 12 outputs a four-word vector ψ2m (of the above-described type, where 0≦m≦7).

Interleaver 14 interleaves the words of each vector φ2m with the words of the corresponding vector ψ2m, to generate an eight-word interleaved row vector which is written into the odd-numbered word locations of a row of memory array 4. When filters 10 and 12 and interleaver 14 have processed all eight such row vectors xi of block 34, and have caused the resulting processed and interleaved data to be written into memory 4, memory 4 contains partially transformed 16×16 word block 36.

Analyzer 8 then sequentially processes only the odd-numbered words of the odd-numbered columns of block 36 (under control of timing and control circuitry 24), to generate block 38. Specifically, for each 8-word column vector yi written from the first (or fifth, ninth, or thirteenth) column of block 36, filter 16 outputs a four-word vector φ2 and filter 18 of analyzer 8 outputs a four-word vector ψ2v. For each 8-word column vector yi written from the third (or seventh, eleventh, or fifteenth) column of block 36, filter 16 outputs a four-word vector φ2h and filter 18 outputs a four-word vector ψ2d. Vectors φ2 represent relatively low spatial frequency data, and vectors ψ2v, ψ2h, and ψ2d represent relatively high 2d spatial frequency data.

Interleaver 20 interleaves the words of each vector φ2 with the words of the corresponding vector ψ2v to generate an interleaved column vector, and interleaver 20 interleaves the words of each vector ψ2h with the words of the corresponding vector ψ2d to generate an interleaved column vector. After all the interleaved column vectors have been sequentially written into different columns of memory 2, the content of memory 2 is second transformed block 38.

Block 38 consists of sixteen transformed data blocks, each consisting of 4×4 words, which together define a pyramidal representation of original image data 30. The transformed data blocks are stored in memory 2 as a repeating pattern of 4×4 word blocks, each 4×4 block consisting of one word (from a vector ψ2) representing low spatial frequency data, and fifteen words representing higher spatial frequency data.

Still with reference again to FIG. 3, the third (final) iteration of the transformation generates third transformed block 42 from second transformed block 38. During the third iteration, filters 10 and 12 are caused by timing and control circuitry 24 to receive and process only every fourth word from every fourth row of block 38 (e.g., filter 10 processes the first, fifth, ninth, and thirteenth words of the first row of block 38, but none of the words of the second, third, and fourth rows of block 38). For each of the four-word row vectors xi written to filter 10 from a row of block 38, filter 10 outputs a two-word vector φ3m (of the above-described type, where 0≦m≦3). Similarly, for each of the same four-word row vectors xi from block 38, filter 12 outputs a four-word vector ψ3m (of the above-described type, where 0≦m≦3).

Interleaver 14 interleaves the words of each vector φ3m with the words of the corresponding vector ψ3m, to generate a four-word interleaved row vector whose words are written into every fourth word location of the first (or fifth, ninth, or thirteenth) row of memory array 4. When filters 10 and 12 and interleaver 14 have processed all four such row vectors xi of block 38, and have caused the resulting processed and interleaved data to be written into memory 4, memory 4 contains partially transformed 16×16 word block 40.

Analyzer 8 then sequentially processes only every fourth word of every fourth column of block 40, to generate block 42. Specifically, for each four-word column vector yi written from the first (or ninth) column of block 40, filter 16 outputs a two-word vector φ3 and filter 18 of analyzer 8 outputs a two-word vector ψ3v. For each four-word column vector yi written from the seventh (or fifteenth) column of block 40, filter 16 outputs a two-word vector φ3h and filter 18 outputs a two-word vector ψ3d. Vectors φ3 represent relatively low spatial frequency data, and vectors ψ3v, ψ3h, and ψ3d represent relatively high spatial frequency data.

Interleaver 20 interleaves the words of each vector φ3 with the words of the corresponding vector ψ3v to generate an interleaved column vector, and interleaver 20 interleaves the words of each vector ψ3h with the words of the corresponding vector ψ3d to generate an interleaved column vector. After all the interleaved column vectors have been sequentially written into different columns of memory 2, the content of memory 2 is third transformed block 42.

Block 42 consists of four transformed data blocks, each consisting of 8×8 words, which together define a pyramidal representation of original image data 30. The transformed data blocks are stored in memory 2 as a repeating pattern of four 8×8 word blocks, where each 8×8 block consists of one word (from a vector φ3) representing lowest spatial frequency data, and sixty-three words representing data having higher spatial frequency.

FIG. 4 represents a reordered version of block 42A (or 42B, 42C, or 42D) of FIG. 3. In FIG. 4, the data word having index pair "a1" (the word in the upper left corner) represents the lowest spatial frequency features of the image, the words having index pairs b2, c3, c4, d3, d4, e5-e8, f5-f8, g5-g8, and h5-h8 represent higher spatial frequency image features along a diagonal axis, the words having index pairs a2-a4, b3, b4, a5-a8, b5-b8, c5-c8, and d5-d8 represent higher spatial frequency image features along a horizontal axis, and the words having index pairs b1, c1, c2, d1, d2, e1-e4, f1-f4, g1-g4, and h1-h4 represent higher spatial frequency image features along a vertical axis. Of the "diagonal" data words, the word b2 represents image data having lowest spatial frequency, the words c3, c4, d3, and d4 represent data having higher spatial frequency, and the words having index pairs e5-e8, f5-f8, g5-g8, and h5-h8 represent data having highest spatial frequency. Of the "vertical" data words, the word b1 represents image data having lowest spatial frequency, the words c1, c2, d1, and d2 represent image data having higher spatial frequency, and the words having index pairs e1-e4, f1-f4, g1-g4, and h1-h4 represent data having highest spatial frequency.

It is apparent from a comparison of FIG. 4 with block 42A of FIG. 3, that the FIG. 4 data order is much more uniform (in terms of spatial frequency) than the data order of block 42A. This can be understood by realizing that, if the FIG. 4 data are read (or written) in line-scan order, strings of words representing identical spatial frequency (e.g., four-word string a5 through a8, and eight-word string d5 through e4) will be read (or written) consecutively. In contrast, if block 42A is read (or written) in line-scan order, no more than two words of identical spatial frequency (e.g., the eighth word in the first line followed by the first word in the second line) will be read (or written) consecutively. Because the data of block 42A are much ordered less uniformly in terms of spatial frequency than are the FIG. 4 data, the data of block 42A are in better form for subsequent coding by conventional hardware in accordance with the conventional JPEG or MPEG compression algorithm. Indeed, block 42 is in optimal form for subsequent coding by such conventional JPEG or MPEG compression circuitry.

FIG. 5 is a block diagram of a preferred embodiment of the invention. Format conversion circuit 50 of the system shown in FIG. 5 receives an image signal (which can be color video signal from a camera), and transforms the image signal into a format suitable for processing in image transformation circuit 52. For example, circuit 50 can be conventional circuitry for receiving a color video signal in color raster scan format, and reformatting it into a stream of N×M image data blocks, each block consisting of digital words in line-scan format.

Image transformation circuit 52 performs the above-described iterative transformation on each N×M image data block received from circuit 50. Thus, for each image data block, circuit 52 outputs a set of interleaved component image blocks, which together define a pyramidal representation of the image data block. In preferred embodiments, M=N=16, and circuit 50 performs a three-iteration transformation on each 16-16 image data block to convert the block into a transformed block having the format of block 42 of FIG. 3. Typically in such embodiments, each 16×16 block received by circuit 52 represents a very small portion of an image, and thus the difference between each pair of corresponding lowest frequency (DC) components of consecutive transformed blocks output from circuit 52 (i.e., the difference between corresponding words from corresponding vectors φ3 in consecutive transformed blocks) is typically small. For example, the difference between the upper left word of block 42B in FIG. 3, and the upper left word of the next block 42B output from circuit 52 (block 42B in the next block 42 output from circuit 52) is typically very small.

All words of the transformed data output from circuit 52 are quantized in quantizer circuit 54, except (typically) the words representing lowest frequency (DC) components. The quantization process reduces the magnitude (or number of bits) of each quantized word. The quantized words output from circuit 54 include more zero-value words than the corresponding unquantized words input to circuit 54. Typically, a single quantizer is applied to all the words within a particular sub-image (i.e., to all words representing one of the above-described vectors, such as vector ψ3d, ψ2h or ψ2v), although unique quantizers can be applied to all words to be quantized (e.g., to all sixty-four words, except the word representing vector φ3, of a block 42B received from circuit 52).

Each block of quantized data that is output from circuit 54 is received by image coding circuit 56, and undergoes coding therein, for example to generate compressed image data representing a compressed version of the corresponding image block previously output from circuit 50. Coding circuit 56 can consist of conventional hardware for performing the conventional JPEG or MPEG compression algorithm on each 8×8 word block of quantized data output from circuit 54. Where circuit 56 is such a conventional JPEG or MPEG compression circuit, the embodiment of the FIG. 7 circuit described below with reference to FIG. 3 is preferably employed to implement circuit 52, although the FIG. 2 apparatus can alternatively be employed for this purpose. An implementation of the FIG. 7 (or FIG. 2) apparatus which generates interleaved component blocks having the form of block 42 in FIG. 3 is desirable for use as image transformation circuit 52 with JPEG or MPEG compression hardware because it generates 8×8 blocks of transformed data (e.g., block 42B of FIG. 3), and because such 8×8 blocks of transformed data are in a form optimal for processing by conventional JPEG or MPEG compression circuitry since the words of each such 8×8 block are ordered nonuniformly in terms of spatial frequency.

FIG. 6 is a diagram of a sequence in which the words (coefficients) of each 8×8 block 42A, 42B, 42C, and 42D of FIG. 3 are typically reordered by coding circuit 56, to implement coding efficiently. As indicated in FIG. 6, coding circuit 56 typically codes the DC coefficient (labeled φ3) first, then the ψ3h coefficient, then the ψ3v coefficient, then the ψ3d coefficient, then the four ψ2h coefficients, then the four ψ2v coefficients, then the four ψ2d coefficients, then the sixteen ψ1h coefficients, then the sixteen ψ1v coefficients, and finally the sixteen ψ1d coefficients.

In typical cases where (as explained above) the difference between each pair of corresponding lowest spatial frequency (DC) components of consecutive transformed blocks output from circuit 52 is small, coding circuit 56 will exploit this property by coding the DC components using a conventional differential pulse code modulation (DPCM) technique. Such a DPCM technique codes the difference between the DC coefficient of the current block and the corresponding DC coefficient of the previous block.

Also due to the typically small size of each block received by circuit 52 (in relation to the size of a full image represented by the input image signal), the higher spatial frequency (non-DC) components of the transformed blocks output from circuit 52 are typically highly redundant and (after quantization) usually contain runs of consecutive zeros. Coding circuit 56 can exploit these properties by coding the non-DC components using a conventional run-length technique in which the upper four bits of the code symbol indicate the number of consecutive zeros before the next non-zero word, and the lower four bits of the code symbol indicate the number of significant bits in the next word.

The block codes from the DPCM and run-length models can be further compressed using entropy encoding, in preferred embodiments of coding circuit 56. Such entropy encoding is conventionally implemented using a Huffman coding circuit. To compress data symbols, a Huffman coder creates shorter codes for frequently occurring symbols and longer codes for infrequently occurring symbols.

The coded data stream output from circuit 56 (typically a stream of compressed data) is transmitted through (or stored within) transmission or storage means 58, and then received (or read from storage) by image decoder circuit 60.

Image decoder circuit 60 performs the inverse operations to those performed by circuit 56, to decode the received data. Inverse quantizer circuit 62 performs the inverse operations to those performed by circuit 52, to dequantize the decoded data output from circuit 60.

Inverse image transformation circuit 64 performs (recursively) the inverse operations performed by circuit 52. In a preferred embodiment, circuit 64 has the same structure does circuit 52 except that each filter of circuit 64's analyzers (e.g., filters 10, 12, 16, and 18) generates an "inverse" set of coefficients to the coefficients generated by the corresponding filter of circuit 52. Each N×M image data block output from circuit 64 is a reconstructed version of a corresponding N×M image data block received by circuit 52.

In preferred embodiments, such as that shown in FIG. 7, the apparatus of the invention includes only one analyzer connected between first and second memories (i.e., analyzer 86 of FIG. 7) and a control means. The control means controls the analyzer to cause it to perform not only the operations performed by the first analyzer (in those embodiments of the invention which employ two analyzers) but also those performed by the second analyzer (in embodiments with two analyzers). In the FIG. 7 apparatus, the control means includes transform address generators 87 and 88, each of which provides appropriate addresses for writing vectors of data from one of memories 82, 84, 92, and 94 to analyzer 86 (or for reading interleaved vectors from analyzer 86 to one of memories 82, 84, 92, and 94), and transform sequence controller 89 for controlling transform address generators 87 and 88.

The FIG. 7 embodiment is designed to increase the rate at which the inventive apparatus can process data (relative to the FIG. 2 embodiment), and to permit processing of video images in real time at a rate of 60 fields per second. The FIG. 7 apparatus includes identical video field memories 82, 84, 92, and 94, quad port bus switches 90 and 91, and analyzer circuit 86 connected between switches 90 and 91. Each of memories 82, 84, 92, and 94 has capacity to store an N×M field of video data (a data block having N rows and M columns). Typically, M is equal to 640 and N is equal to 480.

Switch 90 has two positions: one connecting block address generator 96 and data bus 99 to memory 92 and analyzer 86 and address generator 87 to memory 82; and the other connecting block address generator 96 and data bus 99 to memory 82 and analyzer 86 and address generator 87 to memory 92. Switch 91 also has two positions: one connecting raster address generator 98 and data bus 100 to memory 94 and analyzer 86 and address generator 88 to memory 84; and the other connecting raster address generator 98 and data bus 100 to memory 84 and analyzer 86 and address generator 88 to memory 94.

Switches 90 and 91 are controlled so that, at any time, three operations are simultaneously performed: first, analyzer 86 reads and writes data between a first one of the memories (memory 82 or 92) and a second one of the memories (memory 84 or 94) to implement the inventive transformation on an N×M field of data stored in the first or second memory; second, a previously transformed N×M field of data is read out of a third one of the memories; and third, a new N×M field of data is written into the fourth one of the memories.

Analyzer 86 preferably includes a pair of transform filters (such as filters 10 and 12 of FIG. 2) and an interleaving circuit (such as interleaving circuit 14 of FIG. 2). Under control of transform sequence controller 89, analyzer 86 can receive a pixel stream (i.e., can sequentially receive the words of a data vector) at either of its ports (A and B) from either one of switches 90 and 91. At times when analyzer 86 receives a pixel stream from switch 90, each of its filters processes the stream and supplies a transformed vector to the interleaving circuit, and the interleaving circuit outputs an interleaved data stream to switch 91. At times when analyzer 86 receives a pixel stream from switch 91, each of its filters processes the stream and supplies a transformed vector to the interleaving circuit, and the interleaving circuit outputs an interleaved data stream to switch 90.

Transform address generator 87, under control of transform sequence controller 89, provides appropriate addresses through switch 90 to memory 82 (or 92) for writing vectors of data from that memory to analyzer 86, or for reading interleaved vectors from analyzer 86 to that memory. Transform address generator 88, under control of transform sequence controller 89, provides appropriate addresses through switch 91 to memory 84 (or 94) for writing vectors of data from that memory to analyzer 86, or for reading interleaved vectors from analyzer 86 to that memory. Transform sequence controller 89 not only controls analyzer 86 (in the manner described above) but also transform address generators 87 and 88.

During a data compression mode, the FIG. 7 apparatus receives a stream of uncompressed image data on bus 100, and outputs a stream of transformed uncompressed data on bus 99 for subsequent compression (typically, by quantization and coding). During a data expansion mode, the FIG. 7 apparatus receives a stream of compressed image data on bus 99, and outputs a stream of inversely-transformed compressed data on bus 100 for subsequent expansion.

Raster address generator 98 provides the addresses required to write a field of uncompressed video data (received, in line-scan format, from data bus 100) to memory 84 or 94 (during a data compression mode), and to read a field of video data (in line-scan format) from memory 84 or 94 (during a data expansion mode).

Block address generator 96 provides the addresses required to write interleaved blocks of compressed video data (received on bus 99) to memory 82 or 92 (during a data expansion mode), and to read a field of uncompressed video data (in line-scan format) from memory 84 or 94 (during a data compression mode).

Next, we describe (with reference to FIG. 8) the sequence in which the FIG. 7 apparatus processes data in a data compression mode. For specificity, FIG. 8 assumes that three iterations of the inventive recursive transformation process are performed by analyzer 86 on each field of data to be transformed (one iteration comprises three passes through an analyzer and the other two each comprise two passes through the analyzer). In alternative embodiments, more than three or less than three iterations can be performed on the data comprising each field.

During the period labeled "field 1" in FIG. 8, a first N×M field of video data is written from line 100 through switch 91 into memory 84 in line-scan format (under control of raster address generator 98), and simultaneously, a second (previously transformed) N×M field of data in memory 82 is read (typically on an 8×8 block by 8×8 block basis, when N=640 and M=480) under control of block address generator 96 from memory 82 through switch 90 to bus 99. During the same period, a third field of video data stored in memory 94 undergoes three iterations of the inventive recursive transformation process. This recursive (three-iteration) process includes the following sequence of steps:

during the first iteration (which comprises three passes through analyzer 86), the N×M data field in memory 94 is read (under control of address generator 88 and controller 89), processed in analyzer 86, and the processed (transformed and interleaved) data are written from analyzer 86 through switch 90 to memory 92. Then, half of the processed N×M data field in memory 92 is read (i.e., M vertical vectors in memory 92, each comprising N/2 words, are read) under control of address generator 87 and controller 89, processed in analyzer 86, and the processed (transformed and interleaved) data are written from analyzer 86 through switch 91 back to memory 94. Then, the same N/2×M block of data just processed in analyzer 86 is read from memory 94 (i.e., N/2 horizontal vectors in memory 94, each comprising M words, are read from memory 94) under control of address generator 88 and controller 89, processed in analyzer 86, and the resulting processed (transformed and interleaved) data are written from analyzer 86 through switch 90 back to memory 92 under control of address generator 87 and controller 89;

then, during the second iteration, a subset of the N×M data block in memory 92 are read (under control of address generator 87 and controller 89) to analyzer 86 for processing, and the resulting processed (transformed and interleaved) data are written from analyzer 86 through switch 91 to a subset of the memory locations of memory 94 under control of address generator 88 and controller 89, and data from a subset of the memory locations of memory 94 are then read (under control of address generator 88 and controller 89) to analyzer 86 for processing, and the resulting processed (transformed and interleaved) data are written from analyzer 86 through switch 90 back to memory 92 under control of address generator 87 and controller 89; and

finally, during the third iteration, a smaller subset of the data in memory 92 are read (under control of address generator 87 and controller 89) to analyzer 86 for processing, and the resulting processed (transformed and interleaved) data are written from analyzer 86 through switch 91 to a smaller subset of the memory locations of memory 94 under control of address generator 88 and controller 89, and data from a smaller subset of the memory locations of memory 94 are then read (under control of address generator 88 and controller 89) to analyzer 86 for processing, and the resulting processed (transformed and interleaved) data are written from analyzer 86 through switch 90 back to memory 92 under control of address generator 87 and controller 89.

At the end of the period labeled "field 1" in FIG. 8, quad port switches 90 and 91 are reversed.

Then, during the period labeled "field 2" in FIG. 8, the next field is written from bus 100 into memory 94 while the previously transformed field of data in memory 92 is read (block by block) under control of block address generator 96 from memory 92 through switch 90 to bus 99. During the same period, the field of video data previously written into memory 84 undergoes three iterations of the inventive recursive transformation process. This process is performed in the same way as was the recursive transformation performed during the period labeled "field 1," except that data are written and read between memories 84 and 82 (rather than memories 94 and 92). At the end of the third iteration, the entire contents of memory 84 are written into memory 82. Then, at the end of the period labeled "field 2" in FIG. 8, quad port switches 90 and 91 are again reversed.

Then, during the period labeled "field 3" in FIG. 8, the same steps performed during the period labeled "field 1" are repeated.

The sequence in which the FIG. 7 apparatus processes data in a data expansion mode is identical to that described with reference to FIG. 8, except in that the roles of memories 84 and 82 are reversed and the roles of memories 94 and 92 are reversed, and in that data flow is from bus 99 to bus 100 (rather than from bus 100 to bus 99 as in the compression mode).

It should be apparent that the terms "horizontal" and "vertical" used in the specification, including in the claims, do not refer to any physical orientation of memory locations in a memory array. The term "memory array" implies memory locations organized in a coordinate system having at least two dimensions, so that each location is identified by a set of two or more coordinates. The term "horizontal" refers to a first coordinate in such set, and the term "vertical" refers to a second coordinate in such set. Thus, "horizontal" vector and "row" denote the contents of a set of memory locations having different first coordinates and a common second coordinate, and "vertical" vector and "column" denote the contents of a set of memory locations having different second coordinates and a common first coordinate.

Various modifications in the structure and method of operation of the described embodiments are within the scope and spirit of this invention, and will be apparent to those skilled in the art. Although the invention has been described in connection with specific preferred embodiments, the invention as claimed should not be unduly limited to such specific embodiments.

Carnahan, Shawn V. A.

Patent Priority Assignee Title
10356385, Jun 07 2016 Stock Company Research and Development Center “Electronic Information Computation Systems” Method and device for stereo images processing
10945706, May 05 2017 BIIM ULTRASOUND AS Hand held ultrasound probe
11113796, Feb 09 2018 Delta Electronics, Inc. Image enhancement circuit and method thereof
11744551, May 05 2017 BIIM ULTRASOUND AS Hand held ultrasound probe
5497777, Sep 23 1994 General Electric Company Speckle noise filtering in ultrasound imaging
5668850, May 23 1996 General Electric Company Systems and methods of determining x-ray tube life
5774677, Jun 28 1996 Intel Corporation Dual interleaved block motion compensation
5900953, Jun 17 1997 LIZARDTECH, INC Method and apparatus for extracting a foreground image and a background image from a color document image
5909518, Nov 27 1996 Qualcomm Incorporated System and method for performing wavelet-like and inverse wavelet-like transformations of digital data
6144773, Feb 27 1996 Vulcan Patents LLC Wavelet-based data compression
6161105, Nov 21 1994 ORACLE INTERNATIONAL CORPORATION OIC Method and apparatus for multidimensional database using binary hyperspatial code
6181831, Sep 28 1995 Sony Corporation; Sony United Kingdom Limited Spatial frequency-domain video signal processing
6201897, Nov 09 1998 INTERGRAPH GOVERNMENT SOLUTIONS CORPORATION; Intergraph Corporation Transformation and selective inverse transformation of large digital images
6226414, Apr 20 1994 INPHI CORPORATION Image encoding and decoding method and apparatus using edge synthesis and inverse wavelet transform
6330367, Apr 20 1994 Oki Electric Industry Co., Ltd. Image encoding and decoding using separate hierarchical encoding and decoding of low frequency images and high frequency edge images
6381280, May 30 1997 Vulcan Patents LLC Single chip motion wavelet zero tree codec for image and video compression
6442298, Nov 09 1998 INTERGRAPH GOVERNMENT SOLUTIONS CORPORATION; Intergraph Corporation Transformation and selective inverse transformation of large digital images
6449619, Jun 23 1999 GOOGLE LLC Method and apparatus for pipelining the transformation of information between heterogeneous sets of data sources
6477279, Apr 20 1994 INPHI CORPORATION Image encoding and decoding method and apparatus using edge synthesis and inverse wavelet transform
6680975, Feb 29 1992 Samsung Electronics Co., Ltd. Signal encoding and decoding system and method
6697525, Oct 02 1998 PARTHUSCEVA LTD System method and apparatus for performing a transform on a digital image
6711299, Mar 11 1997 UPLOAD TECHNOLOGIES, S A Wavelet transformation of dithered quantized/reduced color pixels for color bit depth image compression and decompression
6847468, Dec 05 1994 Microsoft Technology Licensing, LLC Progressive image transmission using discrete wavelet transforms
6859563, Mar 30 2001 RICOH CO , LTD Method and apparatus for decoding information using late contexts
6868186, Jul 13 2000 CEVA D S P LTD Visual lossless image compression
6873734, Sep 21 1994 Ricoh Corporation Method and apparatus for compression using reversible wavelet transforms and an embedded codestream
6895120, Mar 30 2001 RICOH CO , LTD 5,3 wavelet filter having three high pair and low pair filter elements with two pairs of cascaded delays
6898323, Feb 15 2001 RICOH CO LTD Memory usage scheme for performing wavelet processing
6898325, Feb 15 2001 Ricoh Company, Ltd. Method and apparatus for clipping coefficient values after application of each wavelet transform
6904175, Mar 11 1997 UPLOAD TECHNOLOGIES, S A Image compression using an interger reversible wavelet transform with a property of precision preservation
6904178, Feb 15 2001 Ricoh Co., Ltd. Method and apparatus for eliminating flicker by quantizing values based on previous quantization
6925209, Feb 15 2001 Ricoh Co., Ltd. Method and apparatus for outputting a codestream as multiple tile-part outputs with packets from tiles being output in each tile-part
6937659, Nov 14 1997 GENERAL DYNAMICS INFORMATION TECHNOLOGY, INC Apparatus and method for compressing video information
6937767, Nov 29 1997 ALGOTEC SYSTEMS LTD Diagnosis oriented lossy compression of medical images
6950558, Mar 30 2001 RICOH CO , LTD , A CORPORATION OF JAPAN Method and apparatus for block sequential processing
6973217, Feb 15 2001 Ricoh Co., Ltd. Method and apparatus for sending additional sideband information in a codestream
6983075, Feb 15 2001 Ricoh Co., Ltd Method and apparatus for performing selective quantization by manipulation of refinement bits
6990247, Sep 20 1994 Ricoh Corporation Multiple coder technique
7003168, Mar 11 1997 UPLOAD TECHNOLOGIES, S A Image compression and decompression based on an integer wavelet transform using a lifting scheme and a correction method
7006697, Mar 30 2001 RICOH CO , LTD Parallel block MQ arithmetic image compression of wavelet transform coefficients
7016545, Sep 21 1994 RICOH CORPORATION, LTD A CORPORATION OF JAPAN Reversible embedded wavelet system implementation
7050642, May 19 1994 Apple Inc Method and apparatus for video compression using microwavelets
7054493, Sep 21 1994 Ricoh Co., Ltd. Context generation
7062101, Mar 30 2001 RICOH CO , LTD Method and apparatus for storing bitplanes of coefficients in a reduced size memory
7062103, Feb 15 2001 Ricoh Co., Ltd. Method and apparatus for specifying quantization based upon the human visual system
7068849, Sep 21 1994 Ricoh Corporation Method and apparatus for compression using reversible wavelet transforms and an embedded codestream
7072520, Feb 15 2001 Ricoh Co., Ltd. Method and apparatus for selecting layers for quantization based on sideband information
7076104, Sep 21 1994 Ricoh Corporation Compression and decompression with wavelet style and binary style including quantization by device-dependent parser
7079690, Feb 15 2001 Ricoh Co., Ltd. Method and apparatus for editing an image while maintaining codestream size
7088869, Mar 30 2001 Ricoh Co., Ltd. 5,3 wavelet filter having three high pair and low pair filter elements with two pairs of cascaded delays
7092118, Dec 05 1994 Microsoft Technology Licensing, LLC Progressive image transmission using discrete wavelet transforms
7095900, Feb 15 2001 Ricoh Co., Ltd. Method and apparatus for performing scalar quantization with a power of two step size
7095907, Jan 10 2002 RICOH CO LTD A CORPRORATION OF JAPAN Content and display device dependent creation of smaller representation of images
7120305, Apr 16 2002 RICOH CO , LTD Adaptive nonlinear image enlargement using wavelet transform coefficients
7139434, Sep 21 1994 Ricoh Co., Ltd. Decoding with storage of less bits for less important data
7164804, Feb 15 2001 Ricoh Co., Ltd. Method and apparatus for eliminating flicker by quantizing values based on previous quantization
7167589, Sep 21 1994 Ricoh Co., Ltd. Disk read technique
7167592, Sep 21 1994 Ricoh Co., Ltd. Method and apparatus for compression using reversible wavelet transforms and an embedded codestream
7190840, Jan 07 2002 HEWLETT-PACKARD DEVELOPMENT COMPANY L P Transform coefficient compression using multiple scans
7215820, Sep 21 1994 Ricoh Co., Ltd. Method and apparatus for compression using reversible wavelet transforms and an embedded codestream
7227999, Sep 21 1994 Ricoh Co., Ltd. Printing system application using J2K
7280252, Dec 19 2001 RICOH CO , LTD Error diffusion of multiresolutional representations
7289677, Sep 21 1994 Ricoh Co., Ltd. Reversible embedded wavelet system implementation
7292657, Feb 29 1992 Samsung Electronics Co., Ltd. Signal compressing signal
7298912, Mar 30 2001 Ricoh Co., Ltd. Method and apparatus for assigning codeblocks to coders operating in parallel
7321695, Sep 21 1994 Ricoh Co., Ltd. Encoder rate control
7397963, Mar 30 2001 Ricoh Co., Ltd. Method and apparatus for storing bitplanes of coefficients in a reduced size memory
7418142, Sep 20 1994 Ricoh Company, Ltd. Method for compression using reversible embedded wavelets
7457473, Mar 30 2001 Ricoh Co., Ltd. Method for block sequential processing
7474791, Jan 10 2002 Ricoh Co., Ltd. Content and display device dependent creation of smaller representations of images
7477792, Feb 15 2001 Ricoh Co., Ltd. Method and apparatus for performing progressive order conversion
7581027, Jun 27 2001 RICOH CO , LTD JPEG 2000 for efficent imaging in a client/server environment
7596590, Jun 30 1999 Canon Kabushiki Kaisha Image communication apparatus and method
7609760, Feb 29 1992 Samsung Electronics Co., Ltd. Signal compressing system
7616687, Feb 29 1992 Samsung Electronics Co., Ltd. Signal compressing system
7634145, Sep 21 1994 Ricoh Corporation Compression and decompression with wavelet style and binary style including quantization by device-dependent parser
7639886, Oct 04 2004 Adobe Inc Determining scalar quantizers for a signal based on a target distortion
7653255, Jun 02 2004 Adobe Inc Image region of interest encoding
7684490, Mar 01 1993 Samsung Electronics Co., Ltd. Signal compressing system
7724821, Feb 29 1992 Samsung Electronics Co., Ltd. Signal compressing system
7724822, Feb 29 1992 Samsung Electronics Co., Ltd. Signal compressing system
7724823, Feb 29 1992 Samsung Electronics Co., Ltd. Signal compressing system
7724824, Feb 29 1992 Samsung Electronics Co., Ltd. Signal compressing system
7724828, Feb 28 1992 Samsung Electronics Co., Ltd. Signal compressing system
7724829, Feb 28 1992 Samsung Electronics Co., Ltd. Signal compressing system
7742522, Feb 29 1992 Samsung Electronics Co., Ltd. Signal compressing system
7742527, Feb 29 1992 Samsung Electronics Co., Ltd. Signal compressing system
7764735, Feb 29 1992 Samsung Electronics Co., Ltd. Signal compressing system
7782956, Feb 28 1992 Samsung Electronics Co., Ltd. Signal compressing system
7787538, Feb 29 1992 Samsung Electronics Co., Ltd. Signal compressing system
7949045, Feb 28 1992 Samsung Electronics Co., Ltd. Signal compressing system
7949046, Feb 28 1992 Samsung Electronics Co., Ltd. Signal compressing system
7953148, Feb 28 1992 Samsung Electronics Co., Ltd. Signal compressing system
7953149, Feb 28 1992 Samsung Electronics Co., Ltd. Signal compressing system
7953150, Feb 28 1992 Samsung Electronics Co., Ltd. Signal compressing system
8036472, Oct 08 2002 NTT DOCOMO, INC. Image encoding method, image decoding method, image encoding apparatus, image decoding apparatus, image encoding program, and image decoding program
8116377, Nov 20 1998 Interval Licensing LLC Low cost video compression using fast, modified Z-coding of wavelet pyramids
8326057, Oct 08 2002 NTT DOCOMO, INC. Image encoding method, image decoding method, image encoding apparatus, image decoding apparatus, image encoding program, and image decoding program
8422808, Oct 08 2002 NTT DOCOMO, INC. Image encoding method, image decoding method, image encoding apparatus, image decoding apparatus, image encoding program, and image decoding program
8422809, Oct 08 2002 NTT DOCOMO, INC. Image encoding method, image decoding method, image encoding apparatus, image decoding apparatus, image encoding program, and image decoding program
8565298, Sep 21 1994 Ricoh Co., Ltd. Encoder rate control
9894379, Jul 10 2001 DIRECTV, LLC System and methodology for video compression
Patent Priority Assignee Title
4974187, Aug 02 1989 Aware, Inc. Modular digital signal processing system
4984076, Jul 27 1988 Kabushiki Kaisha Toshiba Image compression coding system
5014134, Sep 11 1989 AWARE, INC , A MA CORP Image compression method and apparatus
5214507, Nov 08 1991 AT&T Bell Laboratories; American Telephone and Telegraph Company Video signal quantization for an MPEG like coding environment
5214708, Dec 16 1991 Speech information extractor
5241383, May 13 1992 SHINGO LIMITED LIABILITY COMPANY Pseudo-constant bit rate video coding with quantization parameter adjustment
5245589, Mar 20 1992 Method and apparatus for processing signals to extract narrow bandwidth features
5248845, Mar 20 1992 CREATIVE TECHNOLOGY LTD Digital sampling instrument
5253530, Aug 12 1991 Method and apparatus for reflective ultrasonic imaging
5262958, Apr 05 1991 Texas Instruments Incorporated Spline-wavelet signal analyzers and methods for processing signals
WO9102318,
///////
Executed onAssignorAssigneeConveyanceFrameReelDoc
Jan 27 1993ImMIX(assignment on the face of the patent)
Mar 23 1993CARNAHAN, SHAWN V A IMMIX, A DIVISION OF CARLTON INTERNATIONAL CORP ASSIGNMENT OF ASSIGNORS INTEREST 0065300401 pdf
Sep 30 1994CARLTON INTERNATIONAL CORPORATIONSCITEX IM ACQUISITION CORP ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0075770541 pdf
Nov 03 1994SCITEX IM ACQUISITION CORP IMMIX, INC ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0075770549 pdf
Sep 27 1995IMMIX, INC SCITEX DIGITAL VIDEO, INC CHANGE OF NAME SEE DOCUMENT FOR DETAILS 0077150475 pdf
Dec 10 1998SCITEX DIGITAL VIDEO, INC ACCOM, INC ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0105470874 pdf
Jan 21 2000LA SALLE BUSINESS CREDIT, INC ACCOM, INC SECURITY AGREEMENT0119670389 pdf
Date Maintenance Fee Events
Dec 01 1998REM: Maintenance Fee Reminder Mailed.
May 09 1999EXP: Patent Expired for Failure to Pay Maintenance Fees.


Date Maintenance Schedule
May 09 19984 years fee payment window open
Nov 09 19986 months grace period start (w surcharge)
May 09 1999patent expiry (for year 4)
May 09 20012 years to revive unintentionally abandoned end. (for year 4)
May 09 20028 years fee payment window open
Nov 09 20026 months grace period start (w surcharge)
May 09 2003patent expiry (for year 8)
May 09 20052 years to revive unintentionally abandoned end. (for year 8)
May 09 200612 years fee payment window open
Nov 09 20066 months grace period start (w surcharge)
May 09 2007patent expiry (for year 12)
May 09 20092 years to revive unintentionally abandoned end. (for year 12)