systems and methods for converting a picture frame rate from a source video at a first rate to a target video at a second rate via interpolation of an intermediate frame. In one implementation, the system includes a phase plane correlation calculator including a low pass filter and a high pass filter for receiving previous frame data and current frame data where the phase plane correlation calculator is configured to generate a first motion vector based upon low pass representations and high pass representations. The system may also include a motion compensated interpolator that receives the first motion vector and an additional input motion vector and determines a final motion vector for use in interpolation. The system may further include an intermediate frame generator configured to generate the intermediate frame utilizing the final motion vector.
|
1. A method for interpolating an intermediate frame block from (i) a previous frame block and (ii) a current frame block, the method comprising:
processing the previous frame block by a first process;
processing the previous frame block also by a second process different than the first process;
calculating a first candidate motion vector as a function of (i) the current frame block, and (ii) the previous frame block processed by the first process;
calculating a second candidate motion vector from (i) the current frame block, and (ii) the previous frame block processed by the second process;
calculating a quality value for each of the first candidate motion vector and the second candidate motion vector, wherein the quality value indicates accuracy of the respective candidate motion vector in depicting object motion from the previous frame to the current frame;
selecting, based on the quality value corresponding to each of the first candidate motion vector and the second candidate motion vector, a final motion vector from among the first candidate motion vector and the second candidate motion vector; and
generating the intermediate frame block based on the final motion vector.
11. A system for interpolating an intermediate frame block from (i) a previous frame block and (ii) a current frame block, the system comprising:
a first processing unit configured to process the previous frame block by a first process;
a first processing unit configured to process the previous frame block by a second process different than the first process;
a quantization unit configured to
calculate a first candidate motion vector as a function of (i) the current frame block, and (ii) the previous block processed by the first process;
calculating a second candidate motion vector from the current frame block, and (ii) the previous frame block processed by the second process;
a validate engine configured to
calculate a quality value for each of the first candidate motion vector and the second candidate motion vector, wherein the quality value indicates accuracy of the respective candidate motion vector in depicting object motion from the previous frame block to the current frame block; and
select, based on the quality value for each of the first candidate motion vector and the second candidate motion vector, a final motion vector from among the first candidate motion vector and the second candidate motion vector; and
an intermediate frame generator configured to generate the intermediate frame block based on the final motion vector.
2. The method of
3. The method of
applying the respective candidate motion vector to the previous frame block to yield a shifted frame block; and
calculating the quality value as a difference here en the shifted frame block and the current frame block.
4. The method of
5. The method of
6. The method of
7. The method of
processing the current frame block by the first process; and
processing the current frame block by the second process,
wherein the first candidate motion vector is calculated as a function of (i) the current frame block processed by the first process, and (ii) the previous frame block processed by the first process, and
wherein the second candidate motion vector is calculated as a function of (i) the current frame block processed by the second process, and (ii) the previous frame block processed by the second process.
8. The method of
generating a Fourier transform of the current frame block;
generating a Fourier transform of the previous frame block processed by the respective process;
calculating a phase difference between (i) the Fourier transform of the current frame block and (i) the Fourier transform of the previous frame block processed by the respective process; and
performing an inverse Fourier transform on the phase difference to yield a contour map having a peak, wherein a location of the peak on the contour map indicates a movement of an object between the previous frame and current frame, and wherein a height of the peak indicates a size of the object; and
determining the respective candidate motion vector based on the peak in the contour map.
9. The method of
wherein the second candidate motion vector is an initial second candidate motion vector;
wherein the method further includes:
calculating additional first candidate motion vectors as a function of (i) the current frame block and (ii) the previous frame block processed by the first process;
calculating additional second candidate motion vectors from (i) the current frame block and (ii) the previous frame block processed by the second process; and
calculating, for each of the additional first candidate motion vectors and each of the additional second candidate motion vectors, a quality value that indicates accuracy of the respective candidate motion vector in depicting object motion from the previous frame block to the current frame block; and
wherein selecting the final motion vector includes selecting, based on the quality values, the final motion vector from among, the initial and additional first candidate motion vectors and the initial and additional second candidate motion vectors.
10. The method of
12. The system of
13. The system of
applying the respective candidate motion vector to the previous frame block to yield a shifted frame block; and
calculating the quality value as a difference between the shifted frame block and the current frame block.
14. The system of
15. The system of
16. The system of
17. The system of
wherein the first processing unit is further configured to process the current frame block by the first process;
wherein the second processing unit is further configured to process the current frame block by the second process; and
wherein the quantization unit is further configured to
calculate the first candidate motion vector as a function of (i) the current frame block processed by the first process, and (ii) the previous frame block processed by the first process; and
calculate the second candidate motion vector as a function of (i) the current frame block processed by the second process, and (ii) the previous frame block processed by the second process.
18. The system of
generating a Fourier transform of the current frame block;
generating a Fourier transform of the previous frame block processed by the respective process;
calculating a phase difference between (i) the Fourier transform of the current frame block and (i) the Fourier transform of the previous frame block processed by the respective process; and
performing an inverse Fourier transform on the phase difference to yield a contour map having a peak, wherein a location of the peak on the contour map indicates a movement of an object between the previous frame block and current frame block, and wherein a height of each peak indicates a size of the object; and
determining the respective candidate motion vector based on the peaks in the contour map.
19. The system of
wherein the second candidate motion vector is an initial second candidate motion vector;
wherein the quantization unit is further configured to:
calculate additional first candidate motion vectors as a function of (i) the current frame block and (ii) the previous frame block processed by the first process;
calculate additional second candidate motion vectors from (i) the current frame block and (ii) the previous flame block processed by the second process; and
wherein the validate engine is configured to
calculate, for each of the additional the first candidate motion vectors and each of the additional second candidate motion vectors, a quality value that indicates accuracy of the respective candidate motion vector in depicting object motion from the previous frame block to the current frame block; and
select, based on the quality values, the final motion vector from among the initial and additional first candidate motion vectors and the initial and additional second candidate motion vectors.
20. The system of
|
This application is a continuation of U.S. patent application Ser. No. 12/400,227, filed on Mar. 9, 2009, which claims the benefit of U.S. Provisional Application No. 61/051,782, filed May 9, 2008; U.S. Provisional Application No. 61/046,922, filed Apr. 22, 2008; and U.S. Provisional Application No. 61/059,524, filed Jun. 6, 2008, all of which are herein incorporated in their entirety by reference. This application is related to U.S. Non-provisional application Ser. No. 12/400,207, filed Mar. 9, 2009, entitled “Filter Bank Based Phase Correlation Architecture for Motion Estimation” and U.S. Non-provisional application Ser. No. 12/400,220, filed Mar. 9, 2009, entitled “Picture Rate Conversion System for High Definition Video.”
The technology described in this patent document relates generally to the field of picture rate conversion, and more particularly to motion compensated picture rate conversion.
Typical movie films are recorded at 24 Hz, 25 Hz or 30 Hz. Picture rates of common video cameras are 50 Hz and 60 Hz. Commercially available television displays, on the other hand, have picture rates up to and beyond 120 Hz, and employ either progressive or interlaced scanning. Hence, to interface broadcast video with a high-end TV display, the original sequence from the broadcast video needs to be up-converted using, for example, a picture rate converter. A picture rate converter typically operates by interpolating image frames at time instances where the frame sequence from a lower-frequency source device has yet to be registered in a higher-frequency destination display.
In simple picture rate converters, a picture is often repeated in the destination display until the next picture arrives from the source device, which oftentimes results in blur and judder when motion occurs. Motion estimation and compensation circuits may be used in a picture rate converter to reduce these unwanted effects and achieve a high performance conversion for moving sequences. Motion compensation operates by estimating where elements of an interpolated picture would be, based on the direction and speed of the movement of those elements. The direction and speed values may then be expressed as motion vectors and are used to “move” the elements to the correct position in the newly interpolated frame. If this technique is applied correctly, its impact may be immediately visible on any picture sequence involving motion, where the resulting pictures can hardly be distinguished from the original sequences before the up-conversion.
It is thus desirable to provide methods and systems that minimize computational cost associated with motion-compensated picture rate conversion while maximizing its estimation accuracy.
In accordance with the teachings provided herein, systems and methods are provided for converting a picture frame rate from a source video at a first rate to a target video at a second rate via interpolation of an intermediate frame. As an example, a picture frame rate converting system may include a phase plane correlation calculator including a low pass filter and a high pass filter for receiving previous frame data and current frame data where the phase plane correlation calculator is configured to generate a first motion vector based upon low pass representations and high pass representations. The system may also include a motion compensated interpolator that receives the first motion vector and an additional input motion vector and determines a final motion vector for use in interpolation. The system may further include an intermediate frame generator configured to generate the intermediate frame utilizing the final motion vector.
The system may operate on a two channel low-voltage differential signal source video and output a four channel low-voltage differential signal target video. The phase plane correlation calculator may include an Xtensa™ processing unit. Further, the system may include an RGB to luminance converter that receives the previous frame data and the current frame data before the phase plane correlation calculator and converts the input data to a luminance signal.
The system may also include an error concealer configured to receive the final motion vector and determines if error correction is required. The error concealer may be further configured to perform edge propagation and texture synthesis procedures to the generated intermediate frame if it is determined that error correction is required.
Further, the system may include a decision system configured to determine whether to utilize the generated intermediate frame in the target video where the decision system is configured to determine whether to utilize the generated intermediate frame based on a magnitude of error correction required, a quality of error correction performed, or a magnitude of the final motion vector. Further, the decision system may be configured to repeat the previous frame in the target video if the magnitude of error correction required is above a first threshold, the quality of error correction performed is below a second threshold, or a magnitude of input motion vectors is below a third threshold.
The motion compensated interpolator may further include a shift register configured to receive the previous frame data and the first motion vector to shift the previous frame data according to the first motion vector to generate a first shifted candidate. The interpolator may further include a sum of absolute difference calculator configured to receive the first shifted candidate and the current frame data to calculate a score based on a comparison between the two.
The interpolator may also include a first previous frame data buffer configured to receive the previous frame data and make the previous frame data available to the shift register and a first current frame data buffer configured to receive the current frame data and make the current frame data available to the sum of absolute difference calculator. The motion compensated interpolator may further include a second previous frame data buffer such that the first previous frame data buffer and the second previous frame data buffer form a double buffer, and the interpolator may also include a second current frame data buffer such that the first current frame data buffer and the second current frame data buffer form a double buffer. The interpolator may also include a low pass filter positioned between the first current frame data buffer and the sum of absolute difference calculator.
The motion compensated interpolator may also include a motion vector buffer configured to make a plurality of motion vectors available to the shift register for each previous frame data—current frame data set. The motion vector buffer may be a circular buffer that holds up to 16 motion vectors. The shift register may be configured to shift the previous frame data according to each of the plurality of motion vectors to generate a plurality of shifted candidates where the sum of absolute difference calculator is configured to calculate a score for each of the plurality of shifted candidates based on a comparison between the shifted candidate and the current frame data. A final motion vector may be selected from the plurality of motion vectors based on the calculated scores where the motion vector having the lowest sum of absolute difference score is selected as the final motion vector. A smoothing factor may also be considered in selecting the final motion vector.
As a further example, a method of converting a picture frame rate from a source video at a first rate to a target video at a second rate via interpolation of an intermediate frame may include a step of decomposing the previous frame data and the current frame data into low pass representations and high pass representations, respectively. The method may then generate a first motion vector based upon the low pass representations and the high pass representations of the previous frame data and the current frame data. The method may further include a step of selecting a final motion vector from candidate motion vectors based upon a quality calculation. The method may then generate the intermediate frame utilizing the final motion vector.
With reference to
The entire MCPRC system 10 illustrated in
Like the configuration of
As illustrated in
In the illustrated MCPRC systems 10 and 30, the input video signals may range from Standard Definition (NTSC/PAL/SECAM) to High Definition and may be interlaced or progressive-based. In some instances, the video signal resolution is even lower than Standard Definition with low frame rates. For example, the input video signal may be a QVGA (320×240) input at 15 or 30 frames per second from a connector device in a portable media player such as an iPod. In certain instances, the low-resolution video signal may be fed to a video dock in a personal media player or a multimedia cellular phone via a connector device, where the dock may contain an integrated circuit capable of performing spatial and temporal conversions from, for example, 320×160 at 5 fps to 720×480 at 60 fps. Interlaced inputs may be composed of video-originated or film-originated material. Video-originated material may be first de-interlaced and converted from a field rate to a frame rate before being input to MCPRC modules 16.
In
Following storage of the low pass and high pass representations of the input signal, a phase plane correlation calculator (PPC) 78 accesses the filtered representations of the input signals and performs a phase plane correlation calculation on the filtered data. In performing the phase plane correlation calculation, the PPC 78 may utilize a buffer 80 and a processing unit 82 such as an Xtensa™ CPU. The PPC 78 and processing unit 82 perform a phase plane correlation calculation on the filtered representations of the input signal to produce one or more candidate motion vectors. The candidate motion vectors may be stored in a validate motion vector (VMV) buffer 84 prior to access by a VMV engine 86.
The VMV engine 86 receives candidate motion vectors from the VMV buffer 84 as well as previous frame data and current frame data from physical memory 72 via the DDR SDRAM controller 70. The VMV engine 86 determines a final motion vector to be used in generating one or more intermediate frames by comparing the candidate motion vectors stored in the VMV buffer 84 to the received previous frame data and current frame data. For example, the VMV engine 86 may perform a quality calculation for each of the candidate motion vectors and select the candidate motion vector having a best score relating to the accuracy of the candidate motion vector in depicting object motion from the previous frame to the current frame while offering a high degree of picture smoothness. Following selection of a final motion vector based on the quality calculation, the VMV engine 86 generates one or more intermediate frames based on the selected final motion vector for insertion into the final up-converted target video.
In calculating motion vectors for integrating intermediate frames for up-conversion of video, some portions of the intermediate frame may be undefined by the final motion vectors selected by the VMV engine 86. This anomaly in motion based interpolation often occurs at edges of objects appearing in the video. This phenomenon is illustrated in
Following generation of the intermediate frames and error correction, a second formatter 90 reads the target video, which includes previous frame data, current frame data, and the generated intermediate frames, and converts them into the desired output format. In the example of
It should be noted that various modifications may be made to the example of
When block sizes becomes large, the reliability of the estimated motion vectors may decrease. Thus, it is often possible to miss small object motion because small objects do not make large contributions to the correlation surface and are masked by noise in the image. To circumvent this problem, a filter bank based design may be utilized. Filtering the input signal into low pass representations and high pass representations aides the MCPRC system in identifying both large and small object motion within a video frame. Typical motion compensation converters are unable to account for such multiple object movement within a video frame. Thus, incorrect motion vectors may be calculated where multiple objects are moving at the same time, such as in a sports video where players may appear as large objects on the screen moving in one direction while a small ball may also be depicted moving in a different direction or at a different speed. By decomposing input signals into both low pass and high pass representations, both small and large object motion may be better accounted for and more optimal motion compensation may be accomplished. The low pass filtered image captures the global motion or large object motion in the block, and the high pass filtered image captures the small object motion. Because these two motion engines are independent of each other, the problem of failing to compensate for small object motion may be addressed.
The process for generating motion vectors of
Following decomposition, each of the representations are processed by one or more two-dimensional fast Fourier transform calculators (FFTs) 128, 130. The two-dimensional FFTs 128, 130 take the time-domain representations output by the filter band based decomposition and quantization unit 122 and convert the representations into frequency-domain representations: F1(ωx, ωy), F2(ωx, ωy), F3(ωx, ωy), F4(ωx, ωy). Some or all of the frequency-domain representations may be temporarily stored in a frame buffer 136 before proceeding with further calculation. Maximum values at the beginning of a block may be retained and stored for later detection of periodic structures as shown at 132.
Following calculation of the frequency-domain conversions, a phase difference 138 is calculated between the low pass, frequency-domain representations of the previous frame data and the current frame data. For example, the phase difference may be calculated by solving for the “A” and “B” parameters of the following formula:
F2(ωx,ωy)=e−j(Aω
After calculating the phase difference 138 between the previous frame data and the current frame data, a two-dimensional inverse fast Fourier transform (IFFT) 140 is applied to the calculated phase difference 138. The result of the IFFT 140 calculation is a two-dimensional phase plane correlation surface. The phase plane correlation surface may be viewed as a contour map identifying motion between the previous frame and the current frame of the source video. The locations of peaks on the phase plane correlation surface (a1, b1) correspond to motion within the frame block such that:
F2(x2,y2)=F1(x2+n·a1,y2+m·b1).
The height of a peak on the phase correlation surface corresponds to the size of an object that is moving within a block. To locate peaks within the phase correlation surface, a peak search 142 is performed, and based on the identified peaks, a low pass filter based motion vector 144 is determined. The low pass filter based motion vector 144 corresponds to large object motion within a frame block.
A similar process is performed utilizing the high pass frequency-domain representations (F3(ωx, ωy), F4(ωx, ωy)) calculated by the two-dimensional FFT 128. Again, some or all of the high pass frequency-domain representations may be temporarily stored in a frame buffer 136. Following calculation of the high pass frequency-domain representations, a phase difference 146 is calculated between the high pass, frequency-domain representations of the previous frame data and the current frame data. For example, the phase difference 146 may be calculated by solving for the “C” and “D” parameters of the following formula:
F4(ωx,ωy)=e−j(Aω
After calculating the phase difference 146 between the previous frame data and the current frame data, a two-dimensional IFFT 148 is applied to the calculated phase difference 146. The result of the IFFT calculation 148 may be viewed as a second two-dimensional phase plane correlation surface. The locations of peaks on the second phase plane correlation surface (c1, d1) correspond to motion within the frame block such that:
F4(x4,y4)=F3(x4+n·c1,y4+m·d1).
To locate peaks within the second phase correlation surface, a peak search 150 is performed, and based on the identified peaks, a high pass filter based motion vector 152 is determined. The high pass filter based motion vector 152 corresponds to small object motion within a frame block.
One of the difficulties in performing motion compensation on video having a high pixel density is the number of computations to be performed to accomplish high quality compensation. A video having 1920×1080 pixels is made up of about 1000 64×32 pixel blocks where each pixel block includes over 2000 pixels. Thus, an input picture rate of 60 Hz results in about 120 million calculations per second of video. To accommodate this large number of calculations, care may need to be taken regarding selection of components for the phase plane correlation calculator 42.
The choice of the filter bank involves appropriate selection of the low pass and high pass filter. It may be desirable to select filters according to the input signal. For example, for a block size of 64×32 pixels a multi-tap filter of 15×7 pixels may be selected. The filter size may also be selected to be proportional to the input resolution of the video. A higher resolution video has more redundancy of data and may require a larger filter kernel. Gradient edge based detectors such as the Sobel edge detector have been discovered to perform well as high pass filters. The gradient edge based detectors reliably detect the presence of small objects in blocks. To facilitate simplified, high-speed calculations by the FFTs 128, 130, the low pass and high pass data may be quantized to two bit and one bit resolution respectively.
As stated above, the FFTs 128, 130 may be required to perform very large numbers of computations very quickly. Examples of FFTs that may be used in the phase plane correlation calculator 42 include an FFT module that computes a 64 point FFT using a radix 2 decimation-in-time (DIT) implementation. The internal resolution of the FFT and hence the input and output resolutions are 2×16 bit wide (real and complex data). The FFT module is designed in a pipelined fashion and has a latency of around 10 clock cycles. Thus, if the input 64 data points are ready in clock 1, then the output 64 FFT values will be ready after 10 clock cycles latency. This design has a very high throughput. The inverse FFT may be done using the same block as is used for FFTs using a signal that controls the sign specific multipliers in the butterfly structure of the FFT module. A 32 point FFT may also be done using this FFT module by alternate zero insertion of the input data and then taking only the first 32 points of the output. The minimum time taken to compute the horizontal FFTs (HFFT) in this design of a 64×32 block is 32+10 clock cycles and for vertical FFTs (VFFT) is 64+10 clocks. The phase difference calculators 138, 146 may be implemented by modules using a CORDIC implementation to calculate the phase and magnitude of a complex number.
Following an IFFT calculation, the phase plane correlation surface has peaks that correspond to motion in the block. The phase plane correlation calculation process may result in multiple peaks in the correlation surface. An example search process 142, 150 finds the top four values of peaks in the correlation surface. Due to the presence of sub-pixel motion it is possible to have peaks whose values are shared between two adjacent locations. Thus, the search process may calculate a unique peak value in a neighborhood of 8 positions. In other words if a peak is found in position (x,y), then nearby peaks in positions (x+/−1,y+/−1) may not be considered. The location and magnitude of the discovered peaks are then utilized to calculate low pass and high pass based motion vectors 144, 152.
The phase plane correlation calculator 160 of
Simultaneously at clock cycle 200, a block of PLP data can be read from physical memory to buffer A 164. In similar fashion as the CLP data, an HFFT of the PLP data is performed at clock cycle 300 with the results being stored in buffer B 166. A transpose operation is performed at clock cycle 350 transferring the PLP data from buffer B 166 to buffer N 168. A VFFT is performed on the data in buffer N 168 at clock cycle 400 with the results being stored in buffer O 170.
Following phase extraction for the PLP data, the phase differences of the data in buffer P 172 and buffer O 170 are calculated column wise and then an inverse VFFT is performed with the results being stored in buffer N 168 as illustrated at clock cycle 450. Data from buffer N 168 is then copied to buffer B 166 at clock cycle 500 to facilitate row wise processing. This takes 50 clock cycles. An HIFFT operation is performed on the data in buffer B 166 with the results stored in buffer A 164 at clock cycle 550. This takes 50 clock cycles. Due to the conjugate symmetry memory savings procedures discussed above, only 33 samples of the row are available in buffer B 166. The remaining samples are generated by taking the complex conjugate of the existing columns and the full 64 samples are sent for HIFFT. The 64×32 output samples of the HIFFT represent the phase plane correlation surface that are then searched over the next 150 clock cycles beginning at clock cycle 600 to find the top 4 values corresponding to potential peak values of the phase plane correlation surface. Candidate motion vectors are then determined based upon the identified peaks, and the above described steps are repeated for the high pass data.
The controller 192 of
The inputs and outputs of the data processing modules 202, 204, 206 are connected to the multiplexer module 196. Depending upon the operation being performed, individual multiplexers inside the multiplexer module 196 are connected using selection logic from the controller 192. The FFT module 202 is multiplexed to perform HFFT and VFFT operations. Because the phase plane correlation calculator inputs are always real (no imaginary component), the columns of the block after performing an HFFT operation exhibit conjugate symmetry. Because this symmetry is known based on the real inputs, a memory savings can be realized because after a VFFT operation, the columns from 64/2+1 onwards are circularly shifted versions of columns 1 to 64/2. Due to this known data redundancy, storage for nearly half the columns may be eliminated as long as compensations are made downstream based on this data paring. For example databank A 200 may contain 33 sRAMs, 32 bits wide by 34 locations deep, sized to be able to handle the maximum storage and access requirements. Databank B 200 only stores phase values and can, therefore, be much smaller in size than databank A 200.
In the example of
In operation, the phase plane correlation calculator 190 of
PLP data is read from physical memory, passed through the FFT 202 for HFFT, and stored in databank A 200. The PLP HFFT data is transposed in A as was done with the CLP HFFT data previously. A VFFT operation is performed on the data in databank A 200 followed by phase extraction 206 with the resulting phases being stored in databank A 200. The phase differences 204 of the data in databanks A and B 200 are calculated column wise, and the results of an inverse VFFT operation 202 on the phase differences are stored in databank A 200.
The data in databank A 200 is transposed to enable row processing, and an inverse HFFT is performed on the data in databank A 200 using the FFT module 202 with the results being stored in databank A 200. While performing the HIFFT, only 33 samples of the row are available in B due to the storage savings made possible based on the conjugate symmetry of the FFT of real input values as described above. The remaining samples are generated by taking the complex conjugate of the existing columns, and the full 64 samples are sent to HIFFT. A PPC search is done on the data in databank A 200 to find peak values in the phase plane correlation surface, and corresponding output motion vectors are determined and output as illustrated at 210. These steps are then repeated for high pass data.
One method of selecting a final motion vector is by making a quality calculation for each of the candidate motion vectors and selecting the candidate motion vector that yields the best score. Candidate motion vectors may come from estimation calculations such as the phase plane correlation calculation described above, motion vectors from neighboring blocks, from other motion vector estimations such as the global motion calculations described below, as well as a variety of other sources. The quality calculation may include a variety of matching techniques including sum of absolute differences (SAD) and sum of squared differences (SSD). The quality calculation may also include other terms such as a smoothness term. For example, the following quality calculation formula may be utilized where the vector resulting in the lowest cost is selected as the final motion vector:
Cost(x,y,t)=α∫f(Data)+β∫f(smoothness).
The first function corresponds to a distortion measure such as SAD or SSD. If SAD is selected, the first function could be represented as:
f(Data)=Σ|I1(x,y,t1)−I2(x−dxi,y−dyi,t2)|,
where I1 and I2 are the current and previous frames, respectively, and dx, and dye represent a candidate motion vector.
As noted above, picture rate up-conversion may be performed from a variety of source video rates to a variety of target data rates. The factor of the source to target data rates dictates the number of intermediate frames that are interpolated into the target video. For example, an up-conversion from a 60 Hz source video to a 120 Hz target video results in the insertion of 1 frame between frames of the source video in the target video. Thus, one frame is inserted halfway between source video frames resulting in an interpolation factor of 0.5: 60 Hz/120 Hz=0.5. For conversion from a 24 Hz source video to 120 Hz target video, four frames are inserted between source frames in the target video. Inserting four frames between source video frames causes an intermediate frame to be inserted every 0.2 source frames resulting in interpolation factors of 0.2, 0.4, 0.6, and 0.8: 24 Hz/120 Hz=0.2.
The interpolation factor is utilized in generating intermediate frames. A final motion vector selected by the VMV engine 222 corresponds to the detected motion between the previous frame and the target frame. However, in the example of 60 Hz to 120 Hz conversion, the intermediate frame will depict object motion halfway between the previous frame and the target frame. Thus, when calculating the proper motion of objects within a block in an intermediate frame, the final motion vector is multiplied by the interpolation factor, 0.5, to capture object position at the time of interest (i.e., the time of the intermediate frame). Similarly with 24 Hz to 120 Hz conversion, the first intermediate frame utilizes the final motion vector multiplied by the first interpolation factor, 0.2, the second intermediate frame utilizes the final motion vector multiplied by the second interpolation factor, 0.4, the third intermediate frame utilizes the final motion vector multiplied by the third interpolation factor, 0.6, and the fourth intermediate frame utilizes the final motion vector multiplied by the fourth interpolation factor, 0.8.
In operation, the example of
For 5× interpolation (e.g., 24 Hz to 120 Hz), as “i” increases from 0.2 to 0.8, the boundary that defines the set of pixels used for SAD calculation increases in the previous frame and decreases in the current frame. This fact may be used to optimize bandwidth for memory transfers. The VMV engine 222 may use the RGB representation from the frame that has a smaller boundary and the luminance representation from the other frame. Thus, for interpolation factors 0.2 and 0.4, the RGB pixel from the previous frame and the luminance pixel from the current frame are used for SAD calculations. For interpolation factors 0.6 and 0.8, RGB pixels from the current frame and luminance pixels from the previous frames are utilized. The worst case bandwidth required occurs at an interpolation factor of 0.5.
The VMV buffer 254 provides motion vectors to a shift register 262. The shift register 262 receives data associated with the previous frame window through a double buffer 260. The previous frame window of
The SAD calculator 270 also receives current frame data via the current frame read path 264, 266, 268. The RGB block read DMA 264 accesses current frame data via the SDRAM controller 256 and provides the current frame data to the RGB window block 266. The RGB window block 266 is multiple buffered similar to the luminance window 260 to facilitate parallel processing. In the example of
The SAD calculator 270 receives the shifted candidate window from the shift register 262 and current frame data from the current frame read path 264, 266, 268. The SAD calculator 270 performs a quality calculation on the shifted candidate frame and the current frame data by comparing the two inputs. If the candidate motion vector associated with the shifted candidate frame is accurate, the two inputs will be similar resulting in a low SAD score. If the candidate motion vector is poor, the two inputs will be very different resulting in a high SAD score. The SAD calculator 270 selects the candidate motion vector associated with the best quality calculation score as the final motion vector for a pixel.
Following selection of a final motion vector, one or more intermediate frames are generated in the compensator block 272 and stored in the compensation buffer 276, which may be implemented as a double buffer. The dynamic average block 278 receives the generated intermediate frames and applies error correction and concealment procedures. The dynamic average block 278 outputs the intermediate frames to a write-back buffer 280 where the frames await being written to memory via the RGB block write DMA 282.
The steps of
In
In estimating global motion, the most dominant motion in video is translation. In an affine motion representation, the affine parameters that represent the zoom and rotation are small in magnitude and may be estimated using a sparse motion field. Large translation estimation is more difficult in comparison. Phase correlation based estimation, as described above, measures sub-pixel motion with limited complexity. Using the predicted values from the phase correlation values as an initial guess assuming no rotation or scaling in the image, the subsequent RANSAC (RANdom SAmple Consensus) based estimation stage refines the motion parameters.
The affine model is chosen for simplicity. However, the global motion calculation may be extended to any other motion model such as projective, etc. The affine motion model may be expressed as:
X′=A*X+B,
where X=[x,y]T represents a position of a selected pixel group in the previous frame, X′=[x′,y′]T represents a position of a selected pixel group in the current frame, where
represents motion parameters corresponding to zoom and rotation of the selected pixel group from the previous frame to the current frame, and
B=[bx,by]T
represents motion parameters corresponding to pan motion of the pixel group from the previous frame to the current frame.
The selected blocks are sorted based upon their level of activity in step 408. In general, activity may be measured as a sum of absolute difference from the mean, variance, or eigenvalues of a windowed image second moment matrix. These three metrics are listed in increasing order of computational complexity but also in increasing order of reliability. Another possibility is to look for a significant bi-direction intensity gradient inside a block. The sorting is intended to identify promising blocks for motion estimation because low activity blocks are likely to give incorrect motion vectors due to aperture affect. The top Nf blocks from the sorted list are considered promising blocks and are kept under consideration in step 410 while the remaining blocks are discarded. Thus, an adaptive threshold on the activity is implicitly applied.
The translation between each of the high activity Nf promising blocks in the current frame and its counterpart in the previous frame is measured in step 412 based upon the phase correlation values calculated before. Phase-correlation provides two advantages as compared to other methods, in this regard. First, phase correlation measures sub-pixel-accurate translation with reasonably small amount of computations. Second, the translation measurement result is almost independent of minority outlier pixels, which may be due to foreground pixels. When neither background nor foreground is dominant, it gives two motion vectors, one corresponding to each.
After translation measurement is completed in step 412, the data is passed to a RANSAC-based robust least-square model for fitting in step 414 as a set of Nf pentuples (x,y,1,dx,dy), where x,y is the coordinate of the center of the block, dx is the translation of the block along the x-axis, and dy is the translation of the block along the y axis. Step 414 is described in further detail with reference to
where I2 is a 2×2 identity matrix.
The parameter values that are calculated in step 434 may not be consistent with affine parameters calculated over previous iterations of the loop beginning at step 428. If a data-point is consistent with a particular value of affine parameters, then it is said to be in the support set of the affine parameter. Support for the current affine parameters is calculated in step 436. The number of supporting data-points, which exhibit a motion vector consistent with the currently determined affine parameters, is computed by counting the number of data-points for which the norm (e.g., L1 or L2) of the following error metric is below a suitable threshold:
The “best affine parameter” refers to that value of affine parameters encountered so far that had the largest support set. The size of the support set for the best affine parameter is stored in the max support variable. Thus, in step 438, an inquiry is made as to whether the support for the current affine parameters is greater than the value stored in the max support variable. If the current support is greater, the max support value is changed to the current support value, and the best affine parameters array is reset to the current affine parameters value in step 442. If the current support is not greater, then the current max support and best affine parameters values are maintained. In either case, the iteration index is incremented and the loop is restarted as long as the iteration index is less than the max iterations value.
Once the iteration index is equal to the maximum iterations value, branch 448 is taken and data-points not supporting the best affine parameters are discarded in step 450. A least-square plane fitting is applied to the retained data-points in step 452 to determine the refined affine parameters 454 which are utilized to generate an estimated motion vector.
The detailed portion of the target area 522 has an edge propagation technique applied. An edge map of the data is analyzed and the direction of the edge is computed in the neighborhood region near the target area 522. Often, an edge direction is consistent over small distances. Thus, the edge pixels are copied from the neighborhood region and filled in the hole based on a sum of squared differences or other matching criteria. Because detailed regions tend to have more information in the neighboring regions, edge propagation may be performed before texture syntheses. Thus, the entire edge 524 may be restored prior to addressing any regions that require texture filling 526.
One such advanced technique is depicted in
While insertion of a motion compensated intermediate frame into an up-converted target video often improves the target video by increasing sharpness and reducing blur and judder, sometimes it may be preferable to repeat the previous frame when up-converting rather than using the motion interpolated frame.
This written description uses examples to disclose the invention, including the best mode, and also to enable a person skilled in the art to make and use the invention. The patentable scope of the invention may include other examples.
Biswas, Mainak, Namboodiri, Vipin, Patankar, Kaustubh Milind, Srivivasan, Sujith
Patent | Priority | Assignee | Title |
10354394, | Sep 16 2016 | Dolby Laboratories Licensing Corporation | Dynamic adjustment of frame rate conversion settings |
10977809, | Dec 11 2017 | Dolby Laboratories Licensing Corporation | Detecting motion dragging artifacts for dynamic adjustment of frame rate conversion settings |
11302280, | Jan 02 2019 | BEIJING BOE OPTOELECTRONICS TECHNOLOGY CO , LTD ; BOE TECHNOLOGY GROUP CO , LTD | Compensation method and compensation device for VR display and display device |
11734835, | Feb 22 2019 | Nippon Telegraph and Telephone Corporation | Image processing apparatus, image processing method, and program |
9959600, | May 28 2015 | BOE TECHNOLOGY GROUP CO , LTD | Motion image compensation method and device, display device |
Patent | Priority | Assignee | Title |
8462266, | Apr 27 2007 | Sharp Kabushiki Kaisha | Image processing device and method, and image display device and method |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
May 21 2013 | Marvell International Ltd. | (assignment on the face of the patent) | / | |||
Jun 11 2017 | MARVELL INTERNATIONAL LTD | Synaptics Incorporated | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 043853 | /0827 | |
Sep 27 2017 | Synaptics Incorporated | Wells Fargo Bank, National Association | SECURITY INTEREST SEE DOCUMENT FOR DETAILS | 044037 | /0896 |
Date | Maintenance Fee Events |
Nov 01 2018 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Oct 20 2022 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Date | Maintenance Schedule |
May 12 2018 | 4 years fee payment window open |
Nov 12 2018 | 6 months grace period start (w surcharge) |
May 12 2019 | patent expiry (for year 4) |
May 12 2021 | 2 years to revive unintentionally abandoned end. (for year 4) |
May 12 2022 | 8 years fee payment window open |
Nov 12 2022 | 6 months grace period start (w surcharge) |
May 12 2023 | patent expiry (for year 8) |
May 12 2025 | 2 years to revive unintentionally abandoned end. (for year 8) |
May 12 2026 | 12 years fee payment window open |
Nov 12 2026 | 6 months grace period start (w surcharge) |
May 12 2027 | patent expiry (for year 12) |
May 12 2029 | 2 years to revive unintentionally abandoned end. (for year 12) |