A display device displays an image in which each frame is formed by at least two subframes. The display device includes a drive unit configured to drive a plurality of pixels in a pixel array so as to drive pixels in at least two rows based on pixel data of each row of each supplied subframe data. The drive unit drives the plurality of pixels so as to cause pixels in a row not matching a row in the supplied current subframe data to emit light with a first condition, and cause pixels in a row matching the row in the current subframe data to emit light with a second condition. The second condition is a condition in which a light emission amount is larger than in the first condition when causing the pixels to emit light in accordance with identical pixel values.
|
12. A display device that uses image data in which one frame is formed by at least two subframes, and drives a plurality of rows including a first row using data based on the same data, wherein
the one frame includes a first subframe for which the first row and another row among the plurality of rows are driven using data based on drive data of the first row, and a second subframe for which the first row and another row among the plurality of rows are driven using data based on drive data of the other row, and
a light emission amount of at least one pixel included in the first row in the first subframe is larger than a light emission amount thereof in the second subframe.
1. A display device configured to display an image in which each frame is formed by at least two subframes, comprising:
a drive unit configured to drive a plurality of pixels in a pixel array so as to drive pixels in at least two rows based on pixel data of each row of each supplied subframe data, wherein
the drive unit drives the plurality of pixels so as to cause pixels in a row not matching a row in the supplied current subframe data to emit light with a first condition, and cause pixels in a row matching the row in the current subframe data to emit light with a second condition,
the second condition is a condition in which a light emission amount is larger than in the first condition when causing the pixels to emit light in accordance with identical pixel values.
2. The display device according to
in the first condition, a light emission period for causing pixels to emit light is a first light emission period, and in the second condition, a light emission period for causing pixels to emit light is a second light emission period longer than the first light emission period.
3. The display device according to
each of the plurality of pixels includes a transistor configured to control a light emission period, and
the drive unit controls the transistor of each of the plurality of pixels so as to cause pixels in a row not matching a row in supplied current subframe data to emit light in accordance with the first light emission period, and cause pixels in a row matching the row in the current subframe to emit light in accordance with the second light emission period.
4. The display device according to
the drive unit includes a plurality of selection units arranged such that one selection unit is assigned to one row, and each selection unit selects, from a plurality of control signals including a first control signal that defines a first light emission period and a second control signal that defines a second light emission period longer than the first light emission period, a control signal corresponding to supplied current subframe data, and outputs the selected control signal, and
the transistors of pixels in each row are controlled by an output of a corresponding selection unit among the plurality of selection units.
5. The display device according to
the second condition is a condition in which light emission intensity is higher than in the first condition.
6. The display device according to
each of the plurality of pixels includes a memory, and is configured to emit light in accordance with a signal written in the memory, and
the drive unit controls, so as to satisfy the first condition and the second condition, a pixel signal to be written in the memory of each of the plurality of pixels based on supplied current subframe data.
7. The display device according to
the drive unit includes a first data line configured to supply pixel signals to pixels in a row matching a row in first subframe data of supplied subframe data, and a second data line configured to supply pixel signals to pixels in a row matching a row in second subframe data of the supplied subframe data, and
the drive unit controls, so as to satisfy the first condition and the second condition, pixel signals to be supplied to the first data line and the second data line based on supplied current subframe data.
8. The display device according to
the at least two subframes include an even-numbered subframe and an odd-numbered subframe.
9. The display device according to
a determination unit configured to determine that supplied current subframe data is one the at least two subframes.
10. An electronic apparatus comprising:
an image capturing unit configured to capture an object; and
a display device defined in
11. A moving body comprising:
an image capturing unit configured to capture an object; and
a display device defined in
13. The display device according to
the at least two subframes include an even-numbered subframe and an odd-numbered subframe.
14. The display device according to
a determination unit configured to determine the first subframe or the second subframe based on the drive data.
|
This application is a Continuation of International Patent Application No. PCT/JP2021/036922, filed Oct. 6, 2021, which claims the benefit of Japanese Patent Application No. 2020-172134, filed Oct. 12, 2020, both of which are hereby incorporated by reference herein in their entirety.
The present invention relates to a display device, an electronic apparatus, and a moving body.
There is a display device that drives pixels by a multi-line drive method for driving pixels in two or more rows using the pixel data of each row in order to reduce the input data amount and shorten the time required for frame display. In the multi-line drive method, for each subframe, two or more rows to be supplied with the pixel data of each row are changed to distribute an original image and an interpolation image, thereby improving the visibility. Since human vision is less sensitive than the drive cycle of a display, a human perceives an image obtained by integrating and averaging the original image and the interpolation image.
Japanese Patent Laid-Open No. 2003-532145 proposes a method of improving image quality by performing correction such that the average luminance value between two consecutive subframes becomes equal to the original luminance value. However, with the method described in Japanese Patent Laid-Open No. 2003-532145, a memory for holding the luminance of the previous frame for one frame period is necessary, and this increases the circuit scale.
The present invention provides a technique advantageous in improving the quality of a display image with a simple arrangement.
One of aspects of the present invention provides a display device configured to display an image in which each frame is formed by at least two subframes, comprising: a drive unit configured to drive a plurality of pixels in a pixel array so as to drive pixels in at least two rows based on pixel data of each row of each supplied subframe data, wherein the drive unit drives the plurality of pixels so as to cause pixels in a row not matching a row in the supplied current subframe data to emit light with a first condition, and cause pixels in a row matching the row in the current subframe data to emit light with a second condition, the second condition is a condition in which a light emission amount is larger than in the first condition when causing the pixels to emit light in accordance with identical pixel values.
According to the present invention, a technique advantageous in improving the quality of a display image with a simple arrangement is provided.
Further features of the present invention will become apparent from the following description of exemplary embodiments with reference to the attached drawings.
Preferred embodiments according to the present invention will now be described hereinafter with reference to the accompanying drawings. It is to be understood that the following embodiments are not intended to limit the claims of the present invention, and that not all of the combinations of the aspects that are described according to the following embodiments are necessarily required with respect to the means to solve the problems according to the present invention.
The pixel array 10 includes a plurality of pixels arranged so as to form a plurality of rows and a plurality of columns. In
The gate of the transistor Tr1 is connected to a vertical scanning line VSLx, ( . . . , VSLN−1, VSLN, VSLN+1) driven by the vertical scanning control unit 11. The vertical scanning line VSLx is activated when writing a pixel signal in the capacitor C through data DLx (DL1, DL2, . . . ) driven by the horizontal scanning control unit 14. The transistor Tr2 supplies, to the light emitting element EL, a current corresponding to the pixel signal written in the capacitor C. The gate of the transistor Tr3 is connected to a light emission period signal line EPx ( . . . , EPN−1, EPN, EPN+1) driven by the light emission period control unit 13 to control the light emission period of the light emitting element EL.
Image data ID of each frame supplied to the display device DD1 can include even-numbered subframe data which is data of an even-numbered subframe, and odd-numbered subframe data which is data of an odd-numbered subframe. When writing pixel data of the even-numbered subframe data, the transistors Tr1 of the pixels 101 and 102 in the (N−1)th and Nth rows are driven so that the pixels 101 and 102 in the (N−1)th and Nth rows are simultaneously driven. When writing pixel data of the odd-numbered subframe data, the transistors Tr1 of the pixels 102 and 103 in the Nth and (N+1)th rows are driven so that the pixels 102 and 103 in the Nth and (N+1)th rows are simultaneously driven. In order to implement this, the vertical scanning control unit 11 includes a plurality of selection units 11x (111, 112, and 113).
The plurality of selection units 11x can be arranged such that one selection unit 11x is assigned to one row. The image data ID of each frame is supplied to the determination unit 12. For each frame, the determination unit 12 can determine one of two subframes forming the frame the supplied current subframe data belongs to, and output selection signals RAS and RSB indicating the result of the determination. The supplied current subframe data can be the even-numbered subframe data or the odd-numbered subframe data. The selection signals RSA and RSB are complementary or exclusive signals, and the selection signal RSB can be the inverted signal of the selection signal RSA. When writing the pixel data of the even-numbered subframe data, the determination unit 12 can drive the selection signal RSB to the active level and the selection signal RSA to the inactive level. When writing the pixel data of the odd-numbered subframe data, the determination unit 12 can drive the selection signal RSA to the active level and the selection signal RSB to the inactive level. If each frame is formed by at least two subframes, the determination unit 12 can be configured to determine one of the at least two subframes forming the frame the current subframe data belongs to.
Each selection unit 11x of the vertical scanning control unit 11 can be formed by a multiplexer that outputs a signal having the same logic level as the input signal designated by the selection signal from the first and second input signals. The selection unit 11x, to which the selection signal RSA is input, can output a signal having the same logic level as the first input signal if the selection signal RSA is at the active level, and output a signal having the same logic level as the second input signal if the selection signal RSA is at the inactive level. The selection unit 11x, to which the selection signal RSB is input, can output a signal having the same logic level as the first input signal if the selection signal RSB is at the active level, and output a signal having the same logic level as the second input signal if the selection signal RSB is at the inactive level. Each selection unit 11x drives the vertical scanning line VSLx in the corresponding row by the output thereof. For example, when writing the pixel data of the even-numbered subframe data, the selection unit 112 assigned to the Nth row outputs a signal having the same logic level as a row selection signal R N to the vertical scanning signal line VSLN.
In order to write a pixel signal in the capacitor C (that is, charge or discharge), which is used to control the light emission luminance of each pixel 10x, the horizontal scanning control unit 14 outputs the pixel signal to the data line DLx in each column. Here, when writing the pixel data of the even-numbered subframe data, a pixel signal PN corresponding to the pixel data of the Nth row of the even-numbered subframe data is written in the capacitors C of the pixels 101 and 102 in the (N−1)th and Nth rows. That is, the pixel signal PN corresponding to the pixel data of the Nth row is written in the capacitor C of the pixel 102 in the Nth row, and the pixel signal PN corresponding to the pixel data for the pixel 102 in the Nth row, which is adjacent to the pixel 101, is also written in the capacitor C of the pixel 101 in the (N−1)th row. When writing the pixel data of the odd-numbered subframe data, a pixel signal PN+1 corresponding to the pixel data of the (N+1)th row of the odd-numbered subframe data is written in the capacitors C of the pixels 102 and 103 in the Nth and (N+1)th rows. That is, the pixel signal PN+1 corresponding to the pixel data for the (N+1)th pixel 103, which is adjacent to the pixel 102, is written in the capacitor C of the pixel 102 in the Nth row, and the pixel signal PN+1 corresponding to the pixel data for the (N+1)th pixel 103 is also written in the capacitor C of the pixel 102 in the (N+1)th row.
According to the multi-line drive as described above, in two subframe periods, light emission of the pixel 102 in the Nth row is perceived by the human as light emission corresponding to PN/2+PN+1/2. Here, assume that the image data ID is for a still image (repetition of a single image), and the luminance levels of the pixels in the (N−1)th, Nth, and (N+1)th rows are 20, 20, and 80, respectively. In this case, light emission of the pixel 102 in the Nth row is perceived as the integrated average of 20, 20, and 80, that is, a luminance of 50. Accordingly, a deviation from the original pixel data having the luminance level of 20 occurs, and image quality degradation can occur.
With reference to
For each frame, the determination unit 12 can determine, based on the image data ID or a signal accompanying the image data ID, one of at least two subframes (here, the even-numbered subframe and the odd-numbered subframe) forming the frame the supplied current subframe data belongs to. For example, if the image data ID is supplied as an interlace signal, this determination can be made using the characteristic that the start of scanning of a vertical synchronization signal is shifted by 0.5H. Alternatively, one of the even-numbered subframe data and the odd-numbered subframe data that is transmitted first at the start of transmission of the image data ID may be defined, and the determination can be made based on this. Alternatively, for example, if transmission is performed according to a standard such as MIPI or the like, it may be defined that, for example, the even-numbered subframe data is transferred in an event mode and the odd-numbered subframe data is transferred in a pulse mode, and the determination can be made based on this. Alternatively, determination information accompanying the image data may be transmitted as a sideband signal, and the determination can be made based on this. The determination may be made according to other methods. Based on the determination as described above, the determination unit 12 can generate selection signals RSA and RSB. When writing the pixel data of the even-numbered subframe data, the determination unit 12 can drive the selection signal RSB to the active level and the selection signal RSA to the inactive level. When writing the pixel data of the odd-numbered subframe data, the determination unit 12 can drive the selection signal RSA to the active level and the selection signal RSB to the inactive level. Since the selection signal RSA is driven to the active level in the odd-numbered subframe and the selection signal RSB is driven to the active level in the odd-numbered subframe, the selection signals RSA and RSB are alternately driven to the active level.
The vertical scanning control unit 11 generates a row selection signal Rx for controlling vertical scanning (row selection) of the pixel array 10. The row selection signal Rx can be formed by, for example, a shift register. The row selection signal Rx can be driven to the active level in the order of R1, R2, . . . , RN−1, RN, RN+1, and so on over a predetermined period. In multi-line drive, when writing the pixel data of the even-numbered subframe data in a pixel 10x, the vertical scanning control unit 11 can generate a vertical scanning signal such that pixels 101 and 102 in the (N−1)th and Nth rows are simultaneously driven. When writing the pixel data of the odd-numbered subframe data in the pixel 10x, the vertical scanning control unit 11 can generate a vertical scanning signal such that the pixel 102 and a pixel 103 in the Nth and (N+1)th rows are driven simultaneously.
When writing the pixel data of the even-numbered subframe data, a selection unit 111 in the (N−1)th row can output, to a vertical scanning signal line VSLN−1 in the (N−1)th row, a signal having the same logic level as a row selection signal RN of the Nth row in accordance with the selection signal RSA. Further, when writing the pixel data of the even-numbered subframe data, a selection unit 112 in the Nth row can output, to a vertical scanning signal line VSLN in the Nth row, a signal having the same logic level as the row selection signal RN of the Nth row in accordance with the selection signal RSB.
When writing the pixel data of the odd-numbered subframe data, the selection unit 112 in the Nth row can output, to the vertical scanning signal line VSLN− in the Nth row, a signal having the same logic level as a row selection signal RN+1 of the (N+1)th row in accordance with the selection signal RSB. Further, when writing the pixel data of the odd-numbered subframe data, a selection circuit 113 in the (N+1)th row can output, to a vertical scanning signal line VSLN+1 in the (N+1)th row, a signal having the same logic level as the row selection signal RN+1 of the (N+1)th row in accordance with the selection signal RSA.
The drive unit DRV can be configured to drive the pixels 10x in at least two rows based on the pixel data of each row of each subframe data. The drive unit DVR can drive a plurality of the pixels 10x so as to cause the pixel in a row not matching a row in the supplied current subframe data to emit light with a first condition, and cause the pixel in a row matching the row in the current subframe data to emit light with a second condition. The second condition can be a condition in which the light emission amount is larger than in the first condition when causing the pixels to emit light in accordance with identical pixel values. In the first embodiment, in the first condition, the light emission period for causing the pixel to emit light is a first light emission period, and in the second condition, the light emission period for causing the pixel to emit light is a second light emission period longer than the first light emission period.
The light emission period control unit 23 can control a transistor Tr3 of each of the plurality of the pixels 10x so as to cause the pixel 10x in the row not matching the row in the supplied current subframe data to emit light in accordance with the first light emission period. Further, the light emission period control unit 23 can control the transistors Tr3 of the plurality of the pixels 10x so as to cause the pixel in the row matching the row in the supplied current subframe data to emit light in accordance with the second light emission period.
The light emission period control unit 23 can include a period signal generation unit 25. The control signal generation unit 25 can generate a first period control signal EPAx that defines a first light emission period T1, and a second period control signal EPBx that defines a second light emission period T2 longer than the first light emission period T1. The light emission period control unit 23 can include a plurality of selection units 23x (231, 232, and 233). The plurality of selection units 23x can be arranged such that one selection unit 23x is assigned to one row. Each selection unit 23x can select, from the first period control signal EPBx that defines the first light emission period T1 and the second period control signal EPAx that defines the second light emission period T2 longer than the first light emission period T1 (a plurality of control signals including them), the period control signal corresponding the current subframe data, and output the selected period control signal. The transistor Tr3 of the pixel 10x in each row can be controlled by the output of the corresponding selection unit among the plurality of selection units 23x.
In an example, the first light emission period T1 is a light emission period shorter than the light emission period in an optimized comparative example, and the second light emission period T2 is a light emission period longer than the light emission period in the comparative example. The average of the first light emission period T1 and the second light emission period T2 can match the light emission period in the comparative example. For example, assume a case in which, in order to suppress power consumption, the light emission period in the comparative example is designed to be 70% the longest period. In this case, for example, the first light emission period T1 can be 40% the longest period, and the second light emission period T2 can be 100% the smallest period. Alternatively, assume a case in which the light emission period in the comparative example is designed to be 90% the longest period. In this case, for example, the first light emission period T1 can be 80% the longest period, and the second light emission period T2 can be 100% the smallest period.
The first period control signal EPAx and the second period control signal EPBx may be generated so as to cause a light emitting element EL of the pixel 10x to emit light continuously, or may be generated so as to cause the light emitting element EL of the pixel 10x to emit light intermittently.
In the even-numbered subframe, the light emission period of the light emitting element EL of each of the pixels 101 and 103 in the (N−1)th and (N+1)th rows can be controlled in accordance with the first period control signal EPAx that defines the first light emission period T1 which is the relatively short light emission period. Further, in the even-numbered subframe, the light emission period of the light emitting element EL of the pixel 102 in the Nth row can be controlled in accordance with the second period control signal EPBx that defines the second light emission period T2 which is the relatively long light emission period. On the other hand, in the odd-numbered subframe, the light emission period of the light emitting element EL of the pixel 102 in the Nth row can be controlled in accordance with the first period control signal EPAx that defines the first light emission period T1 which is the relatively short light emission period. Further, in the odd-numbered subframe, the light emission period of the light emitting element EL of each of the pixels 101 and 103 in the (N−1)th and (N+1)th rows can be controlled in accordance with the second period control signal EPBx that defines the second light emission period T2 which is the relatively long light emission period.
The horizontal scanning control unit 14 can generate, based on the image data ID, a pixel signal (voltage signal) to be written in a capacitor C of the pixel 10x in each column of the selected row, and output the pixel signal to a data line DLx in each column. Scanning of the pixel array 10 by the vertical scanning control unit 11 and output of the pixel signal to the data line DLx by the horizontal scanning control unit 14 can be controlled in synchronization with each other. In the multi-line drive, in each column, a single pixel signal can be written in two pixels 10x adjacent to each other in the vertical direction.
Each pixel 10x can control light emission of the light emitting element EL based on the vertical scanning signal from the vertical scanning control unit 11, the period control signal from the light emission period control unit 23, and the pixel signal from the horizontal scanning control unit 14. The light emission luminance of the light emitting element EL can be defined by the pixel signal (voltage signal) written in the capacitor C. The light emission period of the light emitting element EL can be controlled by the period control signals EPAx and EPBx.
In the even-numbered subframe, the pixel signal for the pixel 102 in the Nth row can be written in the capacitors C of the pixels 101 and 102 in the (N−1)th and Nth rows via the data line DLx. Then, the light emission period of the light emitting element EL of the pixel 101 in the (N−1)th row can be controlled over the first light emission period T1 defined by the first control signal EPAN−1. The light emission period of the light emitting element EL of the pixel 102 in the Nth row can be controlled over the second light emission period T2 defined by the second control signal EPBN. As exemplarily shown in
In the odd-numbered subframe, the pixel signal for the pixel 103 in the (N+1)th row can be written in the capacitors C of the pixels 102 and 103 in the Nth and (N+1)th rows via the data line DLx. Then, the light emission period of the light emitting element EL of the pixel 102 in the Nth row can be controlled over the first light emission period T1 defined by the first control signal EPAN. The light emission period of the light emitting element EL of the pixel 103 in the (N+1)th row can be controlled over the second light emission period T2 defined by the second control signal EPBN+1. The light emission period of the light emitting element EL of the pixel 103 in the (N+1)th row can be controlled over the second light emission period T2 defined by the first control signal EPAN+1. As exemplarily shown in
In
If the first light emission period T1 is set to be 0.5 times the light emission period in the comparative example, and the second light emission period T2 is set to be 1.5 times the light emission period in the comparative example, the average value of the difference absolute values in the first embodiment is 10.6, and the higher effect is exhibited. However, if the ratio difference of each of the first light emission period T1 and the second light emission period T2 in the first embodiment with respect to the light emission period in the comparative example is increased, the fluctuation amount of the light emission luminance in one frame period increases, and this can be perceived as flicker. To prevent this, it is preferable to control with the upper limit of about 1.5 times, but the present invention is not limited to this.
In the first embodiment, the quality of a display image can be improved by using the supplied current subframe data alone. Accordingly, as compared to the method using a frame memory or the like, the quality of a display image can be improved with a simple arrangement.
In the description so far, for the sake of simplicity, an example has been described in which the light emission period of the pixel having the above-described structure is controlled. In addition to this, methods such as a method of generating, based on a triangular wave, a signal for controlling the light emission period are conceivable. In this case, it is possible to control the light emission period using a control method in which, upon generating the signal for controlling the light emission period, the threshold value for pulse conversion is changed for each row, or a method in which the tilt of the triangular wave is changed for each row.
The light emitting element EL may be an organic EL element, or may be another element. The light emitting element EL may be replaced with a liquid crystal element. In this case, a method is conceivable in which the luminance is adjusted by, for example, controlling the lighting period of a backlight for each row based on the period control signal in the first embodiment. By performing, in accordance with the light emission period, write for clearing electric charges held in the pixel capacitor at the time of driving the liquid crystal, and changing the light emission period for each row in accordance with the period control signal, a similar effect can be obtained.
The configuration has been described above in which the pixels in two rows are mainly driven using the pixel data of one row and the determination unit determines the even-numbered subframe data or the odd-numbered subframe data. However, by extending this embodiment, the present invention can be applied to a configuration for simultaneously driving pixels in more rows and a configuration using input data having two or more kinds of attributes.
With reference to
The pixel array 80 includes a plurality of pixels arranged so as to form a plurality of rows and a plurality of columns. In
The gate (Tr1/G) of the transistor Tr1 is connected to a vertical scanning line VSLx ( . . . , VSLN−1, VSLN, VSLN+1) driven by the vertical scanning control unit 11. The vertical scanning line VSLx is activated when writing a pixel signal in the capacitor C through data DLAx and DLBx (DLA1, DLB1, DLA2, DLB2, . . . ) driven by the horizontal scanning control unit 84. The transistor Tr2 can supply, to the light emitting element EL, a current corresponding to the pixel signal written in the capacitor C.
The drive unit DRV′ can be configured to drive the pixels 80x in at least two rows based on the pixel data of each row of each subframe data. The drive unit DVR′ can drive a plurality of the pixels 80x so as to cause the pixel in a row not matching a row in the supplied current subframe data to emit light with a first condition, and cause the pixel in a row matching the row in the current subframe data to emit light with a second condition. The second condition can be a condition in which the light emission intensity (luminance) is higher than in the first condition.
The horizontal scanning control unit 84 forming a part of the drive unit DRV′ can include the first data line DLAx configured to supply a pixel signal to the pixel in the row not matching the row in the first subframe data of at least two subframe data. The horizontal scanning control unit 84 can also include the second data line DLBx configured to supply a pixel signal to the pixel in the row matching the row in the second subframe data of at least two subframe data. More specifically, the horizontal scanning control unit 84 can include the first data line DLAx configured to supply a pixel signal to the pixels 801 and 803 in the (N−1)th row not matching the Nth row in the odd-numbered subframe data among the even-numbered subframe data and the odd-numbered subframe data. The horizontal scanning control unit 84 can also include the second data line DLBx configured to supply a pixel signal to the pixel 802 in the Nth row matching the Nth row in the even-numbered subframe data among the even-numbered subframe data and the odd-numbered subframe data. In each column, the pixel 80, connected to the first data line DLAx and the pixel 80x connected to the second data line DLBx can be alternately arranged.
When writing the pixel data of the even-numbered subframe data, the transistors Tr1 of the pixels 801 and 802 in the (N−1)th and Nth rows can be driven so that the pixels 801 and 802 in the (N−1)th and Nth rows are simultaneously driven. When writing the pixel data of the odd-numbered subframe data, the transistors Tr1 of the pixels 802 and 803 in the Nth and (N+1)th rows can be driven so that the pixels 802 and 803 in the Nth and (N+1)th rows are simultaneously driven. In order to implement this, the vertical scanning control unit 11 includes a plurality of selection units 11x (111, 112, and 113).
When writing the pixel data of the even-numbered subframe data, the horizontal scanning control unit 84 generates pixel signals to be supplied to the first data line DLAx and the second data line DLBx, respectively, based on the pixel data of the Nth row of the image data ID. At this time, the horizontal scanning control unit 84 generates the pixel signals to be supplied to the first data line DLAx and the second data line DLBx, respectively, so as to satisfy the first condition and the second condition. More specifically, when writing the pixel data of the even-numbered subframe data, the horizontal scanning control unit 84 can make the value of the pixel signal to be supplied to the second data line DLBx larger than the value of the pixel signal to be supplied to the first data line DLAx. More specifically, if the pixel data of the Nth row is d, the horizontal scanning control unit 84 can supply the pixel signal corresponding to d−α (α is a positive numeric value) to the first data line DLAx, and supply the pixel signal corresponding to d+β (β is a positive numeric value) to the second data line DLBx. Here, the value of β may be equal to or different from the value of α. The pixel signal supplied to the first data line DLAx is written in the capacitor C of the pixel 801 in the (N−1)th row via the transistor Tr1 of the pixel 801 in the (N−1)th row. The pixel signal supplied to the second data line DLBx is written in the capacitor C of the pixel 802 in the Nth row via the transistor Tr1 of the pixel 802 in the Nth row. With this, in the even-numbered subframe, the light emission intensity of the pixel 802 in the (N−1)th row driven in accordance with the pixel data of the Nth row becomes higher than the light emission intensity of the pixel 801 in the (N−1)th row driven in accordance with the pixel data of the Nth row.
When writing the pixel data of the odd-numbered subframe data, the horizontal scanning control unit 84 generates pixel signals to be supplied to the first data line DLAx and the second data line DLBx, respectively, based on the pixel data of the (N+1)th row of the image data ID. At this time, the horizontal scanning control unit 84 generates the pixel signals to be supplied to the first data line DLAx and the second data line DLBx, respectively, so as to satisfy the first condition and the second condition. More specifically, when writing the pixel data of the odd-numbered subframe data, the horizontal scanning control unit 84 can make the value of the pixel signal to be supplied to the first data line DLAx larger than the value of the pixel signal to be supplied to the second data line DLBx. More specifically, if the pixel data of the (N+1)th row is d, the horizontal scanning control unit 84 can supply the pixel signal corresponding to d+β (β is the positive numeric value) to the first data line DLAx, and supply the pixel signal corresponding to d−α (α is the positive numeric value) to the second data line DLBx. Here, the value of β may be equal to or different from the value of α. The pixel signal supplied to the first data line DLAx is written in the capacitor C of the pixel 803 in the (N+1)th row via the transistor Tr1 of the pixel 803 in the (N+1)th row. The pixel signal supplied to the second data line DLBx is written in the capacitor C of the pixel 802 in the Nth row via the transistor Tr1 of the pixel 802 in the Nth row. With this, in the odd-numbered subframe, the light emission intensity of the pixel 803 in the (N+1)th row driven in accordance with the pixel data of the (N+1)th row becomes higher than the light emission intensity of the pixel 802 in the Nth row driven in accordance with the pixel data of the (N+1)th row.
The processing of deciding the pixel signals to be supplied to the first data line DLAx and the second data line DLBx, respectively, based on the pixel data of each row (and each column) of the image data ID can also be understood as correction processing for obtaining the pixel signals by correcting the data. This correction processing can be processing of performing linear or nonlinear correction on the pixel data of each row of the image data ID.
Here, when the pixel in the Nth row of the image data ID (original image) is a pixel in a low-luminance region, if the pixel signal is generated by performing correction so as to increase the pixel value of the pixel in the even-numbered field, an unnatural display image can be generated. Therefore, for the pixel in the low-luminance region, it is preferable to generate the pixel signal so as to suppress the luminance to be equal to or lower than the luminance in the original image. For example, for data having the pixel value (luminance value) equal to or smaller than a threshold value Th shown in
According to the first and second embodiments, the pixel in the row not matching the row in the supplied current subframe data is caused to emit light with the first condition, and the pixel in the row matching the row in the current subframe data is caused to emit light with the second condition in which the light emission amount is larger than in the first condition. With this, the quality of a display image in multi-line drive can be improved. In addition, in the second embodiment, arbitrary correction processing can be employed, so that a further improvement in quality of a display image can be expected.
It can also be said that the display device according to each of the first and second embodiments is a display device that uses image data in which one frame is formed by at least two subframes, and drives a plurality of rows including a first row using data based on single data, wherein the one frame includes a first subframe for which the first row and another row among the plurality of rows are driven using data based on drive data of the first row, and a second subframe for which the first row and another row among the plurality of rows are driven using data based on drive data of the other row, and a light emission amount of at least one pixel included in the first row in the first subframe is larger than a light emission amount thereof in the second subframe.
The display device according to this embodiment may be used for a display unit of an image capturing apparatus that includes an optical unit including a plurality of lenses and an image sensor that receives light passing through the optical unit. The image capturing apparatus may include the display unit that displays information acquired by a plurality of image sensors included in an image capturing unit. Alternatively, information may be acquired using the information acquired by the image sensor, and the display unit may display information different from the information acquired by the image sensor. The image capturing apparatus may be a digital camera or a digital video camera.
Since the timing suitable for capturing an image is a short time, it is preferable to display the information as quickly as possible. Therefore, it is preferable to use the display device according the above-described embodiment. The light emitting element can be formed by an organic light emitting element having a high response speed.
The image capturing apparatus 1100 includes an optical unit (not shown). The optical unit includes a plurality of lenses and forms an image on an image sensor housed in the housing 1104. It is possible to adjust the focus by adjusting the relative positions of the plurality of lenses. This operation can be performed automatically. The display device according to the embodiment may include a color filter having red, green, and blue colors. In the color filter, the red, green, and blue colors may be arranged in a delta arrangement. The display device according to the embodiment may be used for the display unit of a mobile terminal. In this case, both a display function and an operation function may be provided. Examples of the mobile terminal include a mobile phone such as a smartphone, a tablet, and a head mounted display.
The illumination apparatus is, for example, an apparatus that illuminates a room. The illumination apparatus may emit light of white, day white, or any other color from blue to red. A light control circuit for controlling the light color may be provided. The illumination apparatus may include the organic light emitting element according to the present invention and a power supply circuit connected thereto. The power supply circuit is a circuit that converts an AC voltage into a DC voltage. Note that white light has a color temperature of 4200K, and day-white light has a color temperature of 5000K. The illumination apparatus may include a color filter. Further, the illumination apparatus according to this embodiment may include a heat dissipation portion. The heat dissipation portion releases the heat in the apparatus to the outside of the apparatus, and examples thereof include a metal having high specific heat, liquid silicon, and the like.
The automobile 1500 may include a body 1503 and windows 1502 attached thereto. The window may be a transparent display as long as it is not a window for checking the front or rear of the automobile. The light emitting device 100 can be applied to the transparent display. In this case, the components such as the electrodes included in the organic light emitting element are formed by transparent members. The moving body according to this embodiment may be a ship, an aircraft, a drone, or the like. The moving body may include a body and a lighting unit provided in the body. The lighting unit may emit light to inform the position of the body. A display panel formed by a display device represented by the display device DD2 or DD3 can be applied to the lighting unit.
Application examples of the display device according to the embodiments described above will be described with reference to
The pair of glasses 1600 further includes a control apparatus 1603. The control apparatus 1603 functions as a power supply that supplies power to the image capturing apparatus 1602 and the display device according to the embodiments. The control apparatus 1603 controls the operation of the image capturing apparatus 1602 and the operation of the display device. An optical system for focusing light to the image capturing apparatus 1602 is formed on the lens 1601.
The line of sight of the user to the displayed image is detected from the captured image of the eyeball obtained by the image capturing operation using the infrared light. A known method can be arbitrarily applied for the line-of-sight detection using the captured eyeball image. As an example, a line-of-sight detection method based on Purkinje images caused by the reflection of the emitted light on the cornea can be used.
More specifically, line-of-sight detection processing is performed based on a pupil-cornea reflection method. The line of sight of the user is detected by using the pupil-cornea reflection method to calculate a line-of-sight vector representing the direction (rotation angle) of the eyeball based on the image of the pupil and the Purkinje images included in the captured image of the eyeball.
A display apparatus according to one embodiment of the present invention may include an image capturing apparatus including a light receiving element, and control a displayed image on the display apparatus based on the line-of-sight information of the user obtained from the image capturing apparatus.
More specifically, in the display apparatus, a first field-of-view region which is gazed upon by the user and a second field-of-view region other than the first field-of-view region are determined based on the line-of-sight information. The first field-of-view region and the second field-of-view region may be determined by a control apparatus of the display apparatus. Alternatively, the first field-of-view region and the second field-of-view region may be determined by an external control apparatus and the display apparatus may receive information corresponding to this determination. Control can be performed in the display region of the display apparatus so that the display resolution of the first field-of-view region will be higher than the display resolution of the second field-of-view region. That is, the resolution of the second field-of-view region may be lowered more than the resolution of the first field-of-view region.
In addition, the display region includes a first display region and a second display region different from the first display region, and a region with a high degree of priority is determined from the first display region and the second display region of the display region based on the line-of-sight information. The first field-of-view region and the second field-of-view region may be determined by the control apparatus of the display apparatus. Alternatively, the first field-of-view region and the second field-of-view region may be determined by an external control apparatus and the display apparatus may receive information corresponding to this determination. Control may be performed so that the resolution of a region with the high degree of priority will be set higher than the resolution of a region other than the region with the high degree of priority. That is, the resolution of a region with a relatively low degree of priority can be set low.
Note that an AI may be used for the determination of the first field-of-view region and the region with the high degree of priority. The AI may be a model configured to estimate, from an image of the eyeball, the angle of the line of sight and the distance to an object as the target of the gaze by using the image of the eyeball and the direction actually gazed upon by the eyeball of the image as the teaching data. The display apparatus, the image capturing apparatus, or an external apparatus may include the AI program. If the AI program is included in an external apparatus, information determined by the AI program will be transmitted to the display apparatus by communication.
In a case in which display control is to be performed based on visual recognition detection, the display device according to the embodiments can be preferably applied to a pair of smart glasses that further includes an image capturing apparatus configured to capture the outside. The smart glasses can display the captured external information in real time.
As has been described above, by using the device using the organic light emitting element according to this embodiment, it is possible to perform stable display even for a long-time display with good image quality.
While the present invention has been described with reference to exemplary embodiments, it is to be understood that the invention is not limited to the disclosed exemplary embodiments. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all such modifications and equivalent structures and functions.
Shikina, Noriyuki, Matsuyama, Kazunori
Patent | Priority | Assignee | Title |
Patent | Priority | Assignee | Title |
10049628, | Dec 01 2014 | Samsung Display Co., Ltd. | Display device and driving method thereof |
5748164, | Dec 22 1994 | CITIZEN FINETECH MIYOTA CO , LTD | Active matrix liquid crystal image generator |
6243055, | Oct 25 1994 | Fergason Patent Properties LLC | Optical display system and method with optical shifting of pixel position including conversion of pixel layout to form delta to stripe pattern by time base multiplexing |
6538664, | Apr 19 2000 | Koninklijke Philips Electronics N.V. | Matrix display device with improved image sharpness |
6952194, | Mar 31 1999 | SEMICONDUCTOR ENERGY LABORATORY CO , LTD | Liquid crystal display device |
7453478, | Jul 29 2004 | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | Address generation in a light modulator |
7605899, | Dec 05 2003 | Canon Kabushiki Kaisha | Electrophoretic dispersion liquid and electrophoretic display device |
7812812, | Mar 25 2003 | Canon Kabushiki Kaisha | Driving method of display apparatus |
8203511, | Mar 09 2007 | Canon Kabushiki Kaisha | Display apparatus |
9354494, | Feb 04 2014 | PANASONIC INTELLECTUAL PROPERTY MANAGEMENT CO., LTD. | Projection type image display apparatus and adjusting method |
20010038374, | |||
20100295844, | |||
20130120477, | |||
20160155405, | |||
20170004783, | |||
20200058690, | |||
CN1216135, | |||
CN1866340, | |||
JP2003532145, | |||
JP2010271365, | |||
JP2016110115, | |||
KR1020160066131, | |||
KR20020025897, | |||
KR20020080248, | |||
WO182280, | |||
WO2015087598, | |||
WO182280, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Mar 16 2023 | Canon Kabushiki Kaisha | (assignment on the face of the patent) | / | |||
Mar 30 2023 | MATSUYAMA, KAZUNORI | Canon Kabushiki Kaisha | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 063587 | /0248 | |
Mar 30 2023 | SHIKINA, NORIYUKI | Canon Kabushiki Kaisha | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 063587 | /0248 |
Date | Maintenance Fee Events |
Mar 16 2023 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Date | Maintenance Schedule |
Apr 30 2027 | 4 years fee payment window open |
Oct 30 2027 | 6 months grace period start (w surcharge) |
Apr 30 2028 | patent expiry (for year 4) |
Apr 30 2030 | 2 years to revive unintentionally abandoned end. (for year 4) |
Apr 30 2031 | 8 years fee payment window open |
Oct 30 2031 | 6 months grace period start (w surcharge) |
Apr 30 2032 | patent expiry (for year 8) |
Apr 30 2034 | 2 years to revive unintentionally abandoned end. (for year 8) |
Apr 30 2035 | 12 years fee payment window open |
Oct 30 2035 | 6 months grace period start (w surcharge) |
Apr 30 2036 | patent expiry (for year 12) |
Apr 30 2038 | 2 years to revive unintentionally abandoned end. (for year 12) |