Provided are a display device and a display method enabling to implement AR with excellent visibility by using a retinal direct drawing display device. The display device includes a display unit that projects a visual information image including visual information onto a retina of a user and displays the visual information in superimposition with a field of view of the user. The display device further includes a visual information control unit that controls displaying of the visual information by the display unit, in which the display unit displays the visual information that is blurred based on situation.
|
16. A display method, comprising:
projecting a visual information image including visual information onto a retina of a first user of a first display device;
displaying the visual information in superimposition with a field of view of the first user;
blurring the visual information based on a specific situation;
controlling sharing of the visual information with a second display device; and
displaying, in superimposition with a region overlapping with the field of view of the first user, the visual information displayed in a field of view of a second user of the second display device.
1. A first display device, comprising:
a projector configured to:
project a visual information image including visual information onto a retina of a first user of the first display device; and
display the visual information in superimposition with a field of view of the first user; and
a processor configured to:
control the projector to display the visual information;
blur the visual information based on a specific situation; and
control sharing of the visual information with a second display device,
wherein the projector is further configured to display, in superimposition with a region overlapping with the field of view of the first user, the visual information displayed in a field of view of a second user of the second display device.
17. A non-transitory computer-readable medium having stored thereon computer-executable instructions, which when executed by a processor, cause the processor to execute operations, the operations comprising:
controlling projection of a visual information image including visual information onto a retina of a first user of a first display device;
controlling display of the visual information in superimposition with a field of view of the first user;
blurring the visual information based on a specific situation;
controlling sharing of the visual information with a second display device; and
controlling display, in superimposition with a region overlapping with the field of view of the first user, of the visual information displayed in a field of view of a second user of the second display device.
2. The first display device according to
3. The first display device according to
control a display position of first visual information to follow a movement of the field of view of the first user, and
control a display position of second visual information to be fixed with respect to a real world.
4. The first display device according to
5. The first display device according to
6. The first display device according to
7. The first display device according to
8. The first display device according to
9. The first display device according to
the visual information includes a virtual operation tool for operating an equipment, and
the processor is further configured to control the equipment based on an operation on the virtual operation tool.
10. The first display device according to
11. The first display device according to
12. The first display device according to
the visual information includes a virtual display, and
the projector is further configured to display the virtual display at a position corresponding to an operation device in the field of view of the first user.
13. The first display device according to
the visual information includes operation support information for support of an operation on the operation device, and
the processor is further configured to control display of the operation support information based on a movement of a hand of the first user with respect to the operation device.
14. The first display device according to
15. The first display device according to
|
This application is a U.S. National Phase of International Patent Application No. PCT/JP2020/026959 filed on Jul. 10, 2020, which claims priority benefit of Japanese Patent Application No. JP 2019-137501 filed in the Japan Patent Office on Jul. 26, 2019. Each of the above-referenced applications is hereby incorporated herein by reference in its entirety.
The present technology relates to a display device, a display method, and a program, and more particularly to a retinal direct drawing display device, a display method, and a program.
Conventionally, it has been proposed to implement augmented reality (AR) by using a retinal direct drawing display device (see, for example, Patent Document 1).
In the retinal direct drawing display device, an image including visual information is projected on a retina of a user, and the image is drawn directly on the retina. As a result, the visual information is displayed in superimposition with a real world in the user's field of view. Furthermore, in the retinal direct drawing display device, even if the user's field of view moves, the visual information is always in focus. Therefore, since the visual information is focused even in a situation of originally being out of focus and blurred, the user may feel uncomfortable and visibility may be reduced.
The present technology has been made in view of such a situation, and an object thereof is to implement AR with excellent visibility by using a retinal direct drawing display device.
A display device according to one aspect of the present technology includes: a display unit configured to project a visual information image including visual information onto a retina of a user and display the visual information in superimposition with a field of view of the user; and a visual information control unit configured to control displaying of the visual information by the display unit, in which the display unit displays the visual information that is blurred in accordance with a situation.
A display method according to one aspect of the present technology includes: displaying visual information that is blurred in accordance with a situation, by using a display device configured to project a visual information image including the visual information onto a retina of a user and display the visual information in superimposition with a field of view of the user.
A program according to one aspect of the present technology causes a computer to execute processing of displaying visual information that is blurred in accordance with a situation, in which the computer is of a display device configured to project a visual information image including the visual information onto a retina of a user and display the visual information in superimposition with a field of view of the user.
In one aspect of the present technology, a visual information image including visual information is projected on a retina of a user, the visual information is displayed in superimposition with a field of view of the user, and the visual information that is blurred in accordance with a situation is displayed.
Hereinafter, an embodiment for implementing the present technology will be described. The description will be given in the following order.
1. Embodiment
2. Modification
3. Other
An embodiment of the present technology will be described with reference to
<Configuration Example of Display Device>
The display device 101 is an eyewear display device of a retinal direct drawing method. The display device 101 projects an image including visual information (hereinafter, referred to as a visual information image) onto a retina of a user in a state of being worn on the head of the user like glasses, and draws the visual information image directly on the retina of the user. As a result, in a field of view of the user, the visual information is displayed in superimposition with a real world, and the user can view the visual information in superimposition with the real world.
The display device 101 includes an input unit 111, an image capturing unit 112, a sound detection unit 113, a sensor unit 114, a control unit 115, a display unit 116, an output unit 117, a communication unit 118, and a storage unit 119. The input unit 111, the image capturing unit 112, the sound detection unit 113, the sensor unit 114, the control unit 115, the display unit 116, the output unit 117, the communication unit 118, and the storage unit 119 are mutually connected via a bus 131.
The input unit 111 includes an input device such as a button or a switch. The input unit 111 is used, for example, for operation of the display device 101 and input of various data, instructions, and the like.
The image capturing unit 112 includes, for example, an image sensor such as a CMOS sensor. The image capturing unit 112 captures an image of a range including a user's field of view. The image capturing unit 112 outputs data (hereinafter, referred to as captured image data) of an image (hereinafter, referred to as a captured image) obtained as a result of image capturing.
The sound detection unit 113 includes, for example, a plurality of microphones. The microphones are provided, for example, at least at three locations of the front, left, and right of the eyewear display device 101 so that a position of a sound source around the user can be detected. Specifically, for example, the microphones are provided at least at three locations of a central portion of a front surface of the eyewear display device 101 and left and right temples. The sound detection unit 113 detects sound around the user and outputs data (hereinafter, referred to as sound data) of the detected sound.
The sensor unit 114 includes an inertial measurement unit (IMU) including an acceleration sensor and an angular velocity sensor, a position sensor that receives a global navigation satellite system (GNSS) signal from a GNSS satellite and detects a current position, and the like. The sensor unit 114 outputs data (hereinafter, referred to as sensor data) indicating a detection result of each sensor.
The control unit 115 includes, for example, a processor including a central processing unit (CPU) or the like. The control unit 115 controls each unit of the display device 101, and performs processing such as generation of a visual information image, control of displaying of visual information, and sharing of visual information with another display device.
The display unit 116 projects a visual information image onto a retina of the user to display the visual information in superimposition within a field of view of the user. Furthermore, as described later, the display unit 116 displays visual information that is blurred in accordance with a situation. Note that a specific configuration example of the display unit 116 will be described later with reference to
The output unit 117 includes an output device such as, for example, a display, a light emitting element, and a speaker. The output unit 117 outputs information indicating a state or the like of the display device 101 by, for example, an image, light, sound, or the like. Furthermore, for example, the output unit 117 includes an output terminal for output of image data, sound data, and the like.
The communication unit 118 includes a communication device and performs wireless communication with surrounding equipment. Note that a communication method of the communication unit 118 is not particularly limited, and for example, short-range wireless communication such as Bluetooth (registered trademark) or near field communication (NFC) is adopted. Furthermore, the communication unit 118 may support a plurality of communication methods.
The storage unit 119 includes, for example, a magnetic storage device such as a read only memory (ROM), a random access memory (RAM), and a flash memory, a semiconductor storage device, an optical storage device, a magneto-optical storage device, or the like. The storage unit 119 stores various programs, data, and the like used for processing of the display device 101.
<Configuration Example of Projection System 201>
The projection system 201 includes a projection device 211 and an optical element 212. Since the projection system 201 is used in a state where a positional relationship between the optical element 212 and an eyeball 202 is fixed, a distance between the optical element 212 and a rotation center of the eyeball 202 does not change even if the eyeball 202 rotates. Therefore, it is not necessary to adjust projection light in accordance with a movement of the eyeball, and it is not necessary to provide an eye tracking device.
The projection device 211 includes the projection optical system 221. The projection optical system 221 includes a two-dimensional array display element 231, a lens 232, and a lens 233.
The two-dimensional array display element 231 includes, for example, a liquid crystal display (LCD), a liquid crystal on silicon (LCOS), or an organic light-emitting diode (OLED). For example, under control of the control unit 115, from illumination light emitted from a light source (not illustrated), the two-dimensional array display element 231 forms projection light for displaying of a visual information image, and emits the projection light. For example, on an optical path between the light source and the two-dimensional array display element 231, an imaging system and a color separation synthesis system may be provided.
The lens 232 and the lens 233 are provided between the two-dimensional array display element 231 and the optical element 212. As illustrated in
As illustrated in
Note that a main light beam of projection light incident on the optical element 212 preferably has a direction in which the projection light can be collected near the pupil after being incident on the optical element 212, and more preferably is substantially parallel to an optical axis. That is, the main light beam of the projection light preferably reaches the optical element 212 as a telecentric light beam. Therefore, it is preferable that the projection optical system 221 is configured such that the projection light having the main light beam substantially parallel to the optical axis is to be incident on the optical element 212.
As illustrated in
Furthermore, the projection optical system 221 is configured such that the projection light is collected near the pupil and reaches the retina 242. That is, the projection light may be projected onto the retina 242 by so-called Maxwellian view. For example, as illustrated in
In the Maxwellian view optical system, since one dot (a minimum display unit) in an image passes through one point on a crystalline lens 241, the one-dot image on the retina is less susceptible to a state of the crystalline lens 241. Therefore, even a user having, for example, myopia, hyperopia, astigmatism, or the like can clearly recognize the image. Furthermore, a virtual image that appears to float in a space is focus-free, and the virtual image comes into focus at any distance from the eye.
The optical element 212 causes the projection light to be collected near the pupil and to reach the retina 242. The optical element 212 is used in a state where a positional relationship with the eyeball 202 is fixed. Preferably, as illustrated in
The optical element 212 includes, for example, a contact lens-shaped optical element, and preferably includes a contact lens-shaped holographic optical element. Since the optical element 212 includes the contact lens-shaped optical element, a field of view in which an image by projection light can be recognized can be enlarged. Furthermore, since the optical element 212 includes the contact-lens-shaped optical element, it is possible to easily enlarge an eye box (that is, a spatial region around the eyeball, in which an image by projection light can be recognized).
In this way, the projection system 201 projects projection light representing a visual information image onto the retina 242, so that a virtual image of visual information appears to float in a space. That is, the virtual visual information is displayed in superimposition with a real world in the user's field of view.
Note that the projection system 201 may be individually provided for both eyes of the user, or may be provided only for one eye. Note that, in the former case, stereoscopic vision can be used, and the user can feel perspective of visual information more naturally.
<Configuration Example of Information Processing Unit 301>
The information processing unit 301 includes a space recognition unit 311, a sound recognition unit 312, an operation recognition unit 313, a visual information control unit 314, an output control unit 315, an equipment control unit 316, and a communication control unit 317.
The space recognition unit 311 performs various types of recognition processing on a space including a field of view of the user who uses the display device 101, that is, a space on which visual information is superimposed by the display device 101, on the basis of captured image data supplied from the image capturing unit 112, sound data supplied from the sound detection unit 113, and sensor data supplied from the sensor unit 114. For example, the space recognition unit 311 performs recognition processing of a type, a position, an orientation, a size, a shape, a feature point, and the like of an object in the space. Furthermore, for example, the space recognition unit 311 performs recognition processing of a range and a movement of the user's field of view, a direction and a movement of a line-of-sight, and the like.
The sound recognition unit 312 performs sound recognition processing on sound data supplied from the sound detection unit 113. For example, the sound recognition unit 312 performs recognition processing of a position of a sound source around the user, a state of sound emitted from the sound source (for example, a type, a volume, a pitch, a tone, sound quality, and the like), an utterance content, and the like.
The operation recognition unit 313 recognizes a user's operation on visual information on the basis of recognition results by the space recognition unit 311 and the sound recognition unit 312. For example, the operation recognition unit 313 recognizes an operation on visual information on the basis of a motion of a hand of the user recognized by the space recognition unit 311. Furthermore, for example, the operation recognition unit 313 recognizes an operation on visual information on the basis of user's voice recognized by the sound recognition unit 312.
The visual information control unit 314 performs generation of visual information, control of displaying of visual information by the display unit 116, and the like. Note that, as described later, the visual information control unit 314 blurs visual information in accordance with a situation. The visual information control unit 314 includes an image processing unit 321, a display control unit 322, and a sharing control unit 323.
The image processing unit 321 generates a visual information image including visual information. Note that, as described later, the image processing unit 321 generates a visual information image including visual information that is blurred in accordance with a situation.
The display control unit 322 supplies data of the visual information image (hereinafter, referred to as visual information image data) to the display unit 116, and controls the display unit 116 to control a position, a timing, and the like at which the visual information is displayed.
The sharing control unit 323 controls a process of sharing visual information with another display device. By performing the process of sharing visual information with another display device, for example, among visual information displayed on another display device, visual information virtually existing in a field of view of a user of the display device 101 is displayed on the display device 101. Furthermore, conversely, among visual information displayed on the display device 101, visual information virtually existing in a field of view of a user of another display device is displayed on the another display device.
The output control unit 315 controls output of various types of information and data by the output unit 117.
The equipment control unit 316 controls equipment (for example, an electrical appliance or the like) around the user. For example, the equipment control unit 316 transmits a control signal to the equipment around the user in accordance with a user operation on visual information, and controls the equipment.
The communication control unit 317 controls communication by the communication unit 118.
<Processing of Display Device 101>
Next, processing of the display device 101 will be described with reference to
<Window Display Control Process>
First, a window display control process executed by the display device 101 will be described with reference to a flowchart of
This process is started when power of the display device 101 is turned on, and is ended when the power is turned off, for example.
In step S1, the display device 101 determines whether or not displaying of a window has been instructed.
For example, the space recognition unit 311 performs a process of detecting a hand in a captured image supplied from the image capturing unit 112. In a case where a hand is detected, the space recognition unit 311 further performs a process of detecting a shape of the hand.
The operation recognition unit 313 performs a process of recognizing a gesture on the basis of a detection result of the shape of the hand. Then, the operation recognition unit 313 recognizes the gesture. Further, in a case where the recognized gesture coincides with a gesture instructing to display the window, it is determined that displaying of the window has been instructed, and the process proceeds to step S2.
For example, as illustrated in
In step S2, the display device 101 displays the window.
Specifically, the image processing unit 321 generates a visual information image including the window, which is one of visual information.
The display control unit 322 supplies visual information image data to the display unit 116, and controls the display unit 116 to display the window in accordance with a position where the gesture is performed, in the user's field of view.
As a result, for example, as illustrated in
Thereafter, the process proceeds to step S3.
Whereas, in step S1, in a case where the operation recognition unit 313 does not recognize a gesture instructing to display the window, it is determined that displaying of the window is not instructed, the process of step S2 is skipped, and the process proceeds to step S3.
In step S3, the space recognition unit 311 determines whether or not the user's field of view or line-of-sight has moved. For example, on the basis of captured image data supplied from the image capturing unit 112 and sensor data of each speed sensor and acceleration sensor included in sensor data supplied from the sensor unit 114, the space recognition unit 311 determines whether or not the user's field of view or line-of-sight has moved. In a case where it is determined that the user's field of view or line-of-sight has moved, the process proceeds to step S4.
In step S4, the display device 101 controls displaying of the visual information in accordance with the movement of the user's field of view or line-of-sight. Specifically, visual information such as a window is displayed in the user's field of view by a process similar to step S2. At this time, the image processing unit 321 and the display control unit 322 generate a visual information image and control a display position of each piece of visual information, in consideration of visual information following the movement of the user's field of view and visual information not following the movement of the user's field of view. Furthermore, the image processing unit 321 blurs the visual information as necessary.
In this example, the window 402 is displayed in superimposition at the same position in a real space without following the movement of the user's field of view. Note that the display position of the window 402 is set on the basis of, for example, a feature point in the real space.
Whereas, the dinosaur 403 existing in the window 402 in the example of
Therefore, the user sees the dinosaur 403 fixed at a predetermined position in the field of view (following the movement of the field of view), and the window 402 fixed at a predetermined position in the real space (fixed with respect to the real space). Furthermore, the user sees the dinosaur 403 as if popping out of the window 402. Moreover, the user sees the dinosaur 403 moving from near (for example, the palm 401) to far (for example, a background (not illustrated) of the dinosaur 403 in
Here, if both the dinosaur 403 located far away and the window 402 located near are in focus even though the dinosaur 403 has moved from near to far, the perspective of the user is disturbed, and the user may feel uncomfortable. Therefore, for example, the image processing unit 321 intentionally blurs the window 402. As a result, the user feels the movement of the dinosaur 403 more naturally.
Note that, when the user returns the field of view to the state of
Furthermore, for example, in the state of
Note that, in a case where the line-of-sight of the user is not directed to any visual information, all the visual information may be blurred. As a result, for example, it is possible to prevent visual information from being constantly in focus within the user's field of view, and it is possible to reduce fatigue of the user's eyes.
Furthermore, for example, a content (a content of visual information) displayed in the window 402 may be changed on the basis of a movement of at least one of the head or the line-of-sight of the user.
For example, in a case where details of the content is represented in a hierarchical structure, a content of a lower hierarchy may displayed in the window 402 in a case where the head or the line-of-sight of the user moves downward, while a content of an upper hierarchy may be displayed in the window 402 in a case where the head or the line-of-sight of the user moves upward.
Moreover, for example, the content in the window 402 may be scrolled in a direction in which the head or the line-of-sight of the user moves.
Thereafter, the process proceeds to step S5.
Whereas, in a case where it is determined in step S3 that the user's field of view and line-of-sight have not moved, the process of step S4 is skipped, and the process proceeds to step S5.
In step S5, the display device 101 determines whether or not a movement of the window has been instructed. For example, in a case where a gesture of the user is recognized by a process similar to step S1, and the recognized gesture coincides with a gesture instructing to move the window, it is determined that a movement of the window has been instructed, and the process proceeds to step S6.
In step S6, the display device 101 moves a position of the window. That is, the display unit 116 stops displaying the window under the control of the visual information control unit 314. As a result, for example, in the example of
Thereafter, the process proceeds to step S7.
Whereas, in a case where it is determined in step S5 that a movement of the window has not been instructed, the process of step S6 is skipped and the process proceeds to step S7.
In step S7, the display device 101 determines whether or not deletion of the window has been instructed. For example, in a case where a gesture of the user is recognized by a process similar to step S1 and the recognized gesture coincides with a gesture instructing to delete the window, it is determined that deletion of the window has been instructed, and the process proceeds to step S8.
In step S8, the display device 101 deletes the window. That is, the display unit 116 stops displaying the window under the control of the visual information control unit 314. As a result, the window is deleted from the user's field of view.
Thereafter, the process returns to step S1, and the processes in and after step S1 are executed.
Whereas, in a case where it is determined in step S7 that deletion of the window has not been instructed, the process returns to step S1, and the processes in and after step S1 are executed.
As described above, the user can display the window at a desired position or move the window, by a simple hand movement. Furthermore, by causing a part of visual information to follow a movement of the user's field of view, visibility of the visual information is improved. Moreover, by focusing on visual information in a line-of-sight direction of the user and blurring other visual information, it is possible to implement natural AR with excellent visibility without disturbing perspective of the user.
<Auditory Perception Support Process>
Next, an auditory perception support process executed by the display device 101 will be described with reference to a flowchart in
The auditory perception support process is, for example, a process of supporting perception of auditory information by using visual information for a user having auditory abnormalities.
This process is started when power of the display device 101 is turned on, and is ended when the power is turned off, for example.
In step S101, the sound recognition unit 312 determines whether or not sound has been detected on the basis of sound data supplied from the sound detection unit 113. This determination process is repeatedly executed until it is determined that sound has been detected, and the process proceeds to step S102 in a case where it is determined that sound has been detected.
In step S102, the sound recognition unit 312 performs sound recognition. Specifically, the sound recognition unit 312 performs a sound recognition process on sound data supplied from the sound detection unit 113, and recognizes a relative position of a sound source, from which the sound is emitted, with respect to the user. Furthermore, the sound recognition unit 312 recognizes a state such as a type, a volume, a pitch, a tone, and sound quality of the sound. Furthermore, in a case where the type of sound is person's voice, the sound recognition unit 312 recognizes utterance contents.
In step S103, the display device 101 displays auditory perception support information. Specifically, the image processing unit 321 generates a visual information image including auditory perception support information that is visual information regarding sound emitted from a sound source around the user. The display control unit 322 supplies the visual information image data to the display unit 116, and controls the projection system 201 to display the auditory perception support information in accordance with a position of the sound source, in the field of view of the user.
Thereafter, the process returns to step S101, and the processes in and after step S101 are executed.
In this example, a person 501 and a person 502 in the user's field of view are talking. Note that only a part of an arm of the person 502 is included in the field of view of the user, and the face is not shown. Furthermore, sound state information 511, which is visual information indicating a recognized state of voice of the person 501, is displayed in superimposition with the person 501. Moreover, utterance information 512, which is visual information indicating utterance contents of the person 501, is displayed below the sound state information 511. Furthermore, sound state information 513 indicating a recognized state of voice of the person 502 is displayed immediately to the left of the person 502. Moreover, utterance information 514 indicating utterance contents of the person 502 is displayed below the sound state information 513.
The sound state information 511 and the sound state information 513 visually express the sound state by a waveform shape, size, number, color, line thickness, and the like.
Note that, as the sound including many pieces of sound, for example, cases are assumed where many people are talking at the same place, where sound of many instruments is included, and the like.
Furthermore, in a case where the sound is person's voice, for example, the sound state information may represent a feeling recognized from the person's voice. For example, a color of a curve representing the sound state information is set to green for gentle voice, the color of the curve representing the sound state information is set to blue for sad voice, and the color of the curve representing the sound state information is set to red for angry voice.
As described above, for example, a user having auditory abnormalities can perceive more detailed and accurate auditory information. For example, the user can recognize utterance contents of the person 501 and the person 502 from the utterance information 512 and the utterance information 514. Furthermore, the user can recognize a position of a sound source of each piece of voice, that is, a position of a subject of each conversation, on the basis of positions of the sound state information 511 and the utterance information 512, and the sound state information 513 and the utterance information 514. Moreover, by using the sound state information 511 and the sound state information 513, the user can recognize a state of voice that cannot be recognized only with the utterance information 512 and the utterance information 514.
Note that a display position of the sound state information does not necessarily need to be accurately matched with the position of the sound source, and may be matched to an extent to allow a direction of the sound source to be roughly known. Furthermore, the display position of the sound state information does not need to correspond to a position of the sound source in an up-down direction but may correspond only to a position in a left-right direction.
<Virtual Display Displaying Process>
Next, a virtual display displaying process executed by the display device 101 will be described with reference to a flowchart of
This process is started when power of the display device 101 is turned on, and is ended when the power is turned off, for example.
In step S201, the space recognition unit 311 determines whether or not a keyboard has been detected. This determination process is repeatedly executed until the space recognition unit 311 determines that the keyboard has been detected, and the process proceeds to step S202 in a case where the space recognition unit 311 determines that the keyboard has been detected.
In step S202, the display device 101 displays a virtual display in accordance with a position of the keyboard.
Specifically, the image processing unit 321 generates a visual information image including a virtual display, which is one of visual information. The display control unit 322 supplies visual information image data to the display unit 116, and controls the display unit 116 to display the virtual display in accordance with the position of the keyboard, in the field of view of the user.
In this example, a keyboard 602 is placed on a desk 601 in a real space in the user's field of view. No information is printed on a key top of each key of the keyboard 602.
The display control unit 322 sets a display position of the virtual display 603 on the basis of feature points (for example, four corners of the keyboard 602 and the like) of the keyboard 602 recognized by the space recognition unit 311. For example, the display position of the virtual display 603 is set to a position at which the user can easily view while using the keyboard 602. Then, the display control unit 322 controls the display unit 116 to display the virtual display 603 at the set display position in the field of view of the user.
Note that, for example, the image processing unit 321 may generate a visual information image in which a periphery of the virtual display 603 is blurred. As a result, in the user's field of view, the vicinity of a boundary of the virtual display 603 with a real world is blurred. Then, the user receives an impression that the virtual display 603 is spatially wider than the actual size, and is prevented from receiving an illusion that information has disappeared near the boundary of the virtual display 603. As a result, visibility and workability of the user are improved.
In step S203, the display control unit 322 determines whether or not displaying of a key top is set to a variable mode. In a case where it is determined that displaying of a key top is set to the variable mode, the process proceeds to step S204.
Here, the variable mode is a mode in which information displayed on a key top of each key of the keyboard changes in accordance with a situation.
In step S204, the space recognition unit 311 determines whether or not a finger has approached a key top. In a case where it is determined that the finger has approached a key top, the process proceeds to step S205.
In step S205, the display device 101 displays the key top.
Specifically, the image processing unit 321 generates a visual information image including visual information to be displayed on the key top, as operation support information for support of the operation of the keyboard. The display control unit 322 supplies the visual information image data to the display unit 116, and controls the display unit 116 to display the visual information on a key top of each key of the keyboard in the field of view of the user.
At this time, the visual information displayed on the key top changes in accordance with a situation. For example, the visual information displayed on the key top changes in accordance with information displayed on the virtual display or a function being operated. For example, a character or a symbol assigned to each key, a name or an icon of a function assigned to each key, or the like is displayed on the key top of each key in the field of view of the user.
Furthermore, it is not always necessary to display visual information on all the keys of the keyboard. For example, the visual information may be displayed only on a key top of a key to which the user's finger approaches and keys around the key.
Thereafter, the process proceeds to step S207.
Whereas, in a case where it is determined in step S204 that a finger has not approached a key top, the process of step S205 is skipped, and the process proceeds to step S207.
Furthermore, in a case where it is determined in step S203 that displaying of a key top is set to a fixed mode, the process proceeds to step S206.
Here, the fixed mode is a mode in which information displayed on a key top of each key of the keyboard is fixed without changing in accordance with a situation.
In step S206, the key top is displayed similarly to the process of step S205.
At this time, the visual information displayed on the key top of each key is fixed. For example, a character or a symbol printed on a key top of each key of a commercially available keyboard is displayed on the key top of each key in the field of view of the user.
Thereafter, the process proceeds to step S207.
In step S207, the space recognition unit 311 determines whether or not a key top is hidden by a hand. In a case where there is a key whose key top is invisible by the user's hand among the keys of the keyboard in the user's field of view, the space recognition unit 311 determines that a key top is hidden by the hand, and the process proceeds to step S208.
In step S208, the display device 101 displays visual information of the hidden portion of the key top in superimposition with the hand.
Specifically, the image processing unit 321 generates a visual information image including visual information for a key for which visual information needs to be displayed, among keys whose key tops are invisible by the user's hand in the user's field of view. The display control unit 322 supplies visual information image data to the display unit 116, and controls the display unit 116 to display the visual information for the key whose key top is hidden, to be superimposed on the hand of the user in the field of view of the user.
Note that, for example, visual information for a key pressed by the user may be displayed separately from visual information for other keys. For example, in a case a key with a character “4” illustrated in
Thereafter, the process returns to step S201, and the processes in and after step S201 are executed.
Whereas, in a case where it is determined in step S207 that the key top is not hidden by the hand, the process returns to step S201, and the processes in and after step S201 are executed.
In this way, for example, by simply placing the keyboard 602 in the field of view, the virtual display 603 is displayed at an appropriate position in the field of view, and the user can perform work or the like by using the keyboard 602 and the virtual display 603.
Furthermore, the virtual display 603 does not occupy a physical space. Therefore, for example, the user can carry the keyboard 602 and easily perform work at any place as long as there is a space where the keyboard 602 can be placed. Moreover, since the virtual display 603 is visible only to the user and is invisible to others, for example, the user can perform highly confidential work even in a public place.
<Visual Information Sharing Process>
Next, a visual information sharing process executed by the display device 101 will be described with reference to a flowchart in
This process is started when power of the display device 101 is turned on, and is ended when the power is turned off, for example.
In step S301, the display device 101 shares information regarding the user's field of view with a surrounding display device. Specifically, the communication unit 118 transmits information regarding the user's field of view to the surrounding display device under control of the communication control unit 317. Furthermore, under the control of the communication control unit 317, the communication unit 118 receives information regarding a field of view of a user of the surrounding display device, from the surrounding display device.
The information regarding the user's field of view of includes, for example, captured image data of a range including the user's field of view of and sensor data indicating an acceleration, an angular velocity, and a current position of the display device.
Note that, hereinafter, in a case where the user of the display device 101 and the user of the surrounding display device are distinguished, the former is referred to as an own user, and the latter is referred to as another user.
In step S302, the sharing control unit 323 calculates a positional relationship of fields of view of individual users. Specifically, the sharing control unit 323 calculates a range of a field of view of the own user in a real world on the basis of a captured image, an acceleration, an angular velocity, and a current position of the display device 101. Furthermore, the sharing control unit 323 calculates a range of a field of view of another user in the real world on the basis of a captured image, an acceleration, an angular velocity, and a current position of the surrounding display device. Moreover, on the basis of the range of the field of view of the own user and the range of the field of view of another user, the sharing control unit 323 calculates a shared region in which fields of view overlap between the users.
Therefore, a shared region Rall in which the region Ra, the region Rb, and the region Rc overlap is a region in which the fields of view of the users A to C overlap. A shared region Rab in which only the region Ra and the region Rb overlap is a region in which the fields of view of the user A and the user B overlap. A shared region Rbc in which only the region Rb and the region Rc overlap is a region in which the fields of view of the user B and the user C overlap. A shared region Rca in which only the region Rc and the region Ra overlap is a region in which the fields of view of the user C and the user A overlap.
In step S303, the display device 101 shares visual information with a surrounding display device. Specifically, the sharing control unit 323 generates visual information shared data including image data for displaying visual information existing within a field of view of another user among visual information displayed by the display device 101, and including a display position of the visual information in a real world. The display position of the visual information in the real world indicates a position where the visual information is displayed in superimposition in the real world in the field of view of the own user. The communication unit 118 transmits the visual information shared data to the surrounding display device under the control of the communication control unit 317.
Furthermore, the communication unit 118 receives similar visual information shared data from the surrounding display device under the control of the communication control unit 317.
For example, in a case of the example of
Moreover, among visual information virtually displayed in the region Rb by the display device B, the display device 101 receives, from the display device B, visual information shared data related to visual information virtually displayed in the shared region Rall and the shared region Rab. Furthermore, among visual information virtually displayed in the region Rc by the display device C, the display device 101 receives, from the display device C, visual information shared data related to visual information virtually displayed in the shared region Rall and the shared region Rca.
Note that, hereinafter, in a case where visual information displayed by the display device 101 and visual information displayed by the surrounding display device are distinguished, the former is referred to as original visual information, and the latter is referred to as surrounding visual information.
In step S304, the display device 101 displays visual information.
Specifically, the image processing unit 321 generates a visual information image including the original visual information and the surrounding visual information. The display control unit 322 supplies the visual information image data to the display unit 116, and controls the display unit 116 to display the original visual information and the surrounding visual information in the user's field of view.
Thereafter, the process returns to step S301, and the processes in and after step S301 are executed.
Then, the building group 705 in a shared region R is visual information shared between the own user and the user 703. That is, the building group 705 is viewed from different directions by the own user and the user 703 via the respective display devices.
As a result, for example, the own user and the user 703 can have a conversation or the like while viewing the virtual building group 705 that does not exist in a real world.
Whereas, only the own user can see the window 706. Therefore, the own user can perform various kinds of work by using the window 706 without being seen by surrounding users.
Note that, the window 706 is shifted to the left from the line-of-sight direction of the own user, and thus may be blurred.
Here, in order to clearly recognize the shared region, for example, it is desirable to individually display visual information indicating the shared region R in
Furthermore, for example, as illustrated in
For example,
Note that patterns (for example, a color, a design, and the like) of the shadow 722a and the shadow 722b are distinguished so that the corresponding users can be identified. Furthermore, in order to enable unique identification of each user in a case where a plurality of pieces of visual information is shared, shadow patterns are distinguished for each user, and shadow patterns for the same user are unified.
Moreover, for example, as illustrated in
For example,
In this example, glitter 733, which is visual information for identifying an own user who is the owner of the visual information 731, is displayed around the visual information 731. Furthermore, glitter 734 for identifying another user who is the owner of the visual information 732 is displayed around the visual information 732.
Note that the glitter 733 and the glitter 734 are distinguished in pattern so that the owner can be identified. Furthermore, in order to enable unique identification of the owner of each piece of visual information, the glitter patterns are distinguished for each user, and the glitter patterns of the same owner are unified.
As described above, the visual information can be shared between the users.
Note that each user can individually move a position of the original visual information by a predetermined operation. Whereas, each user cannot move a position of the shared visual information. Therefore, each user can supply the original visual information with another user or cancel sharing with another user, by moving the position of the original visual information. Furthermore, each user can select a user with which the original visual information is shared, by moving the position of the original visual information.
<Equipment Control Process>
Next, an equipment control process executed by the display device 101 will be described with reference to a flowchart in
This process is started when power of the display device 101 is turned on, and is ended when the power is turned off, for example.
In step S401, the display device 101 determines whether or not displaying of an operation tool has been instructed. For example, in a case where a gesture of the user is recognized by a process similar to step S1 in
In step S402, the display device 101 displays the operation tool.
Specifically, the image processing unit 321 generates a visual information image including the operation tool, which is one of visual information. The display control unit 322 supplies the visual information image data to the display unit 116, and controls the display unit 116 to display a virtual operation tool in the field of view of the user.
Note that the operation tool is displayed, for example, in the field of view of the user so as to overlap with one piece of equipment among pieces of equipment that can be operated by the operation tool or in the vicinity of the one piece of equipment. Then, the equipment becomes an operation target by the operation tool.
Note that, hereinafter, equipment to be operated by the operation tool is referred to as operation target equipment.
Alternatively, the operation tool is displayed at a position designated by the user in the field of view of the user.
In this example, an air conditioner 801 and a TV 802 are present as equipment that can be operated by the operation tool in the user's field of view. Then, when the user stands still for a while with a palm 803 facing upward, an operation tool 804 is displayed on the right side of the TV 802.
The operation tool 804 is semitransparent and circular. This example shows an example in which an operation target of the operation tool 804 is set to the TV 802, and power supply and a volume of the TV 802 can be operated by the operation tool 804. Furthermore, visual information 805 indicating that the operation target equipment is the TV 802 and the volume is set to 30 is displayed on the right side of the field of view.
Thereafter, the process proceeds to step S403.
Whereas, in a case where it is determined in step S401 that displaying of the operation tool has not been instructed, the process of step S402 is skipped and the process proceeds to step S403.
In step S403, the operation recognition unit 313 determines whether or not the operation tool has been operated. For example, in a case where a gesture of the user is recognized by a process similar to step S1 in
In step S404, the display device 101 performs processing corresponding to an operation content.
For example, the image processing unit 321 generates a visual information image in which displaying of the operation tool is updated, in accordance with the operation content. The display control unit 322 supplies the visual information image data to the display unit 116, and controls the display unit 116 to display the operation tool at a position corresponding to the operation content in the field of view of the user.
Furthermore, in a case where an operation is performed on the operation target equipment by using the operation tool, the equipment control unit 316 generates a control signal corresponding to the operation content. The communication unit 118 transmits the control signal to the operation target equipment under the control of the communication control unit 317.
Whereas, the operation target equipment performs processing on the basis of the control signal. As a result, the user can operate the operation target equipment by using the operation tool.
Moreover, in a case where the operation tool is superimposed on or brought close to equipment different from the operation target equipment, the equipment control unit 316 changes the operation target equipment to the equipment.
Thereafter, the process proceeds to step S405.
Whereas, in a case where it is determined in step S403 that the operation tool has not been operated, the process of step S404 is skipped and the process proceeds to step S405.
In step S405, the display device 101 determines whether or not deletion of the operation tool has been instructed. For example, in a case where a gesture of the user is recognized by a process similar to step S1 in
In step S406, the display device 101 deletes the operation tool. That is, the display unit 116 stops displaying the operation tool under the control of the visual information control unit 314. As a result, the operation tool is deleted from the user's field of view.
Thereafter, the process returns to step S401, and the processes in and after step S401 are executed.
Whereas, in a case where it is determined in step S405 that deletion of the operation tool has not been instructed, the process returns to step S401, and the processes in and after step S401 are executed.
As described above, the user can operate various kinds of equipment by operating the operation tool virtually displayed in the field of view.
Furthermore, in a conventional method in which a user visually recognizes visual information displayed on a display in front of both eyes of the user, in order to cause an operation tool to be displayed near the user, it is necessary to display the operation tool at a position close to an inside (on a side between both eyes) of each display. However, when the operation tool is too close to the inside of each display, there is a possibility to cause crossed eyes of the user. Therefore, it is necessary to separate the user and the display position of the operation tool by a predetermined distance or more. Therefore, the user needs to operate the operation tool separated by the predetermined distance or more, and operability is deteriorated.
Whereas, in the retinal direct drawing method, since there is no possibility of crossed eyes, a distance between the user and the display position of the operation tool is not particularly limited. Therefore, the operation tool can be displayed near the user, and operability is improved.
<Visual Information Remote Display Process>
Next, a visual information remote display process executed by the display device 101 will be described with reference to a flowchart in
The visual information remote display process is a process of externally controlling visual information displayed on the display device 101.
In step S501, the display device 101 transmits an image of a user's field of view. Specifically, the communication unit 118 transmits captured image data supplied from the image capturing unit 112 to an external device under the control of the communication control unit 317.
The external device is not particularly limited. For example, the external device includes a personal computer (PC), a tablet terminal, a smartphone, a mobile phone, a server, and the like. Furthermore, the display device 101 may directly communicate with an external device or may communicate via a network.
In step S502, the display device 101 receives visual information to be superimposed with the user's field of view.
For example, another user (hereinafter, referred to as a remote user) who uses the external device generates visual information to be displayed in a field of view of the user of the display device 101 by using the external device, and specifies a display position of the visual information in a captured image acquired from the display device 101. Then, the external device generates remote visual information data including image data for displaying the visual information and including the display position of the visual information in the captured image, and transmits to the display device 101.
Whereas, the communication unit 118 of the display device 101 receives the remote visual information data under the control of the communication control unit 317.
In step S503, the display device 101 displays the received visual information. Specifically, the image processing unit 321 generates a visual information image including visual information received from the external device. The display control unit 322 supplies the visual information image data to the display unit 116, and controls the display unit 116 to display the visual information received from the external device in the user's field of view. At this time, the display control unit 322 controls a display position of the visual information so as to substantially match the position specified on the captured image, in the field of view of the user.
Thereafter, the visual information remote display process ends.
In this example, there is a town in the user's field of view. Then, a map 901, which is one of visual information, is displayed at an upper left corner in the user's field of view. In the map 901, a pointer 911 indicating a destination is displayed. Furthermore, a pointer 902, which is visual information indicating a destination, is displayed on a central building in the user's field of view.
For example, according to the above-described process, the remote user can control to display the map 901 and the pointer 902 in the field of view of the user of the display device 101 from the external device, indicate destination to the user, and guide the user to the destination.
Note that without limiting to this example, for example, by remotely controlling visual information in the user's field of view, the remote user can give an instruction or provide information to the user from a distance.
Hereinafter, a modification of the above-described embodiment of the present technology will be described.
<Modification Regarding Configuration of System>
In the above description, an example has been described in which the display device 101 independently generates and displays visual information, but for example, the processing by the display device 101 may be shared by a plurality of devices.
The display system 1001 includes a display device 1011, an information processing apparatus 1012, and a server 1013. The information processing apparatus 1012 and the server 1013 are connected to each other via a network 1021, and can communicate with each other. Furthermore, the display device 1011 and the information processing apparatus 1012 can directly communicate with each other.
The display device 1011 has, for example, at least a function similar to that of the display unit 116 of the display device 101 in
Note that the information processing apparatus 1012 or the server 1013 may include functions other than the function of generating the visual information image among the functions provided to the display device 101.
Furthermore, a type of the information processing apparatus 1012 is not particularly limited. For example, the information processing apparatus 1012 includes a PC, a tablet terminal, a smartphone, a mobile phone, or the like.
<Modification Regarding Projection System>
The configuration example of the projection system 201 described above with reference to
<First Modification of Projection System>
The projection system 1101 includes a projection device 1111 and an optical element 212. The projection device 1111 includes a projection optical system 1121.
The projection optical system 1121 includes a light source 1131, a scanning mirror 1132, and a lens 1133.
The light source 1131 emits a light beam toward the scanning mirror 1132. As the light source 1131, for example, a light emitting diode (LED), a laser diode (LD), or the like is used.
The scanning mirror 1132 may two-dimensionally scan the laser beam emitted from the light source 1131 to cause the laser beam to reach the optical element 212. For example, a micro electro mechanical systems (MEMS) mirror is used as the scanning mirror 1132. The scanning mirror 1132 may move a direction of the laser beam at a high speed so that an image is formed on a retina 242.
The lens 1133 is provided between the light source 1131 and the optical element 212. A laser beam (projection light) projected from the light source 1131 is refracted by the lens 1133 and incident on the optical element 212.
<Second Modification of Projection System>
The projection system 1201 differs from the projection system 201 in that an optical element 1211 is provided instead of the optical element 212.
The optical element 1211 is used in a state of not being in contact with a surface of an eyeball.
<Third Modification of Projection System>
The projection system 1251 differs from the projection system 1201 in that an optical element 1261 is provided instead of the optical element 1211.
The optical element 1261 is identical to the optical element 1211 in that the optical element 1261 is used in a state of not being in contact with a surface of an eyeball, but is different in that the optical element 1261 has a curved surface.
Note that the projection system is not limited to the above-described configuration as long as the retinal direct drawing can be realized.
<Other Modifications>
Although
Furthermore, similarly to displaying of a key top of the keyboard 602, operation support information for support of an operation of the operation device may be displayed as visual information in the field of view of the user. For example, a function of an operation device, an operation method, and the like are assumed as the operation support information.
Furthermore, in the above-described example, an example has been described in which only one window or virtual display is displayed in the user's field of view, but two or more windows or virtual displays may be displayed. In this case, for example, a window or a virtual display in a line-of-sight direction of the user may be set to an active state, and other windows or virtual displays may be set to an inactive state. Then, the window or the virtual display in the active state may be focused, and the window or the virtual display in the inactive state may be blurred.
Moreover, for example, the present technology can also be applied to a case where AR is implemented by using a display device other than the retinal direct drawing method, except for the embodiment of blurring the visual information.
The series of processes described above can be executed by hardware or also executed by software. When the series of processes are performed by software, a program that configures the software is installed in a computer. Here, examples of the computer include, for example, a computer that is built in dedicated hardware, a general-purpose personal computer that can perform various functions by being installed with various programs, and the like.
The program executed by the computer can be provided by being recorded in, for example, a removable medium (for example, a removable medium constituting the storage unit 119 of
Note that the program executed by the computer may be a program that performs processing in a time series according to an order described in this specification, or may be a program that performs processing in parallel or at necessary timing such as when a call is made.
Furthermore, in this specification, the system means a set of a plurality of components (a device, a module (a part), and the like), and it does not matter whether or not all the components are in the same housing. Therefore, a plurality of devices housed in separate housings and connected via a network, and a single device with a plurality of modules housed in one housing are both systems.
Note that the embodiment of the present technology is not limited to the above-described embodiment, and various modifications can be made without departing from the scope of the present technology.
For example, the present technology can have a cloud computing configuration in which one function is shared and processed in cooperation by a plurality of devices via a network.
Furthermore, each step described in the above-described flowchart can be executed by one device, and also shared and executed by a plurality of devices.
Moreover, in a case where one step includes a plurality of processes, the plurality of processes included in the one step can be executed by one device, and also shared and executed by a plurality of devices.
<Combination Example of Configuration>
The present technology can also have the following configurations.
(1)
A display device including:
(2)
The display device according to (1) described above, in which
(3)
The display device according to (2) described above, in which
(4)
The display device according to (3) described above, in which
(5)
The display device according to any one of (1) to (4) described above, in which
(6)
The display device according to any one of (1) to (5) described above, in which
(7)
The display device according to (6) described above, in which
(8)
The display device according to any one of (1) to (7) described above, in which
(9)
The display device according to any one of (1) to (8) described above, in which
(10)
The display device according to (9) described above, in which
(11)
The display device according to any one of (1) to (10) described above, in which
(12)
The display device according to (11) described above, in which
(13)
The display device according to any one of (1) to (12) described above, in which
(14)
The display device according to (13) described above, in which
(15)
The display device according to any one of (1) to (14) described above, in which
(16)
The display device according to (15) described above, in which
(17)
The display device according to (16) described above, in which
(18)
The display device according to any one of (1) to (17) described above, in which
(19)
A display method including:
(20)
A program for causing a computer to execute processing of:
Note that the effects described in this specification are merely examples and are not limited, and other effects may be present.
Tsuda, Takamoto, Moriya, Shoichiro, Yoshimori, Daisuke
Patent | Priority | Assignee | Title |
Patent | Priority | Assignee | Title |
9348141, | Oct 27 2010 | Microsoft Technology Licensing, LLC | Low-latency fusing of virtual and real content |
20180004289, | |||
20180174366, | |||
20190025815, | |||
CN107562184, | |||
CN107683497, | |||
JP2006308674, | |||
JP2015118832, | |||
JP2018004950, | |||
KR1020180004018, | |||
TW201804314, | |||
WO2006118057, | |||
WO2016203792, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Jul 10 2020 | SONY GROUP CORPORATION | (assignment on the face of the patent) | / | |||
Dec 08 2021 | TSUDA, TAKAMOTO | SONY GROUP CORPORATION | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 058675 | /0161 | |
Dec 22 2021 | MORIYA, SHOICHIRO | SONY GROUP CORPORATION | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 058675 | /0161 | |
Dec 25 2021 | YOSHIMORI, DAISUKE | SONY GROUP CORPORATION | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 058675 | /0161 |
Date | Maintenance Fee Events |
Jan 18 2022 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Date | Maintenance Schedule |
Dec 26 2026 | 4 years fee payment window open |
Jun 26 2027 | 6 months grace period start (w surcharge) |
Dec 26 2027 | patent expiry (for year 4) |
Dec 26 2029 | 2 years to revive unintentionally abandoned end. (for year 4) |
Dec 26 2030 | 8 years fee payment window open |
Jun 26 2031 | 6 months grace period start (w surcharge) |
Dec 26 2031 | patent expiry (for year 8) |
Dec 26 2033 | 2 years to revive unintentionally abandoned end. (for year 8) |
Dec 26 2034 | 12 years fee payment window open |
Jun 26 2035 | 6 months grace period start (w surcharge) |
Dec 26 2035 | patent expiry (for year 12) |
Dec 26 2037 | 2 years to revive unintentionally abandoned end. (for year 12) |