Provided is an information processing apparatus that includes a movement history acquiring section that acquires a movement history, based on an absolute coordinate, of a gazing point of a first user that gazes at a display region, in which the absolute coordinate is in a three-dimensional space displayed in the display region. The information processing apparatus further includes a display controller that controls a display unit that displays the display region to visualize, to a second user, the movement history of the gazing point. This allows the gazing point of the first user to be shared with the second user without hiding a target object.
|
15. An information processing method comprising:
acquiring eye gaze information indicating a gazing direction of a first user toward a display unit that displays a display region;
acquiring information indicating a position and an attitude of an imaging unit that captures an image of the display region;
calculating a movement history of a gazing point of the first user that gazes at the display region, wherein
the movement history of the gazing point is calculated based on the eye gaze information and the information indicating the position and the attitude of the imaging unit, and
the gazing point has an absolute coordinate in a three-dimensional space displayed in the display region; and
controlling the display unit to display the display region to visualize, to a second user, the movement history of the gazing point.
1. An information processing apparatus, comprising:
circuitry configured to:
acquire eye gaze information indicating a gazing direction of a first user toward a first display unit that displays a display region;
acquire information indicating a position and an attitude of an imaging unit that captures an image of the display region;
calculate a movement history of a gazing point of the first user that gazes at the display region, wherein
the movement history of the gazing point is calculated based on the eye gaze information and the information indicating the position and the attitude of the imaging unit, and
the gazing point has an absolute coordinate in a three-dimensional space displayed in the display region; and
control the first display unit to display the display region to visualize, to a second user, the movement history of the gazing point.
16. A non-transitory computer-readable medium having stored thereon computer-executable instructions which, when executed by a computer, cause the computer to execute operations, the operations comprising:
acquiring eye gaze information indicating a gazing direction of a first user toward a display unit that displays a display region;
acquiring information indicating a position and an attitude of an imaging unit that captures an image of the display region;
calculating a movement history of a gazing point of the first user that gazes at the display region, wherein
the movement history of the gazing point is calculated based on the eye gaze information and the information indicating the position and the attitude of the imaging unit, and
the gazing point has an absolute coordinate in a three-dimensional space displayed in the display region; and
controlling the display unit to display the display region to visualize, to a second user, the movement history of the gazing point.
2. The information processing apparatus according to
3. The information processing apparatus according to
4. The information processing apparatus according to
5. The information processing apparatus according to
6. The information processing apparatus according to
estimate a local coordinate of the gazing point based on the eye gaze information of the first user and depth information of the display region; and
convert the local coordinate of the gazing point into the absolute coordinate.
7. The information processing apparatus according to
8. The information processing apparatus according to
9. The information processing apparatus according to
10. The information processing apparatus according to
11. The information processing apparatus according to
12. The information processing apparatus according to
13. The information processing apparatus according to
the display region comprises an affected site region of a patient,
the first user comprises an operator,
the second user comprises an assistant that operates the imaging unit that captures the image of the display region, and
the imaging unit comprises a medical imaging unit.
14. The information processing apparatus according to
|
This application is a U.S. National Phase of International Patent Application No. PCT/JP2019/025279 filed on Jun. 26, 2019, which claims priority benefit of Japanese Patent Application No. JP 2018-167822 filed in the Japan Patent Office on Sep. 7, 2018. Each of the above-referenced applications is hereby incorporated herein by reference in its entirety.
The present disclosure relates to an information processing apparatus, an information processing method, and a program.
Recently, a technique has been developed that shares the experience of a first user with a second user through the vision, in a case where there are multiple users.
Patent Literature 1 discloses a technique in which an image expressing a field of view of the first user is generated on the basis of a captured image captured by an imaging unit worn on the head of the first user, and in which the image is presented to the second user. With this method, it is possible to provide an appropriate image to the second user when sharing the experience through the vision between the users.
In addition, in recent years, a technology that displays a motion of a user's eye gaze as a circle, a line, a heat map, or the like may sometimes be used to visualize and analyze the eye gaze.
Patent Literature 1: Japanese Unexamined Patent Application Publication No. 2015-95802
However, in a technique disclosed in Patent Literature 1, although a field of view of a first user is expressed to a second user, it is not possible for the second user to understand the event or the like that the first user focuses on in the field of view, and it is not possible for the second user to sufficiently understand the intention of the first user.
In addition, if a circle or a line, a heat map, or the like is used to represent the focused event in the field of view of the first user, a location viewed by the first user or the second user may be hidden.
For example, in a medical field, an assistant may sometimes operate various devices to perform a surgery while drawing the intention of an operator. At this time, drawing the operator's intention has been left to the assistant's ability, which can influence the smoothness of the surgery, etc. depending on the assistant's skill. In the medical field, an affected site region may be hidden if the event focused by the operator is indicated by the circle, the line, the heat map, or the like, which can possibly influence the surgical operation.
Accordingly, although it is not limited to the medical field, a case in which a target object gazed by the first user is hidden upon expressing the field of view of the first user to the second user can influence the second user sharing the field of view with the first user. Therefore, a technique has been demanded that shares a gazing point of the first user with the second user without hiding the target object.
The present disclosure provides an information processing apparatus that includes: a movement history acquiring section that acquires a movement history, based on an absolute coordinate, of a gazing point of a first user that gazes at a display region, in which the absolute coordinate is in a three-dimensional space displayed in the display region; and a display controller that controls a display unit that displays the display region to visualize, to a second user, the movement history of the gazing point.
In addition, the present disclosure provides an information processing method that causes a processor to execute a process that includes: acquiring a movement history, based on an absolute coordinate, of a gazing point of a first user that gazes at a display region, in which the absolute coordinate is in a three-dimensional space displayed in the display region; and controlling a display unit that displays the display region to visualize, to a second user, the movement history of the gazing point.
In addition, the present disclosure provides a program that causes a computer to function as: a movement history acquiring section that acquires a movement history, based on an absolute coordinate, of a gazing point of a first user that gazes at a display region, in which the absolute coordinate is in a three-dimensional space displayed in the display region; and a display controller that controls a display unit that displays the display region to visualize, to a second user, the movement history of the gazing point.
According to the present disclosure, the display unit is controlled to visualize, to the second user, the movement history of the gazing point of the first user.
As described above, according to the present disclosure, it is possible to share the gazing point of the first user with the second user without hiding a target object when the first user gazes at the target object. This allows the second user to understand the intention of the first user with respect to the target object. It is to be noted that the above-described effects are not necessarily limitative, and any of effects described in the present specification or other effects that may be understandable from the present specification may be achieved in addition to the above-described effects or instead of the above-described effects.
In the following, a preferred embodiment of the present disclosure is described in detail with reference to the accompanying drawings. It is to be noted that in the present specification and the drawings, components having substantially the same functional configuration are assigned the same reference numeral to avoid the repetition of description.
It is to be noted that description is given in the following order.
(1. Technical Overview)
A technique of the present disclosure relates to a technique that visualizes, to a second user, a movement history of a gazing point without hiding a target object when a first user gazes at a target object. In the present embodiment, a medical field will be exemplified, and a scene will be described in which a surgery is performed using an endoscope, where the first user is an operator and the second user is an assistant. However, the technique of the present disclosure is not limited to the medical field.
As described above, in the surgery, the affected site is presented to the operator 2, the assistant 1, and the like as the affected site region through the monitors or the like. In particular, in the endoscopic surgery, the assistant 1 called a scopist appropriately controls the surgical field captured by the endoscope 3, thereby creating an environment in which the operator 2 is able to perform the surgery smoothly. Accordingly, the assistant 1 has been required to understand or infer a location (a gazing point) that the operator 2 is gazing at in the surgical field, and to present the appropriate affected site region to the operator 2.
However, because the understanding or inference of the gazing point depend on the skill of the assistant 1, it has been difficult for the unskilled assistant 1 or the like to present the appropriate affected site region to the operator 2. Therefore, in a technique of the present disclosure, a movement history of the gazing point of a first user exemplified by the operator 2 is visualized and presented, thereby enabling a second user exemplified by the assistant 1 to appropriately understand the intention of the first user. Thus, in the medical field, the movement history of the gazing point of the operator 2 is visualized and presented to the assistant 1 without hiding the affected site, thereby facilitating the assistant 1 to properly understand the intention of the operator 2. The assistant 1 is able to predict a subsequent action to be taken by the operator 2 and provide the operator 2 with an appropriate surgical environment.
(2. Technical Details)
(2.1. Each Configuration and Function)
Referring to
The information processing system 1000 includes an eye gaze detection device 310, a local coordinate camera 320, a position detection device 330, a movement history calculator 210, an information processing apparatus 100, and the display unit 10.
The eye gaze detection device 310 is a device that detects eye gaze information of the operator 2. The eye gaze detection device 310 may be attached to the operator 2 or may be installed in an external environment without being attached to the operator 2. In a case where the eye gaze detection device 310 is to be attached to the operator 2, the eye gaze detection device 310 may be attached to the head or the face of the operator 2 to detect the eye gaze of the operator 2. The eye gaze information contains information on an orientation or a gazing direction of the face of the operator 2. The gaze refers to visual fixation of any location in a display region during viewing of the display region.
The eye gaze information may also be information acquired by detection of eye movement, pupil detection, or the like. For example, the eye gaze information may be acquired by imaging eyes of a user with a camera or the like and performing an image process. A known technique may be used as a technique for detecting the eye gaze information. For example, a pupil corneal reflection method or the like using a corneal reflection image in which light of a light source is reflected in a cornea may be used.
In a case where the eye gaze detection device 310 is installed in the external environment, the eye gaze detection device 310 may be installed in an operating room, for example, on a ceiling or a wall, to detect the eye gaze information of the operator 2. The eye gaze detection device 310 may be an infra-red camera, an imaging unit, etc.
The local coordinate camera 320 has a function of imaging a target object. For example, the local coordinate camera 320 may be provided at a tip of the endoscope 3. The local coordinate camera 320 captures an image of an organ or the like of the patient 8 at the time of the surgery as the target object. The local coordinate camera 320 has a function of capturing an image of the organ and acquiring position information of the organ with respect to the local coordinate camera 320. The position information of the organ may be a spatial coordinate of the organ based on the local coordinate camera 320, that is, a local coordinate. The position information of the organ may also be information on a direction and a distance from the local coordinate camera 320 to the organ. The local coordinate camera 320 may be a stereo-type camera that captures images of the organ from a plurality of angles, and the local coordinate camera 320 acquires the position information of the organ by performing the image process on the captured images captured from the plurality of angles.
In the above description, the local coordinate camera 320 is the imaging unit as an example of the present embodiment. However, the local coordinate camera 320 is not limited in its form as long as it is possible to grasp the position information of the target object. For example, a known sensor or the like that acquires the position information may be used instead of the local coordinate camera 320. Such a sensor may be provided separately from the imaging unit provided in the endoscope 3. The sensor may be a depth sensor, and may acquire the position information from the depth sensor to the organ. The depth sensor may specifically be a sensor, such as a ToF (Time of Flight) sensor, that measures a distance between the target object and the sensor from the time it takes for the projected laser to reciprocate to the target object. In addition, the position information may be acquired by a projection-type device provided with a projector and the imaging unit. In this technique, a pattern projected against the target object by the projector or the like is treated as an image captured by a camera of a projector viewpoint, and triangulation is performed between the image of the projector viewpoint and an image captured by the camera to acquire the position information of the target object. The imaging unit provided in the endoscope 3 is an exemplary surgical imaging unit in the present disclosure.
The position detection device 330 has a function of detecting a position and an attitude of the local coordinate camera 320. A location at which the position detection device 330 is not limited, as long as a position of the position detection device 330 in the three-dimensional space is apparent and as long as it is possible to detect the local coordinate camera 320. For example, the position detection device 330 is installed in the operating room, and performs imaging of the endoscope 3 that has the local coordinate camera 320, and performs an image process of a captured image to detect an absolute coordinate of the local coordinate camera 320 in the three-dimensional space. The position detection device 330 may be an infra-red camera, an imaging unit, etc.
In the above description, the position detection device 330 is the imaging unit as an example of the present embodiment. However, the position detection device 330 is not limited in its form as long as it is possible to grasp the position and the attitude of the local coordinate camera 320. For example, the position detection device 330 may be a magnetic sensor that detects an absolute coordinate of the local coordinate camera 320 by measuring a magnetic field emitted from an excitation coil.
The movement history calculator 210 has a function of calculating a movement history of the gazing point gazed by the operator 2, on the basis of pieces of information acquired by the eye gaze detection device 310, the local coordinate camera 320, and the position detection device 330. A method of calculating the movement history by the movement history calculator 210 will be described later.
The gazing point indicates a point in the display region with respect to the display region, and also indicates a point corresponding to the three-dimensional space with respect to the display region.
The information processing apparatus 100 includes a movement history acquiring section 110 and a display controller 120. The information processing apparatus 100 has a function of acquiring the movement history of the gazing point calculated by the movement history calculator 210 and visualizing the movement history by controlling the display unit 10. The movement history is visualized by the display unit 10, but is so displayed as not to hide the target object displayed in the display region.
The movement history acquiring section 110 acquires the movement history of the gazing point calculated by the movement history calculator 210 and outputs the acquired movement history to the display controller 120. The movement history acquiring section 110 may continuously perform a communication with the movement history calculator 210 at all times to acquire the movement history, or may discontinuously perform the communication to acquire the movement history. The movement history acquiring section 110 may acquire all or a part of the movement history calculated by the movement history calculator 210.
The display controller 120 has a function of acquiring the movement history of the gazing point from the movement history acquiring section 110 and controlling the display unit 10 so as to visualize the movement history. The display controller 120 changes a display style of the movement history to visualize a gazing region. Further, the display controller 120 may vary the display style depending on the display unit 10 to be used. The display styles of the movement history will be described later.
The display unit 10 is controlled by the display controller 120 and has a function of displaying the movement history in the display region. The display unit 10 is not limited in its form as long as the display unit 10 is able to present the captured image captured by the endoscope 3 to the user and visualize the movement history. For example, the display unit 10 may be a display unit used by being attached to the user, such as a head-mounted display (HMD: Head Mounted Display), or may be a display unit, such as a stationary monitor. The HMD is an example of a wearable display unit in the present disclosure.
With the above-described respective configurations, the movement history of a gazing point 113 of the operator 2 may be acquired by the movement history acquiring section 110 of the information processing apparatus 100, and the assistant monitor 10 may be controlled by the display controller 120 to display the movement history to the assistant.
(2.2. Calculation of Movement History)
Referring to
The tip of the endoscope 3 is provided with the local coordinate camera 320 that performs the imaging of the inside of the body of the patient 8 and detects a position of the organ in the body. The local coordinate camera 320 may be a stereo-type camera. The captured image captured by the local coordinate camera 320 is displayed on the operator monitor 20. The operator 2 gazes at a display region 21 of the operator monitor 20. In
The local coordinate camera 320 detects the position of the organ of the patient 8 based on the local coordinate camera 320 as the spatial coordinate of a local coordinate system X (hereinafter referred to as a local coordinate). Thus, the local coordinate of the position of the organ corresponding to the display region 21 of the operator monitor 20 is known. Here, the eye gaze detection device 310 acquires the eye gaze information of the operator 2 with respect to the display region 21, and the movement history calculator 210 acquires the gazing direction 112 in which the operator 2 gazes.
The local coordinate camera 320 acquires the position information of the organ from the local coordinate camera 320 as depth information. The depth information includes information on a distance and a direction from the local coordinate camera 320 to the organ. The movement history calculator 210 estimates, from the gazing direction 112 with respect to the display region 21 and the position information of the organ, the local coordinate of the gazing point at which the operator 2 gazes. Note that, in a case where it is not possible to estimate, from the gazing direction and the depth information, the local coordinate which is based on the local coordinate camera 320, the local coordinate of the gazing point may be estimated on the assumption that the organ is distant from the local coordinate camera 320 by a predetermined distance. In addition, the gazing direction with respect to the display region 21 may be estimated from both eye congestion to estimate the local coordinate.
Further, the position detection device 330 that detects the position and the attitude of the local coordinate camera 320 is provided in the operating room. The position detection device 330 acquires, as an absolute coordinate of an absolute coordinate system Y in the three-dimensional space, the position and the attitude of the local coordinate camera 320 in the three-dimensional space. In the local coordinate camera 320, the position of the organ is represented as the local coordinate. Accordingly, knowing the absolute coordinate of the local coordinate camera 320 makes it possible to convert the position of the organ into the absolute coordinate.
The movement history calculator 210 thus estimates that an intersection of a depth of the organ and the gazing direction is the gazing point, on the basis of the depth information of the captured image of the local coordinate camera 320 and the eye gaze information of the operator 2. Specifically, the operator 2 performs the surgery while viewing, in the viewing range 111, the captured image displayed in the display region 21 of the operator monitor 20. The eye gaze detection device 310 detects the gazing direction 112 in which the operator 2 gazes as a surgical site in the viewing range 111. The movement history calculator 210 has already acquired the depth information in which the position of the organ is represented as the spatial coordinate in the captured image to be displayed on the operator monitor 20; hence, it is possible for the movement history calculator 210 to determine the intersection of the gazing direction 112 and the captured image to be displayed on the operator monitor 20 as the gazing point 113. The gazing point 113 is grasped by the movement history calculator 210 as the spatial coordinate of the local coordinate from a point in the display region of the operator monitor 20, and the spatial coordinate of the local coordinate is further converted as the absolute coordinate. By grasping the gazing point 113 in the form of the absolute coordinate, it is possible for the movement history calculator 210 to display the gazing point 113 to the assistant 1 without involving the movement of the gazing point 113 even if the position and the attitude of the local coordinate camera 320 have changed.
(2.3. Storing of Movement History)
Referring to
In the assistant monitor 10 illustrated in
The gazing region of the gazing point D, the gazing region of the gazing point E, and the gazing region of the gazing point F are overlapped with the gazing region 122c. The gazing region may be so displayed that the display regions corresponding to the respective gazing points are overlapped with each other. Further, for the gazing region, a point having equal distances from a plurality of gazing points may be treated as one gazing point, and the gazing region may include such one gazing point. For example, a region of 2 degrees from the user's gazing point may be treated as a high-resolution region, and the region corresponding to the high-resolution region may be treated as the gazing region.
In
Thus, instead of displaying only the most recent gazing point, the past gazing points are displayed to display the movement history, allowing the assistant 1 to estimate the intention of the gazing of the operator 2. However, if a large amount of past gazing points are displayed, it becomes difficult to understand where the most recent gazing point is. Accordingly, the movement history of the gazing points exceeding the predetermined number is not displayed.
(2.4. Displaying of Movement History)
The movement history of the gazing point may be displayed in various display styles on the display unit 10 by the display controller 120. The display controller 120 controls the display unit 10 such that the gazing region including the gazing point in the display region and the region other than the gazing region are displayed in different display styles. By the display style, it is possible to increase visibility of the gazing region as compared with the visibility of the region other than the gazing region. The visibility indicates a level of easiness of viewing when the assistant 1 visually recognizes the captured image displayed in the display region. It is possible to attract the attention of the assistant 1 to the gazing region by increasing the visibility of the gazing region. For example, it is possible to make the visibility different between the gazing region and the region other than gazing region in each pixel in the display region, by the saturation, the brightness, the resolution, and the like.
The higher the saturation, the clearer the color display becomes, and the lower the saturation, more monochromatic the displaying becomes. Therefore, the visibility of the gazing region may be increased by making the saturation of the gazing region higher than the saturation of the region other than the gazing region.
The higher the brightness, whiter the displaying becomes, and the lower the brightness, blacker the displaying becomes. Therefore, the visibility of the gazing region may be increased by increasing the brightness of the gazing region.
The higher the resolution is, the higher the definition is, and the lower the resolution is, the lower the definition is. Therefore, the visibility of the gazing region may be increased by making the resolution of the gazing region higher than the resolution of the region other than the gazing region.
For example, with reference to
Thus, the color level r increases, the color property increases, and the visibility increases for any pixel in which the distances from the respective gazing points more overlap with each other, Note that, here, an upper limit is set for the maximum color level rMAX, and the maximum color level rMAX may be displayed in a case where a value exceeding the upper limit is calculated as a value of the color level r.
In addition, in general, in a case where the color hue of each of the pixels includes red R, green G, and blue B, the monochrome hue that performs monochromatizating thereof is calculated by ((0.299×R)+(0.587×G)+(0.114×B)). At this time, the hue to be drawn on each of the pixels as the display style of the gazing region may be expressed by the following expression (1).
((color hue)×(color level r))+((monochrome hue)×(1−(color level r))) (1)
In the expression (1), the term ((monochrome hue)×(1−color level r)) may be further multiplied by a coefficient greater than zero and less than one to relatively increase the color level. Specifically, by multiplying the coefficient by 0.5, the monochrome level may be halved, and an influence of the monochrome hue of the region other than the gazing region in the display region may be weakened to increase the visibility of the pixel of the gazing region that displays the color hue.
Although
The above-described gazing region is represented in a circular shape with the gazing point being the center, but the gazing region may be represented in another figure. In addition, a size of the figure may be determined on an as-necessary basis. As the figure becomes larger, it is possible to attract the attention of the assistant 1 to a wider range of the affected site region, and the assistant 1 becomes more aware of a change in the affected site, etc. As the figure becomes smaller, the attention of the assistant 1 is more attracted to the center of the gazing point, allowing the assistant 1 to address a fine change in position of the endoscope 3.
The display style of the gazing region is not limited to the distance d from the gazing point as described above, and may be determined by the elapsed time from the occurrence of the gazing point. For example, as illustrated in
Further, as illustrated in
Next, referring to
Referring to
(3. Operation Flow)
The respective configurations of the information processing system 1000 have been described above. Next, an operation flow of the information processing system 1000 will be described with reference to
First, an image of the surgical field of the patient 8 is captured by the endoscope 3 provided with the local coordinate camera 320 (S102). The local coordinate camera 320 may be a plurality of imaging units provided in the endoscope 3, and may be of a stereo type. Note that, in the following, an operation flow in a case of performing gazing point sharing is described, and the captured image captured by the imaging unit is displayed on the display unit as it is in a case of where the gazing point sharing is not performed. The instructions for sharing and cancelling the sharing of the gazing point with respect to the information processing system 1000 may be performed by the operator 2 or may be performed by the assistant 1. A change in the display style at this time may involve a gradually change in the display style in accordance with the instructions of sharing and cancelling the sharing of the gazing point.
When the sharing of the gazing point is instructed, the local coordinate camera 320 acquires, from the captured image, the position information such as the position of the organ in the surgical field of the patient 8 (S104). In this case, the position information such as the position of the organ may be represented as the spatial coordinate that is based on the position of the local coordinate camera 320. The position of the organ includes information such as a depth of each point at any point.
Next, the eye gaze detection device 310 detects the eye gaze information of the operator 2 (S106).
Next, the movement history calculator 210 estimates the local coordinate of the gazing point, on the basis of the position information of the organ acquired from the captured image and the eye gaze information of the operator 2 with respect to the display region of the operator monitor 20 (S108).
Next, the position detection device 330 detects the position and the attitude of the endoscope 3. By detecting the position and the attitude of the endoscope 3, the position and the attitude of the endoscope 3 is acquired in the form of the absolute coordinate (S110).
Next, because the position and the attitude of the endoscope 3 are acquired in the form of the absolute coordinate, the movement history calculator 210 converts the local coordinate of the gazing point into the absolute coordinate (S112).
Next, the movement history calculator 210 stores the gazing point converted into the absolute coordinate (S114).
Next, the movement history calculator 210 determines whether the number of stored gazing points has exceeded the predetermined number (S116). If the number of stored gazing points exceeds the predetermined number (S116/Yes), the movement history calculator 210 causes the process to proceed to the next process.
Next, the movement history calculator 210 deletes the oldest gazing point out of the stored gazing points (S118). Note that, if the number of stored gazing points is less than the predetermined number (S116/No), the movement history calculator 210 causes the process to proceed to the next process without performing the step (S118) of deleting the gazing point.
Next, the movement history acquiring section 110 of the information processing apparatus 100 acquires the movement history of the gazing point, and the display controller 120 performs mapping of the movement history on the captured image displayed in the display region of the display unit 10.
Next, the display controller 120 displays the gazing region containing the respective gazing points in the movement history in color, and displays the region other than the gazing region in monochrome (S122).
Next, the information processing system 1000 determines whether to terminate the sharing of the gazing point (S124). If the information processing system 1000 determines to terminate the sharing of the gazing point (S124/Yes), the operation of sharing the gazing point is terminated. Further, if the information processing system 1000 determines not to terminate the sharing of the gazing point (S124/No), the captured image is acquired from the endoscope 3 again (S102) to repeat the operation up to the above continuously. The determination of whether or not to terminate the sharing of the gazing point may be made on the basis of whether or not the information processing system 1000 has acquired the instructions to cancel the sharing of the gazing point from the assistant 1 or the operator 2, for example.
The configuration and the operation flow of the information processing system 1000 according to this embodiment have been described. Although the present embodiment exemplifies that the movement history of the gazing point of the operator is displayed in real time, the movement history of the gazing point may be displayed by being recorded together with the captured image captured by the imaging unit and reproduced by the display unit or the like. At this time, the captured image gazed by the operator, the position and the attitude of the endoscope, and the absolute coordinate of the gazing point are recorded in the information processing system 1000.
(4. Modification Examples)
In this section, examples of specific examples of the display unit 10 during the surgery will be described separately for each scene. The display unit 10 is not limited only to a case where the display unit 10 described in the above embodiment is the monitor display unit, but may be any of various wearable terminals to be attached to and used by the user.
(4.1. Operator Wears Loupe-Type Display Unit)
Referring to
In
The loupe-type first display unit 20A is further provided with a loupe built-in eye gaze detector (not illustrated) as the eye gaze detection device 310. The position detection device 330 that performs imaging of the loupe-type first display unit 20A may be installed in an operating room or the like.
On the basis of pieces of information acquired from the eye gaze detection device 310, the local coordinate camera 320, and the position detection device 330, the assistant monitor 10 displays the captured image, thereby visualizing and displaying the movement history of the gazing point of the operator 2.
(4.2. Operator Wears Loupe-Type Display Unit and Assistant Wears HMD-Type Display Unit)
Referring to
In
The loupe-type first display unit 20A is further provided with the loupe built-in eye gaze detector (not illustrated) as the eye gaze detection device 310. The position detection device 330 that performs the imaging of the loupe-type first display unit 20A may be installed in the external environment such as the operating room.
The captured image captured by the loupe built-in camera may be displayed on the loupe-type first display unit 20A. Further, in a display region of the HMD-type second display unit 10A, the captured image is displayed as the affected site region as viewed from the HMD-type second display unit 10A worn by the assistant 1, in view of a relationship between a position and an attitude of the loupe built-in camera and the position and the attitude of the HMD-type second display unit 10A.
On the basis of pieces of information acquired from the eye gaze detection device 310, the local coordinate camera 320, and the position detection device 330, the HMD-type second display unit 10A displays the captured image, thereby visualizing and displaying the movement history of the gazing point of the operator 2.
(4.3. Operator Wears HMD-Type Display Unit)
Referring to
In
(4.4. Operator and Assistant Operate Microscope)
Referring to
In a case of the microscope, instead of the endoscope 3 described above, the imaging unit is a microscope built-in camera that is built into the microscope. The position detection device that detects a position and an attitude of the microscope-built-in camera may be built in the microscope as well, or may be installed in the external environment for detection thereof.
The operator 2 visually recognizes the captured image through a display region of a microscope eyepiece 20B provided on the microscopic. In addition to the microscopic eyepiece 20B, the microscope is provided with a microscope built-in eye gaze detector as the eye gaze detection device, which enables the eye gaze detection of the operator 2.
The movement history of the gazing point of the operator 2 is visualized and displayed in the display region of the microscopic eyepiece 10B provided on the microscope. In the present embodiment, the movement history is visualized in the display region of the microscopic eyepiece 10B for the assistant, but the movement history may be displayed in a display region such as a monitor that is not provided in the microscope.
(5. Hardware Configuration Example)
Referring to
As illustrated in
The CPU 901 functions as an arithmetic processing unit and a control unit, and controls overall operations in the information processing apparatus 900 in accordance with various programs. Further, the CPU 901 may be a microprocessor. The ROM 902 stores programs, arithmetic parameters, etc. that the CPU 901 uses. The RAM 903 temporarily stores a program that is used in execution of the CPU 901, parameters that change appropriately in execution thereof, and the like. The CPU 901 may form, for example, the movement history calculator 210 or the display controller 120 illustrated in
The CPU 901, the ROM 902, and the RAM 903 are connected to each other by the host bus 904a including a CPU bus and the like. The host bus 904a is connected to the external bus 904b such as a PCI (Peripheral Component Interconnect/Interface) bus via the bridge 904. Note that the host bus 904a, the bridge 904, and the external bus 904b are not necessarily configured separately, and these functions may be packaged in a single bus.
The input unit 906 is implemented by, for example, a unit through which the user inputs information, such as a mouse, a keyboard, a touch panel, a button, a microphone, a switch, or a lever. Further, the input unit 906 may be a remote-control unit that uses infrared ray or other electromagnetic waves, or may be external connection equipment such as a mobile phone or a PDA compatible with operations of the information processing apparatus 900. Further, the input unit 906 may include, for example, an input control circuit or the like that generates an input signal on the basis of information inputted by the user using the input means described above and outputs the generated input signal to the CPU 901. It is possible for the user of the information processing apparatus 900 to input various data or provide instructions for a processing operation to the information processing apparatus 900 by operating the input unit 906.
The display unit 907 is formed by a unit that is able to notify the user of acquired information visually or audibly. Examples of such a unit include displays such as a CRT display, a liquid crystal display, a plasma display, an EL display, a laser projector, an LED projector, or a lamp, and sound output units such as a speaker or a headphone. The display unit 907 outputs, for example, results obtained through various processes performed by the information processing apparatus 900. Specifically, the display unit 907 visually displays the results obtained through various processes performed by the information processing apparatus 900 in a variety of formats, such as text, images, tables, graphs, etc. Meanwhile, the sound output unit converts audio signals including reproduced sound data, acoustic data or the like into analog signals and outputs the analog signals audibly. The display unit 907 visualizes and displays the movement history of the gazing point acquired by the movement history acquiring section 110 illustrated in
The storage unit 908 is a data storing unit formed as an example of a memory section of the information processing apparatus 900. The storage unit 908 is implemented by, for example, a magnetic memory section device such as an HDD, a semiconductor memory device, an optical memory device, a magneto-optical memory device, or the like. The storage unit 908 may include a storage medium, a recording unit for recording data on the storage medium, a reading unit for reading data from the storage medium, a deletion unit for deleting data recorded on the storage medium, etc. The storage unit 908 stores a program to be executed by the CPU 901, various data, and various externally acquired data, etc. The storage unit 908 stores, for example, various parameters and the like used when the movement history of the gazing point is to be calculated by the movement history calculator 210 illustrated in
The drive 909 is a reader/writer for a storage medium, and is incorporated in or externally attached to the information processing apparatus 900. The drive 909 reads information recorded on a removable storage medium mounted thereon, such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory, and outputs the information to the RAM 903. The drive 909 is also able to write information on the removable storage medium.
The connection port 911 is an interface to be connected to external equipment, and is a connector to the external equipment that is able to transmit data through, for example, a USB (Universal Serial Bus) or the like.
The communication unit 913 is, for example, a communication interface formed by a communication unit or the like for connection to a network 920. The communication unit 913 may be, for example, a communication card or the like for wired or wireless LAN (Local Area Network), LTE (Long Term Evolution), Bluetooth (registered trademark), or WUSB (Wireless USB). Further, the communication unit 913 may be a router for optical communication, a router for ADSL (Asymmetric Digital Subscriber Line), modems for various types of communications, or the like. The communication unit 913 is able to transmit and receive signals or the like to and from the Internet or other communication equipment in accordance with predetermined protocols such as TCP/IP, for example. The communication unit 913 is provided in, for example, various configurations illustrated in
Note that the network 920 is a wired or wireless transmission path for information transmitted from units connected to the network 920. For example, the network 920 may include a public network such as the Internet, a telephone network or a satellite communication network, various LANs (Local Area Networks) including Ethernet (registered trademark), WAN (Wide Area Network), and the like. Further, the network 920 may include a dedicated line network such as an IP-VPN (Internet Protocol-Virtual Private Network).
Further, in the information processing apparatus 900, it is possible to create a computer program for causing the hardware such as a CPU, a ROM, or a RAM incorporated in the information processing system 1000 to perform functions equivalent to the configurations of the information processing system 1000 according to the present embodiment described above. A recording medium storing the computer program may also be included in the scope of a technique according to the present disclosure.
As above, the preferred embodiment of the present disclosure has been described in detail with reference to the accompanying drawings; however, the technical scope of the present disclosure is not limited to these examples. It is apparent that those having ordinary skill in the technical field of the present disclosure could easily arrive at various modified examples or revised examples within the meaning of the technical concept recited in claims, and it is understood that these also should naturally fall under the technical scope of the present disclosure.
For example, in the above embodiment, the operator and the assistant in the scene of the surgery in the medical field are taken as an example, but the present technology is not limited to such an example. For example, the present technology may be used in any use in presenting the movement history of the gazing point of the first user to the second user. Specifically, the present technology may be used in the field of education. In a case where a student performs learning such as surgery, it is possible for the student to intuitively understand a point to be gazed at the time of the surgery by learning through a picture on which the movement history of the gazing point of the operator is displayed, thereby enabling effective learning. Meanwhile, it is possible for an instructor to instruct, for example, a timing and where to gaze by understanding the movement history of the gazing point of the student.
For example, the present technology may also be used to prevent overlooking or the like upon monitoring of a surveillance camera in a monitoring room. In a case of monitoring of the surveillance camera, in the display region of the captured image captured by the surveillance camera, the gazing region of an observer indicates that the surveillance has been performed, and the region other than the gazing region indicates that the surveillance has not yet been performed. Thus, it is possible to perform the monitoring more finely by prompting the monitoring of the region other than the gazing region. Therefore, by applying a technique of the present disclosure, it is possible to achieve an effect of preventing overlooking of the surveillance camera through attracting the attention of the observer to the region other than the gazing region, by changing the display style so that the visibility of the region other than the gazing region becomes high and the visibility of the gazing region becomes low. In particular, in a case where there are multiple observers, it is possible to prevent overlapping of the gazing points by displaying the movement history of the gazing point gazed by each of the observers, thereby making it possible to efficiently perform the surveillance operation. As a type of the display style, the movement history of the gazing point may be displayed in monochrome and the display region other than the gazing region may be displayed in color to enhance visibility in a case of the surveillance camera.
Furthermore, the effects described in the present specification are only explanatory or exemplary and not limitative. That is, a technique according to the present disclosure may achieve other effects that should be apparent from the description of the present specification by those skilled in the art in addition to the above-described effects or instead of the above-described effects.
It is to be noted that the following configurations also fall within the technical scope of the present disclosure.
The information processing apparatus according to (2), in which the movement history is derived by converting a local coordinate into the absolute coordinate, the local coordinate being based on the imaging unit and estimated using: the eye gaze information of the first user; and depth information of the display region.
Patent | Priority | Assignee | Title |
Patent | Priority | Assignee | Title |
20060170652, | |||
20180344138, | |||
20190114823, | |||
20190253743, | |||
20190354174, | |||
20190354176, | |||
CN108778093, | |||
CN109716396, | |||
EP1686554, | |||
JP2006293605, | |||
JP2009194697, | |||
JP201595802, | |||
WO2017183353, | |||
WO2017191700, | |||
WO2018051592, | |||
WO2018079166, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Jun 26 2019 | Sony Corporation | (assignment on the face of the patent) | / | |||
Jan 12 2021 | NODA, TAKURO | Sony Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 055432 | /0455 |
Date | Maintenance Fee Events |
Feb 26 2021 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Date | Maintenance Schedule |
Oct 25 2025 | 4 years fee payment window open |
Apr 25 2026 | 6 months grace period start (w surcharge) |
Oct 25 2026 | patent expiry (for year 4) |
Oct 25 2028 | 2 years to revive unintentionally abandoned end. (for year 4) |
Oct 25 2029 | 8 years fee payment window open |
Apr 25 2030 | 6 months grace period start (w surcharge) |
Oct 25 2030 | patent expiry (for year 8) |
Oct 25 2032 | 2 years to revive unintentionally abandoned end. (for year 8) |
Oct 25 2033 | 12 years fee payment window open |
Apr 25 2034 | 6 months grace period start (w surcharge) |
Oct 25 2034 | patent expiry (for year 12) |
Oct 25 2036 | 2 years to revive unintentionally abandoned end. (for year 12) |