The disclosed display system may include (1) a display element that is transparent and includes a first surface facing a viewing region and a second surface opposite the first surface, the second surface facing an external environment, (2) a light-blocking shutter that faces and overlaps at least a portion of the second surface of the display element, (3) a display-driving subsystem that displays images visible to a user at the viewing region via display light emitted from the first surface of the display element during active display intervals, and (4) a shutter-driving subsystem that alternately switches the state of the light-blocking shutter between a blocking state during the active display intervals and a pass-through state during inactive display intervals when the display light is not emitted from the display element. Various other systems, devices, and methods are also disclosed.
|
15. A computer-implemented method, comprising:
displaying images visible to a user via display light emitted from one of a plurality of pixels included in a display element during active display intervals, wherein:
the display element is transparent;
the plurality of pixels are individually controllable; and
the one of the plurality of pixels is pulsed multiple times within a burst sub-period of consecutive cycle periods;
refraining from displaying the images to the user via any of the plurality of pixels included in the display element during inactive display intervals, each of the plurality of pixels remaining inactive within an OFF sub-period of the consecutive cycle periods;
alternately switching a light-blocking shutter between a blocking state in which light transmission through the light-blocking shutter is blocked and a pass-through state in which light transmission through the light-blocking shutter is enabled;
maintaining the light-blocking shutter in the blocking state during the entirety of the burst sub-period in which the one of the plurality of pixels is pulsed multiple times; and
maintaining the light-blocking shutter in the pass-through state during the entirety of the OFF sub-period in which each of the plurality of pixels remains inactive; and
wherein each of the consecutive cycle periods consists of the burst sub-period and the OFF sub-period.
1. A display system, comprising:
a display element that is transparent and comprises a plurality of pixels that are individually controllable;
a light-blocking shutter that overlaps at least a portion of a surface of the display element that faces an external environment;
a display-driving subsystem that:
displays images visible to a user via display light emitted from one of the plurality of pixels included in the display element during active display intervals, the one of the plurality of pixels being pulsed multiple times within a burst sub-period of consecutive cycle periods; and
refrains from displaying the images to the user via any of the plurality of pixels included in the display element during inactive display intervals, each of the plurality of pixels remaining inactive within an OFF sub-period of the consecutive cycle periods; and
a shutter-driving subsystem that:
alternately switches the light-blocking shutter between a blocking state in which light transmission through the light-blocking shutter is blocked and a pass-through state in which light transmission through the light-blocking shutter is enabled;
maintains the light-blocking shutter in the blocking state during the entirety of the burst sub-period in which the one of the plurality of pixels is pulsed multiple times; and
maintains the light-blocking shutter in the pass-through state during the entirety of the OFF sub-period in which each of the plurality of pixels remains inactive; and
wherein each of the consecutive cycle periods consists of the burst sub-period and the OFF sub-period.
14. A head-mounted display device, comprising:
an optical subsystem that provides a user with composite views of displayed images and an external environment, the optical subsystem comprising:
a transparent display element that comprises a plurality of pixels that are individually controllable; and
a light-blocking shutter that overlaps at least a portion of a surface of the transparent display element that faces the external environment;
a display-driving subsystem that:
displays images visible to the user via display light emitted from one of the plurality of pixels included in the transparent display element during active display intervals, the one of the plurality of pixels being pulsed multiple times within a burst sub-period of consecutive cycle periods; and
refrains from displaying the images to the user via any of the plurality of pixels included in the transparent display element during inactive display intervals, each of the plurality of pixels remaining inactive within an OFF sub-period of the consecutive cycle periods; and
a shutter-driving subsystem that:
alternately switches the light-blocking shutter between a blocking state in which light transmission through the light-blocking shutter is blocked and a pass-through state in which light transmission through the light-blocking shutter is enabled;
maintains the light-blocking shutter in the blocking state during the entirety of the burst sub-period in which the one of the plurality of pixels is pulsed multiple times; and
maintains the light-blocking shutter in the pass-through state during the entirety of the OFF sub-period in which each of the plurality of pixels remains inactive; and
wherein each of the consecutive cycle periods consists of the burst sub-period and the OFF sub-period.
2. The display system of
3. The display system of
4. The display system of
5. The display system of
6. The display system of
7. The display system of
8. The display system of
9. The display system of
10. The display system of
the display element further comprises a pass-through region through which light from the external environment is transmitted at least when the light-blocking shutter is in the pass-through state; and
the shutter-driving subsystem drives the light-blocking shutter to block light transmission through a light-blocking region of the light-blocking shutter overlapping at least a portion of the pass-through region of the display element.
11. The display system of
12. The display system of
the light-blocking region overlaps a selected portion of the pass-through region of the display element so as to block externally-directed light from at least one of a light-emitting region of the display element or a portion of the user's face illuminated by the display light; and
the light-blocking region does not overlap a surrounding portion of the pass-through region of the display element outside of the selected portion.
13. The display system of
16. The computer-implemented method of
17. The computer-implemented method of
|
The accompanying drawings illustrate a number of exemplary embodiments and are a part of the specification. Together with the following description, these drawings demonstrate and explain various principles of the instant disclosure.
Throughout the drawings, identical reference characters and descriptions indicate similar, but not necessarily identical, elements. While the exemplary embodiments described herein are susceptible to various modifications and alternative forms, specific embodiments have been shown by way of example in the drawings and will be described in detail herein. However, the exemplary embodiments described herein are not intended to be limited to the particular forms disclosed. Rather, the instant disclosure covers all modifications, equivalents, and alternatives falling within the scope of the appended claims.
The present disclosure is generally directed to systems and methods for minimizing external light leakage from artificial-reality displays. Artificial-reality systems are widely gaining in popularity for use in a growing number of activities. For example, artificial-reality systems (including augmented-reality, mixed-reality, and virtual-reality devices, such as headsets) may enable users to experience events, such as interactions with people in a computer-generated simulation of a three-dimensional world or viewing images superimposed on a real-world view. Such devices and headsets are also commonly used for purposes other than recreation. For example, governments may use such devices for military training, medical professionals may use the devices to simulate surgery, and engineers may use such devices as design visualization aids.
Artificial-reality systems may enable users to have immersive or semi-immersive experiences in which digital display elements block or obscure all or a substantial portion of a user's view of his or her real-world environment. Some augmented-reality systems produce electronically-generated image elements that are overlaid on a user's view of his or her real-world environment. Additionally, in mixed-reality and virtual-reality systems, different portions of a captured scene or an artificially-generated scene may be displayed to a user.
Augmented-reality and mixed-reality headsets often include optical see-through displays that transmit light from a user's ambient environment to the user's eyes such that their surroundings are visible. At the same time, various image elements may be displayed to the user to form a composite image. Undesirably, such headsets may display images that are visible, not only to the user, but often to other individuals as well. While headsets, such as augmented-reality glasses, have become more compact and convenient for wearers, such devices may inadvertently leak display light externally so as to be visible to other individuals. For example, augmented-reality glasses may display images to a user via transparent display elements in the lenses that also leak light into the external environment. Additionally, light from the display elements may be undesirably visible on a portion of a user's face, such as the user's eyes. Such leaked light may be visually unpleasing or distracting to other observers and may be considered socially unacceptable in many contexts. Additionally, if the displayed content is visible to others, the light leakage may pose a privacy or security issue to users.
As will be explained in greater detail below, the systems and methods disclosed herein may effectively mitigate the undesirable leakage of display light from artificial-reality devices while maintaining the transparency of optical see-through display elements essential to the effective operation of such devices. Embodiments of the present disclosure may have one or more advantages over conventional artificial-reality systems, such as augmented-reality and mixed-reality systems. For example, by utilizing a light-blocking shutter in combination with a light-emitting display element that produces images via short pulses of light, the amount of light escaping into the user's surroundings may be effectively reduced. While such a light-blocking shutter may be disposed in the field of view of both a head-mounted display wearer and other individuals in the wearer's environment, it may be made unnoticeable by running such a shutter at a short duty cycle corresponding to that of the display element. For example, a light-blocking shutter may be driven to block light from at least a portion of the display area in conjunction with emission of light from the display element. Thus, while the light-blocking member may temporarily block light from passing through, for example, the lens region of an augmented-reality headset, the lenses may still appear to be transparent during use since, during the majority of the viewing time, the light-blocking shutter is not activated and does not block the passage of light therethrough. Conversely, while the display element may be driven in short pulses while remaining in an off state between such pulses for the majority of the time, the emitted light may still be apparent to the wearer and may be viewed as a continuous image while light is nonetheless blocked to the view of external observers by the light-blocking shutter. Additionally, since the light-blocking shutter is driven at the same time that the light-emitting element displays an image, the contrast and visibility of the image may be enhanced from the wearer's perspective and may increase the perceived brightness of the display images.
The following will provide, with reference to
As shown, display system 100 may include a display-driving subsystem 104 (e.g., a display driver) that presents, via display element 102, a variety of content to a user, including virtual views of an artificially rendered virtual-world environment and/or augmented views of a physical, real-world environment. In some examples, the augmented views may be augmented with computer-generated elements (e.g., two-dimensional (2D) or three-dimensional (3D) images, 2D or 3D video, sound, etc.). In some examples, images 106 may convey a scene (e.g., a captured scene, an artificially-generated scene, an evolving artificial-reality environment, or a combination of the same) to the user.
In various embodiments, display element 102 may include a single electronic display or multiple electronic displays (e.g., a display for each eye of a user). Examples of display element 102 may include a liquid crystal display (LCD), an organic light emitting diode (OLED) display, an inorganic light emitting diode (ILED) display, an active-matrix organic light-emitting diode (AMOLED) display, a transparent organic light emitting diode (TOLED) display, a waveguide display, a digital light processing (DLP) display, a microLED display, a liquid crystal on silicon (LCOS) display, a ferroelectric liquid crystal on silicon (fLCOS) display, or another suitable transparent display, or some combination thereof. As shown, display element 102 may include a plurality of display regions 103 (e.g., pixels or subpixels) that are individually controllable. In some examples, each of display regions 103 may be transparent or partially transparent, at least when not displaying an image or video, such that ambient light from a user's real-world environment is able to pass through the display region.
As shown in
Light-blocking shutter 110 may include a plurality of blocking regions 111 that are individually addressable and switchable between at least a blocking state (e.g., an opaque state or a shuttered state) in which light is blocked from passing through the region and a pass-through state (e.g., a transparent state or an un-shuttered state) in which light is allowed to pass through the region. In some examples, each of blocking regions 111 may be oriented relative to a corresponding one or more of display regions 103, such that, when in a blocking state, the blocking region blocks light from reaching or passing through a corresponding display region. In some embodiments, light-blocking shutter 110 may be a liquid crystal layer that is divided into regular or irregular regions that can be independently turned dark or clear. Light-blocking shutter 110 may use any appropriate approach for blocking or transmitting light, including e-inks, reflective devices such as MEMS mirrors or thin film liquid crystal switchable mirrors, photochromic materials, electrochromic elements, switchable scattering materials and the like. For example, spatially selective electric field application to an electrochromic layer may be used to achieve localized light blocking or transmittance. The light blocking effect may be achieved by switchable reflectivity, absorption, or scattering of the light and may involve polarization imparting elements (e.g., an absorbing linear polarizing film in addition to the switchable shutter).
In at least one embodiment, each of blocking regions 111 of light-blocking shutter 110 may be an optical shutter (e.g., a liquid-crystal shutter or mechanical shutter) capable of blocking or controlling light transmittance. An optical shutter may have pixels that have dark or light states. In a dark state, the pixel may be generally opaque and block light from a portion of the local environment from reaching the user's view. In a light state, the pixel may be relatively transparent, allowing light from the local environment to pass through the pixel and reach the user's view. In some examples, a blocking element may include pixels with multiple available state options (for example, clear/dim/opaque or clear/variable dimness). For example, an optical shutter, such as a liquid crystal shutter, may have pixels that have at least one gray state in addition to the dark and light states.
In some embodiments, light-blocking shutter 110 may include guest-host liquid crystal shutters that include, for example, a liquid crystal layer (such as a nematic liquid crystal) incorporating a dye (e.g., including dye molecules oriented within a nematic liquid crystal). In some embodiments, a nematic liquid crystal molecule may include a dye moiety. In some embodiments, a liquid crystal layer may include a polymer dispersed liquid crystal (PDLC, where the liquid crystal may be a nematic liquid crystal) or a polymer-stabilized liquid crystal, such as a polymer stabilized cholesteric texture.
According to some embodiments, light emitted by a display element may be sufficiently blocked by obscuring only a portion of a region overlapping the display element.
Because the light beam may not be projected from regions outside of display region 406, the light leakage may be effectively blocked by obscuring a limited region overlapping display region 406 and a region extending around display region 406, while still allowing light to pass through other portions of the display element 404. For example, as shown in
In some embodiments, light-blocking region 408 may additionally or alternatively be positioned to block light reflected from a portion of user's eye 202 such that the light spot is not visible to external users. Light-blocking shutter 402 may be configured to only block light at light-blocking region 408. In additional embodiments, light-blocking shutter 402 may be capable of blocking light from other regions outside of light-blocking region 408 such that as the user moves their eye, which may result in a change in the position of the display region of display element 404 (e.g., in response to the detected movement of user's eye 202 by an eye-tracking system), then the portion of light-blocking shutter 402 that is blocked may be shifted correspondingly. Additionally, in some embodiments, the one or more portions of light-blocking shutter 402 may be blocked based on detected positions of external viewers (e.g., based on detection by external sensors of the system). In at least one example, portions of light-blocking shutter 402 that block light may additionally or alternatively be determined based on a position or orientation of the wearer's head so as to ensure that light from display element 404 is not visible.
The shutter driving waveform 604 and display driving waveform 608 may be driven in the active state for small proportions of each cycle and may be driven according to a repeating waveform, with cycle periods CP1 and CP2 being illustrated in
Additionally, as shown in
In an embodiment, the timing of the display pulse 606 and the shutter pulse 602 will be controlled by the display-driving subsystem 104 with the shutter-driving subsystem 112 providing the necessary voltage/AC signal to control the light-blocking shutter 110 (see
In some embodiments, the duration of each display pulse 606 may vary in each cycle period (CP1, CP2, etc.). Similarly, the duration of each shutter pulse 602 may vary in each cycle period.
The active and inactive display intervals 626 and 628 may each have any suitable duration, such as a duration of from approximately one nanosecond to one or more milliseconds. For example, pulse-code modulation (PCM) and/or pulse-width modulation (PWM) driven superluminescent light emitting diodes (SLEDS) for LCOS and/or fLCOS displays may have active display intervals 626 as short as approximately one nanosecond and inactive display intervals 628 of from approximately one nanosecond to several nanoseconds. The pulsed sub-period 622 may include a plurality of active display intervals 626 having the same pulse widths or different pulse widths.
Each of light-blocking regions 802 and 806 of light-blocking shutters 800 and 810 may be an optical shutter element (e.g., a liquid-crystal shutter element or mechanical shutter element) capable of blocking or controlling light transmittance. An optical shutter element may have pixels that have dark or light states. In a dark state, the pixel may be generally opaque and block light from a portion of the local environment from reaching the user's view. In a light state, the pixel may be relatively transparent, allowing light from the local environment to pass through the pixel and reach the user's view. In some examples, a blocking element may include pixels with multiple available state options (for example, clear/dim/opaque or clear/variable dimness). For example, an optical shutter element, such as a liquid crystal shutter element, may have pixels that have at least one gray state in addition to the dark and light states.
In some embodiments, light-blocking shutters 800 and 810 may include guest-host liquid crystal shutters that include, for example, a liquid crystal layer (such as a nematic liquid crystal) incorporating a dye (e.g., including dye molecules oriented within a nematic liquid crystal). In some embodiments, a nematic liquid crystal molecule may include a dye moiety. In some embodiments, a liquid crystal layer may include a polymer dispersed liquid crystal (PDLC, where the liquid crystal may be a nematic liquid crystal) or a polymer-stabilized liquid crystal, such as a polymer stabilized cholesteric texture. In various embodiments, light-blocking shutters 800 and 810 may include polarizer-based liquid crystal shutters that are capable of switching at the disclosed operating frequencies (e.g., in a range of about 60 Hz to 2000 Hz). In some embodiments, light-blocking shutters 800 and 810 may include electrostatically activated micro-blinds based devices.
In some examples, electroactive region 900 may include a single pair of electrodes, such as electrodes 909 and 911, or multiple pairs of electrodes (not shown) that may be patterned across a region of an electroactive element (e.g., similar to electroactive region 900). In particular, electroactive region 900 may correspond to an aperture associated with the electroactive region used in a head-mounted display (“HMD”). This may be done in order to create spatially controllable light scattering (e.g., scattering that is implemented with differing magnitudes at different regions of electroactive region 900). Further, while electrodes may serve to impart an electric field onto electroactive region 900 and modify the electroactive region properties as described herein, in other examples electroactive region 900 may be switched with at least a partially non-electrical technique. In particular, electroactive region 900 may be switched based on a mechanical compression of electroactive region 900 or may be switched using acoustic waves that may propagate through electroactive region 900.
As nanovoids 912 are compressed, the size of nanovoids 912 may become several orders of magnitude smaller than wavelengths of light incident on electroactive region 900. In this case, the amount of light scattered from electroactive region 900 due to nanovoids 912 may be minimized during compression. Further, the interaction of electromagnetic fields with nanovoids 912 having a size that is substantially smaller than wavelengths of incident light may lead to Rayleigh scattering of the incident light from nanovoids 912. As the size of nanovoids 912 in electroactive region 900 increases, the amount of scattered light from electroactive region 900 may also increase. If nanovoids 912 are in the same or substantially the same size range as wavelengths of incident light, a Mie scattering mechanism may describe the scattering of the light from nanovoids 912.
In some examples, when there is little or no electric field applied to electroactive region 900 by electrodes 909 and 911, the size of nanovoids 912 may be less than about 26000 nm and greater than about 400 nm. As noted, the application of electric field across electroactive region 900 may result in a mechanical compression of electroactive region 900 from a prior uncompressed state. The magnitude of electric fields across the electrodes 909 and 911 can be selected to change the size and density of the nanovoids 912 to achieve a desired amount of transparency between the compressed and uncompressed states. In a compressed state, nanovoids 912 of electroactive region 900 may be reduced, in the compressed state, to sizes of from about 0.1 nm to about 50 nm based on the magnitude of the applied electrical field.
In particular, if a beam of visible wavelength light (e.g., incident light 1018) is directed towards electroactive region 900 when electroactive region 900 is in a compressed state, electroactive region 900 may scatter relatively little light because the size of nanovoids 912 may be much smaller than the wavelength of light. Accordingly, electroactive region 900 may be transparent in the visible portion of the electromagnetic spectrum. In another embodiment, by modulating the electrical field applied across electroactive region 900 as a function of time, electroactive region 900 can serve as a switchable component, such as, for example, a switchable screen, a time-varying beam-block, and/or a time-varying intensity modulator for light in the visible and near-infrared portion of the electromagnetic spectrum.
In
The waveguide configuration 1400 may include a waveguide 1402, which may be formed from a glass or plastic material. The waveguide 1402 may include a coupling area 1404 and a decoupling area formed by decoupling elements 1406A on a top surface 1408A and decoupling elements 1406B on a bottom surface 1408B, in some embodiments. The area within the waveguide 1402 (in-between the decoupling elements 1406A and 1406B) may be considered a propagation area 1410, in which light images received from the projector device 1450 and coupled into the waveguide 1402 by coupling elements included in the coupling area 1404 may propagate laterally within the waveguide 1402.
The coupling area 1404 may include a coupling element 1412 configured and dimensioned to couple light of a predetermined wavelength, e.g., red, green, or blue light. When a white light emitter array is included in the projector device 1450, the portion of the white light that falls in the predetermined wavelength may be coupled by each of the coupling elements 1412. In some embodiments, the coupling elements 1412 may be gratings, such as Bragg gratings, dimensioned to couple a predetermined wavelength of light. In some examples, the gratings of each coupling element 1412 may exhibit a separation distance between gratings associated with the predetermined wavelength of light that the particular coupling element 1412 is to couple into the waveguide 1402, resulting in different grating separation distances for each coupling element 1412. Accordingly, each coupling element 1412 may couple a limited portion of the white light from the white light emitter array, when included. In other examples, the grating separation distance may be the same for each coupling element 1412. In some examples, coupling element 1412 may be or include a multiplexed coupler.
As shown in
The portion of light that is not projected out of the waveguide 1402 by the decoupling element 1406A may be reflected off the decoupling element 1406B. The decoupling element 1406B may reflect all incident light back toward the decoupling element 1406A, as depicted. Accordingly, the waveguide 1402 may combine the red image 1420A, the blue image 1420B, and the green image 1420C into a polychromatic image instance, which may be referred to as a pupil replication 1422. The polychromatic pupil replication 1422 may be projected toward a viewing region that includes a user's eye (e.g., user's eye 202 in
As shown in
In some embodiments, movement of the projector assembly 1524, or a portion thereof, may be substantially constrained to directions parallel to a major surface of the optical combiner 1520, such as along the X- and Y-directions depicted in
In some embodiments, cameras 1602, 1604, 1606, 1608, and/or 1610 may each be mounted flush with surfaces of head-mounted display 1600 (rather than protruding from head-mounted display 1600). In one embodiment, camera 1602 may be located behind camera 1604 (relative to the front of head-mounted display 1600) and/or may be angled at a downward angle, such as 45° downward. In some embodiments, camera 1602 may be located at a different downward angle, such as 30°, 60°, or any other appropriate angle. Similarly, camera 1608 may be located behind camera 1606 and/or may be angled at a downward angle. In some embodiments, cameras 1604, 1606, and 1610 may all be mounted on the same surface of the head-mounted display. In other embodiments, cameras 1604 and/or 1606 may be mounted on one front surface of the head-mounted display while camera 1610 may be mounted on a separate front surface of the head-mounted display. In some examples, cameras 1602, 1604, 1606, 1608, and/or 1610 may be used to capture a digital view (e.g., a 2D or 3D view) of a user's real-world environment. In some examples, display surface 1612 may be opaque to visible light, preventing the user from a direct view of the real-world environment. Alternatively, display surface 1612 may be at least partially transparent to visible light (i.e., light having a wavelength of about 380 nm to about 750 nm) and may allow light to pass through to an underlying display. In some embodiments, display surface 1612 may be a display or a light-blocking device.
In some examples, depth-sensing subsystem 1702 may capture data describing depth information characterizing a real-world environment surrounding some or all of head-mounted display system 1700. In some embodiments, depth-sensing subsystem 1702 may characterize a position or velocity of head-mounted display system 1700 and/or objects within the real-world environment. Depth-sensing subsystem 1702 may compute a depth map using collected data (e.g., based on a captured light according to one or more computer-vision schemes or algorithms, by processing a portion of a structured light pattern, by time-of-flight (ToF) imaging, simultaneous localization and mapping (SLAM), etc.). In some examples, the depth maps may be used to generate a model of the real-world environment surrounding head-mounted display system 1700. Accordingly, depth-sensing subsystem 1702 may be referred to as a localization and modeling subsystem or may be a part of such a subsystem.
In some examples, image-capturing subsystem 1704 may include one or more optical image sensors or cameras 1705 that capture and collect image data from a user's real-world environment. In some embodiments, cameras 1705 may provide stereoscopic views of a user's real-world environment that may be used by tracking subsystem 1710 to identify and track real-world objects. In some embodiments, the image data may be processed by tracking subsystem 1710 or another component of image-capturing subsystem 1704 to generate a three-dimensional model of the user's real-world environment and the objects contained therein. In some examples, image-capturing subsystem 1704 may include simultaneous localization and mapping (SLAM) cameras or other cameras that include a wide-angle lens system that captures a wider field-of-view than may be captured by the eyes of the user.
In some examples, IMU 1708 may generate data indicating a position and/or orientation of head-mounted display system 1700 based on measurement signals received from one or more of sensors 1706 and from depth information received from depth-sensing subsystem 1702 and/or image-capturing subsystem 1704. For example, sensors 1706 may generate one or more measurement signals in response to motion of head-mounted display system 1700. Examples of sensors 1706 include one or more accelerometers, one or more gyroscopes, one or more magnetometers, another suitable type of sensor that detects motion, a type of sensor used for error correction of IMU 1708, or some combination thereof. Based on the one or more measurement signals from one or more of position sensors 1706, IMU 1708 may generate data indicating an estimated current position, elevation, and/or orientation of head-mounted display system 1700 relative to an initial position and/or orientation of head-mounted display system 1700. For example, sensors 1706 may include multiple accelerometers to measure translational motion (forward/back, up/down, left/right) and multiple gyroscopes to measure rotational motion (e.g., pitch, yaw, roll). As described herein, image-capturing subsystem 1704 and/or depth-sensing subsystem 1702 may generate data indicating an estimated current position and/or orientation of head-mounted display system 1700 relative to the real-world environment in which head-mounted display is used.
Tracking subsystem 1710 may include one or more processing devices or physical processors that (1) identifies and tracks objects in a user's real-world environment and (2) detects pass-through triggering events involving the objects in accordance with information received from one or more of depth-sensing subsystem 1702, image-capturing subsystem 1704, sensors 1706, IMU 1708, and gaze-estimation subsystem 1712. In some embodiments, tracking subsystem 1710 may monitor real-world objects that can be observed by depth-sensing subsystem 1702, image-capturing subsystem 1704, and/or by another system. Tracking subsystem 1710 may also receive information from one or more eye-tracking cameras included in some embodiments of shutter-driving subsystem 1701 to track a user's gaze. In some examples, a user's gaze angle may inform shutter-driving subsystem 1701 of which region of a light-blocking device (e.g., light-blocking shutter 402 in
As illustrated in
At step 1820, one or more of the systems described herein may block light transmission through a light-blocking shutter during the active display intervals. For example, shutter-driving subsystem 112 may block light transmission through light-blocking shutter 110 during the active display intervals, as shown, for example, in
In some embodiments, blocking light transmission through the light-blocking shutter during the active display intervals may include blocking light transmission through the light-blocking shutter during the entirety of each of the active display intervals. In various examples, the method may further include blocking light transmission through the light-blocking shutter during a portion of at least one of an inactive display interval immediately prior to or immediately following each of the active display intervals (see, e.g.,
At step 1830, one or more of the systems described herein may allow light transmission through the light-blocking shutter during at least a portion of inactive display intervals when the display light is not emitted from the display element. For example, shutter-driving subsystem 112 may allow light transmission through the light-blocking shutter 110 during at least a portion of inactive display intervals when the display light is not emitted from the display element, as shown, for example, in
The preceding description has been provided to enable others skilled in the art to best utilize various aspects of the exemplary embodiments disclosed herein. This exemplary description is not intended to be exhaustive or to be limited to any precise form disclosed. Many modifications and variations are possible without departing from the spirit and scope of the instant disclosure. The embodiments disclosed herein should be considered in all respects illustrative and not restrictive. Reference should be made to the appended claims and their equivalents in determining the scope of the instant disclosure. By way of example and not limitation, some embodiments of the present disclosure include the following:
Example 1: A display system, comprising: (1) a display element that is transparent and comprises a first surface facing a viewing region and a second surface opposite the first surface, the second surface facing an external environment; (2) a light-blocking shutter that faces and overlaps at least a portion of the second surface of the display element; (3) a display-driving subsystem that displays images visible to a user at the viewing region via display light emitted from the first surface of the display element during active display intervals; and (4) a shutter-driving subsystem that alternately switches the state of the light-blocking shutter between a blocking state in which light transmission through the light-blocking shutter is blocked during the active display intervals and a pass-through state in which light transmission through the light-blocking shutter is enabled during at least a portion of inactive display intervals when the display light is not emitted from the display element.
Example 2: The display system of example 1, wherein the shutter-driving subsystem maintains the light-blocking shutter in the blocking state during the entirety of each of the active display intervals.
Example 3: The display system of any of examples 1 and 2, wherein the shutter-driving subsystem maintains the light-blocking shutter in the blocking state during a portion of at least one of an inactive display interval immediately prior to or immediately following each of the active display intervals.
Example 4: The display system of any of examples 1-3, wherein the display-driving subsystem displays the images during consecutive cycle periods each comprising one or more active display intervals and one or more inactive display intervals.
Example 5: The display system of any of examples 1-4, wherein each of the one or more active display intervals of each cycle period has a duration of from approximately 0.1% to approximately 50% of the total duration of the cycle period.
Example 6: The display system of any of example 1-5, wherein, for each cycle period: (1) the one or more active display intervals comprise a plurality of active display intervals pulsed during a burst sub-period; and (2) the one or more inactive display intervals comprise a plurality of inactive display intervals alternating with the active display intervals during the pulsed cycle sub-period and an extended inactive display interval occurring during an OFF sub-period.
Example 7: The display system of any of examples 1-6, wherein the pulsed sub-period has a duration of from approximately 0.00001% to approximately 30% of the total duration of the cycle period.
Example 8: The display system of any of examples 1-7, wherein the shutter-driving subsystem maintains the light-blocking shutter in the blocking state during the entirety of the pulsed sub-period.
Example 9: The display system of any of examples 1-8, wherein the light-blocking shutter comprises one light-blocking region or multiple light-blocking regions that are individually operable in response to signals from the shutter-driving subsystem.
Example 10: The display system of any of examples 1-9, wherein the light-blocking shutter comprises at least one of a liquid crystal shutter or a mechanical shutter, the light blocking being accomplished by absorption, reflection or scattering of the light.
Example 11: The display system of any of examples 1-10, wherein the light-blocking shutter further comprises an additional shutter layer comprising an electrochromic material that is selectively transmissive.
Example 12: The display system of any of examples 1-11, wherein the display element comprises at least one of a waveguide or an optical combiner.
Example 13: The display system of any of examples 1-12, wherein: (1) the display element further comprises a pass-through region through which light from the external environment is transmitted at least when the light-blocking shutter is in the pass-through state; and (2) the shutter-driving subsystem drives the light-blocking shutter to block light transmission through a light-blocking region of the light-blocking shutter overlapping at least a portion of the pass-through region of the display element.
Example 14: The display system of any of examples 1-13, wherein the light-blocking region substantially or entirely overlaps the pass-through region of the display element.
Example 15: The display system of any of examples 1-14, wherein (1) the light-blocking region overlaps a selected portion of the pass-through region of the display element so as to block externally-directed light from at least one of a light-emitting region of the display element or a portion of the user's face illuminated by the display light; and (2) the light-blocking region does not overlap a surrounding portion of the pass-through region of the display element outside of the selected portion.
Example 16: The display system of any of examples 1-15, wherein the shutter-driving subsystem drives the light-blocking shutter to block light transmission through a surrounding region of the light-blocking shutter outside of the light-blocking region to a lesser extent than light-blocking region.
Example 17: A head-mounted display device, comprising: (1) an optical subsystem that provides a user with composite views of displayed images and an external environment, the optical subsystem comprising: (a) a transparent display element that comprises a first surface facing a viewing region and an opposite second surface facing the external environment, and (b) a light-blocking shutter that faces and overlaps at least a portion of the second surface of the display element; (2) a display-driving subsystem that displays images visible to the user at the viewing region via display light emitted from the first surface of the display element during active display intervals; and (3) a shutter-driving subsystem that alternately switches the state of the light-blocking shutter between a blocking state in which light transmission through the light-blocking shutter is blocked during the active display intervals and a pass-through state in which light transmission through the light-blocking shutter is enabled during at least a portion of inactive display intervals when the display light is not emitted.
Example 18: A computer-implemented method, comprising: (1) displaying images visible to a user at a viewing region via display light emitted during active display intervals from a first surface of a display element facing the viewing region, wherein the display element is transparent and further comprises a second surface opposite the first surface, the second surface facing an external environment; (2) blocking light transmission through a light-blocking shutter during the active display intervals, wherein the light-blocking shutter faces and overlaps at least a portion of the second surface of the display element; and (3) allowing light transmission through the light-blocking shutter during at least a portion of inactive display intervals when the display light is not emitted from the display element.
Example 19: The computer-implemented method of example 18, wherein blocking light transmission through the light-blocking shutter during the active display intervals comprises blocking light transmission through the light-blocking shutter during the entirety of each of the active display intervals.
Example 20: The computer-implemented method of any of examples 18 and 19, further comprising blocking light transmission through the light-blocking shutter during a portion of at least one of an inactive display interval immediately prior to or immediately following each of the active display intervals.
Embodiments of the present disclosure may include or be implemented in conjunction with various types of artificial-reality systems. Artificial reality is a form of reality that has been adjusted in some manner before presentation to a user, which may include, e.g., a virtual reality, an augmented reality, a mixed reality, a hybrid reality, or some combination and/or derivative thereof. Artificial-reality content may include completely generated content or generated content combined with captured (e.g., real-world) content. The artificial-reality content may include video, audio, haptic feedback, or some combination thereof, any of which may be presented in a single channel or in multiple channels (such as stereo video that produces a three-dimensional (3D) effect to the viewer). Additionally, in some embodiments, artificial reality may also be associated with applications, products, accessories, services, or some combination thereof, that are used to, e.g., create content in an artificial reality and/or are otherwise used in (e.g., to perform activities in) an artificial reality.
Artificial-reality systems may be implemented in a variety of different form factors and configurations. Some artificial-reality systems may be designed to work without near-eye displays (NEDs), an example of which is augmented-reality system 1900 in
Turning to
As shown, augmented-reality system 1900 may not necessarily include an NED positioned in front of a user's eyes. Augmented-reality systems without NEDs may take a variety of forms, such as head bands, hats, hair bands, belts, watches, wrist bands, ankle bands, rings, neckbands, necklaces, chest bands, eyewear frames, and/or any other suitable type or form of apparatus. While augmented-reality system 1900 may not include an NED, augmented-reality system 1900 may include other types of screens or visual feedback devices (e.g., a display screen integrated into a side of frame 1902).
The embodiments discussed in this disclosure may also be implemented in augmented-reality systems that include one or more NEDs. For example, as shown in
In some embodiments, augmented-reality system 2000 may include one or more sensors, such as sensor 2040. Sensor 2040 may generate measurement signals in response to motion of augmented-reality system 2000 and may be located on substantially any portion of frame 2010. Sensor 2040 may represent a position sensor, an inertial measurement unit (IMU), a depth camera assembly, or any combination thereof. In some embodiments, augmented-reality system 2000 may or may not include sensor 2040 or may include more than one sensor. In embodiments in which sensor 2040 includes an IMU, the IMU may generate calibration data based on measurement signals from sensor 2040. Examples of sensor 2040 may include, without limitation, accelerometers, gyroscopes, magnetometers, other suitable types of sensors that detect motion, sensors used for error correction of the IMU, or some combination thereof.
Augmented-reality system 2000 may also include a microphone array with a plurality of acoustic transducers 2020(A)-2020(J), referred to collectively as acoustic transducers 2020. Acoustic transducers 2020 may be transducers that detect air pressure variations induced by sound waves. Each acoustic transducer 2020 may be configured to detect sound and convert the detected sound into an electronic format (e.g., an analog or digital format). The microphone array in
In some embodiments, one or more of acoustic transducers 2020(A)-(F) may be used as output transducers (e.g., speakers). For example, acoustic transducers 2020(A) and/or 2020(B) may be earbuds or any other suitable type of headphone or speaker.
The configuration of acoustic transducers 2020 of the microphone array may vary. While augmented-reality system 2000 is shown in
Acoustic transducers 2020(A) and 2020(B) may be positioned on different parts of the user's ear, such as behind the pinna or within the auricle or fossa. Or, there may be additional acoustic transducers 2020 on or surrounding the ear in addition to acoustic transducers 2020 inside the ear canal. Having an acoustic transducer 2020 positioned next to an ear canal of a user may enable the microphone array to collect information on how sounds arrive at the ear canal. By positioning at least two of acoustic transducers 2020 on either side of a user's head (e.g., as binaural microphones), augmented-reality device 2000 may simulate binaural hearing and capture a 3D stereo sound field around about a user's head. In some embodiments, acoustic transducers 2020(A) and 2020(B) may be connected to augmented-reality system 2000 via a wired connection 2030, and in other embodiments, acoustic transducers 2020(A) and 2020(B) may be connected to augmented-reality system 2000 via a wireless connection (e.g., a Bluetooth connection). In still other embodiments, acoustic transducers 2020(A) and 2020(B) may not be used at all in conjunction with augmented-reality system 2000.
Acoustic transducers 2020 on frame 2010 may be positioned along the length of the temples, across the bridge, above or below display devices 2015(A) and 2015(B), or some combination thereof. Acoustic transducers 2020 may be oriented such that the microphone array is able to detect sounds in a wide range of directions surrounding the user wearing the augmented-reality system 2000. In some embodiments, an optimization process may be performed during manufacturing of augmented-reality system 2000 to determine relative positioning of each acoustic transducer 2020 in the microphone array.
In some examples, augmented-reality system 2000 may include or be connected to an external device (e.g., a paired device), such as neckband 2005. Neckband 2005 generally represents any type or form of paired device. Thus, the following discussion of neckband 2005 may also apply to various other paired devices, such as charging cases, smart watches, smart phones, wrist bands, other wearable devices, hand-held controllers, tablet computers, laptop computers and other external compute devices, etc.
As shown, neckband 2005 may be coupled to eyewear device 2002 via one or more connectors. The connectors may be wired or wireless and may include electrical and/or non-electrical (e.g., structural) components. In some cases, eyewear device 2002 and neckband 2005 may operate independently without any wired or wireless connection between them. While
Pairing external devices, such as neckband 2005, with augmented-reality eyewear devices may enable the eyewear devices to achieve the form factor of a pair of glasses while still providing sufficient battery and computation power for expanded capabilities. Some or all of the battery power, computational resources, and/or additional features of augmented-reality system 2000 may be provided by a paired device or shared between a paired device and an eyewear device, thus reducing the weight, heat profile, and form factor of the eyewear device overall while still retaining desired functionality. For example, neckband 2005 may allow components that would otherwise be included on an eyewear device to be included in neckband 2005 since users may tolerate a heavier weight load on their shoulders than they would tolerate on their heads. Neckband 2005 may also have a larger surface area over which to diffuse and disperse heat to the ambient environment. Thus, neckband 2005 may allow for greater battery and computation capacity than might otherwise have been possible on a stand-alone eyewear device. Since weight carried in neckband 2005 may be less invasive to a user than weight carried in eyewear device 2002, a user may tolerate wearing a lighter eyewear device and carrying or wearing the paired device for greater lengths of time than a user would tolerate wearing a heavy standalone eyewear device, thereby enabling users to more fully incorporate artificial-reality environments into their day-to-day activities.
Neckband 2005 may be communicatively coupled with eyewear device 2002 and/or to other devices. These other devices may provide certain functions (e.g., tracking, localizing, depth mapping, processing, storage, etc.) to augmented-reality system 2000. In the embodiment of
Acoustic transducers 2020(1) and 2020(J) of neckband 2005 may be configured to detect sound and convert the detected sound into an electronic format (analog or digital). In the embodiment of
Controller 2025 of neckband 2005 may process information generated by the sensors on neckband 2005 and/or augmented-reality system 2000. For example, controller 2025 may process information from the microphone array that describes sounds detected by the microphone array. For each detected sound, controller 2025 may perform a direction-of-arrival (DOA) estimation to estimate a direction from which the detected sound arrived at the microphone array. As the microphone array detects sounds, controller 2025 may populate an audio data set with the information. In embodiments in which augmented-reality system 2000 includes an inertial measurement unit, controller 2025 may compute all inertial and spatial calculations from the IMU located on eyewear device 2002. A connector may convey information between augmented-reality system 2000 and neckband 2005 and between augmented-reality system 2000 and controller 2025. The information may be in the form of optical data, electrical data, wireless data, or any other transmittable data form. Moving the processing of information generated by augmented-reality system 2000 to neckband 2005 may reduce weight and heat in eyewear device 2002, making it more comfortable to the user.
Power source 2035 in neckband 2005 may provide power to eyewear device 2002 and/or to neckband 2005. Power source 2035 may include, without limitation, lithium ion batteries, lithium-polymer batteries, primary lithium batteries, alkaline batteries, or any other form of power storage. In some cases, power source 2035 may be a wired power source. Including power source 2035 on neckband 2005 instead of on eyewear device 2002 may help better distribute the weight and heat generated by power source 2035.
As noted, some artificial-reality systems may, instead of blending an artificial reality with actual reality, substantially replace one or more of a user's sensory perceptions of the real world with a virtual experience. One example of this type of system is a head-worn display system, such as virtual-reality system 2100 in
Artificial-reality systems may include a variety of types of visual feedback mechanisms. For example, display devices in augmented-reality system 2000 and/or virtual-reality system 2100 may include one or more liquid crystal displays (LCDs), light emitting diode (LED) displays, organic LED (OLED) displays, and/or any other suitable type of display screen. Artificial-reality systems may include a single display screen for both eyes or may provide a display screen for each eye, which may allow for additional flexibility for varifocal adjustments or for correcting a user's refractive error. Some artificial-reality systems may also include optical subsystems having one or more lenses (e.g., conventional concave or convex lenses, Fresnel lenses, adjustable liquid lenses, etc.) through which a user may view a display screen.
In addition to or instead of using display screens, some artificial-reality systems may include one or more projection systems. For example, display devices in augmented-reality system 2000 and/or virtual-reality system 2100 may include micro-LED projectors that project light (using, e.g., a waveguide) into display devices, such as clear combiner lenses that allow ambient light to pass through. The display devices may refract the projected light toward a user's pupil and may enable a user to simultaneously view both artificial-reality content and the real world. Artificial-reality systems may also be configured with any other suitable type or form of image projection system.
Artificial-reality systems may also include various types of computer vision components and subsystems. For example, augmented-reality system 1900, augmented-reality system 2000, and/or virtual-reality system 2100 may include one or more optical sensors, such as two-dimensional (2D) or 3D cameras, time-of-flight depth sensors, single-beam or sweeping laser rangefinders, 3D LiDAR sensors, and/or any other suitable type or form of optical sensor. An artificial-reality system may process data from one or more of these sensors to identify a location of a user, to map the real world, to provide a user with context about real-world surroundings, and/or to perform a variety of other functions.
Artificial-reality systems may also include one or more input and/or output audio transducers. In the examples shown in
While not shown in
By providing haptic sensations, audible content, and/or visual content, artificial-reality systems may create an entire virtual experience or enhance a user's real-world experience in a variety of contexts and environments. For instance, artificial-reality systems may assist or extend a user's perception, memory, or cognition within a particular environment. Some systems may enhance a user's interactions with other people in the real world or may enable more immersive interactions with other people in a virtual world. Artificial-reality systems may also be used for educational purposes (e.g., for teaching or training in schools, hospitals, government organizations, military organizations, business enterprises, etc.), entertainment purposes (e.g., for playing video games, listening to music, watching video content, etc.), and/or for accessibility purposes (e.g., as hearing aids, visuals aids, etc.). The embodiments disclosed herein may enable or enhance a user's artificial-reality experience in one or more of these contexts and environments and/or in other contexts and environments.
Some augmented-reality systems may map a user's and/or device's environment using techniques referred to as “simultaneous location and mapping” (SLAM). SLAM mapping and location identifying techniques may involve a variety of hardware and software tools that can create or update a map of an environment while simultaneously keeping track of a user's location within the mapped environment. SLAM may use many different types of sensors to create a map and determine a user's position within the map.
SLAM techniques may, for example, implement optical sensors to determine a user's location. Radios including WiFi, Bluetooth, global positioning system (GPS), cellular or other communication devices may be also used to determine a user's location relative to a radio transceiver or group of transceivers (e.g., a WiFi router or group of GPS satellites). Acoustic sensors such as microphone arrays or 2D or 3D sonar sensors may also be used to determine a user's location within an environment. Augmented-reality and virtual-reality devices (such as systems 1900, 2000, and 2100 of
As detailed above, the computing devices and systems described and/or illustrated herein broadly represent any type or form of computing device or system capable of executing computer-readable instructions, such as those contained within the subsystems and/or modules described herein. In their most basic configuration, these computing device(s) may each include at least one memory device and at least one physical processor.
In some examples, the term “memory device” generally refers to any type or form of volatile or non-volatile storage device or medium capable of storing data and/or computer-readable instructions. In one example, a memory device may store, load, and/or maintain one or more of the modules described herein. Examples of memory devices include, without limitation, Random Access Memory (RAM), Read Only Memory (ROM), flash memory, Hard Disk Drives (HDDs), Solid-State Drives (SSDs), optical disk drives, caches, variations or combinations of one or more of the same, or any other suitable storage memory.
In some examples, the term “physical processor” generally refers to any type or form of hardware-implemented processing unit capable of interpreting and/or executing computer-readable instructions. In one example, a physical processor may access and/or modify one or more modules stored in the above-described memory device. Examples of physical processors include, without limitation, microprocessors, microcontrollers, Central Processing Units (CPUs), Field-Programmable Gate Arrays (FPGAs) that implement softcore processors, Application-Specific Integrated Circuits (ASICs), portions of one or more of the same, variations or combinations of one or more of the same, or any other suitable physical processor.
Although illustrated as separate elements, the subsystems and/or modules described and/or illustrated herein may represent portions of a single module or application. In addition, in certain embodiments one or more of these subsystems and/or modules may represent one or more software applications or programs that, when executed by a computing device, may cause the computing device to perform one or more tasks. For example, one or more of the subsystems and/or modules described and/or illustrated herein may represent subsystems and/or modules stored and configured to run on one or more of the computing devices or systems described and/or illustrated herein. One or more of these subsystems and/or modules may also represent all or portions of one or more special-purpose computers configured to perform one or more tasks.
In addition, one or more of the subsystems and/or modules described herein may transform data, physical devices, and/or representations of physical devices from one form to another. For example, one or more of the subsystems and/or modules recited herein may receive or detect a pass-through event involving a real-world object in a user's real-world environment, transform the pass-through event into light-blocking instructions, output a result of the transformation to a light-blocking device that includes a plurality of blocking regions that selectively blocks or transmits light according to the light-blocking instructions, and use the result of the transformation to selectively pass a view of the real-world object through a display device to the user. Additionally or alternatively, one or more of the subsystems and/or modules recited herein may transform a processor, volatile memory, non-volatile memory, and/or any other portion of a physical computing device from one form to another by executing on the computing device, storing data on the computing device, and/or otherwise interacting with the computing device.
In some embodiments, the term “computer-readable medium” generally refers to any form of device, carrier, or medium capable of storing or carrying computer-readable instructions. Examples of computer-readable media include, without limitation, transmission-type media, such as carrier waves, and non-transitory-type media, such as magnetic-storage media (e.g., hard disk drives, tape drives, and floppy disks), optical-storage media (e.g., Compact Disks (CDs), Digital Video Disks (DVDs), and BLU-RAY disks), electronic-storage media (e.g., solid-state drives and flash media), and other distribution systems.
Embodiments of the instant disclosure may include or be implemented in conjunction with an artificial reality system. Artificial reality is a form of reality that has been adjusted in some manner before presentation to a user, which may include, e.g., a virtual reality, an augmented reality, a mixed reality, a hybrid reality, or some combination and/or derivatives thereof. Artificial reality content may include completely generated content or generated content combined with captured (e.g., real-world) content. The artificial reality content may include video, audio, haptic feedback, or some combination thereof, any of which may be presented in a single channel or in multiple channels (such as stereo video that produces a three-dimensional effect to the viewer). Additionally, in some embodiments, artificial reality may also be associated with applications, products, accessories, services, or some combination thereof, that are used to, e.g., create content in an artificial reality and/or are otherwise used in (e.g., perform activities in) an artificial reality. The artificial reality system that provides the artificial reality content may be implemented on various platforms, including a head-mounted display (HMD) connected to a host computer system, a standalone HMD, a mobile device or computing system, or any other hardware platform capable of providing artificial reality content to one or more viewers.
The process parameters and sequence of the steps described and/or illustrated herein are given by way of example only and can be varied as desired. For example, while the steps illustrated and/or described herein may be shown or discussed in a particular order, these steps do not necessarily need to be performed in the order illustrated or discussed. The various exemplary methods described and/or illustrated herein may also omit one or more of the steps described or illustrated herein or include additional steps in addition to those disclosed.
Unless otherwise noted, the terms “connected to” and “coupled to” (and their derivatives), as used in the specification and claims, are to be construed as permitting both direct and indirect (i.e., via other elements or components) connection. In addition, the terms “a” or “an,” as used in the specification and claims, are to be construed as meaning “at least one of.” Finally, for ease of use, the terms “including” and “having” (and their derivatives), as used in the specification and claims, are interchangeable with and have the same meaning as the word “comprising.”
Valentine, Gareth, Sears, Jasmine Soria
Patent | Priority | Assignee | Title |
11385463, | Sep 18 2019 | Apple Inc.; Apple Inc | Display device using active shutter |
11650447, | Mar 19 2019 | Toppan Printing Co., Ltd. | Light control sheet, light control device and method of controlling the light control sheet |
11662586, | Mar 06 2020 | Magic Leap, Inc. | Angularly selective attenuation of light transmission artifacts in wearable displays |
11693252, | Oct 17 2019 | MAGIC LEAP, INC | Attenuation of light transmission artifacts in wearable displays |
11950022, | Apr 24 2020 | Apple Inc. | Head-mounted devices with forward facing cameras |
11992373, | Dec 10 2019 | Globus Medical, Inc | Augmented reality headset with varied opacity for navigated robotic surgery |
ER4707, |
Patent | Priority | Assignee | Title |
10847075, | Apr 10 2019 | META PLATFORMS TECHNOLOGIES, LLC | Error correction for display device |
5579161, | Nov 12 1992 | Olympus Optical Co., Ltd. | Image display apparatus |
6055110, | Jul 02 1996 | Compound Photonics Limited | Compact display system controlled by eye position sensor system |
6215532, | Jul 27 1998 | MIXED REALITY SYSTEMS LABORATORY INC | Image observing apparatus for observing outside information superposed with a display image |
6396639, | Feb 04 1999 | Olympus Corporation | Viewing optical system and image display apparatus using the same |
6518939, | Nov 08 1996 | Olympus Optical Co., Ltd. | Image observation apparatus |
6760169, | May 07 1997 | Olympus Corporation | Prism optical element, image observation apparatus and image display apparatus |
8687661, | Apr 13 2012 | Coherent, Inc | Pulsed CO2 laser output-pulse shape and power control |
9989767, | Apr 17 2014 | tooz technologies GmbH | Spectacle lens for a display device that can be fitted on the head of a user and generates an image, and display device with such a spectacle lens |
20060033992, | |||
20060250322, | |||
20070086495, | |||
20120050143, | |||
20120062444, | |||
20120119978, | |||
20120242716, | |||
20130127980, | |||
20130285885, | |||
20130286053, | |||
20140218272, | |||
20150178939, | |||
20150243068, | |||
20150260992, | |||
20160154244, | |||
20160187666, | |||
20160320620, | |||
20160360074, | |||
20170032729, | |||
20170115395, | |||
20170192237, | |||
20170199396, | |||
20170256095, | |||
20170323615, | |||
20180182354, | |||
20180286320, | |||
20180348524, | |||
20190041642, | |||
20190227321, | |||
20190265476, | |||
20190273911, | |||
20190372306, | |||
20200029032, | |||
20200051483, | |||
20200158943, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Jul 29 2019 | Facebook Technologies, LLC | (assignment on the face of the patent) | / | |||
Jul 30 2019 | SEARS, JASMINE SORIA | Facebook Technologies, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050075 | /0870 | |
Jul 30 2019 | VALENTINE, GARETH | Facebook Technologies, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050075 | /0870 | |
Mar 18 2022 | Facebook Technologies, LLC | META PLATFORMS TECHNOLOGIES, LLC | CHANGE OF NAME SEE DOCUMENT FOR DETAILS | 060199 | /0876 |
Date | Maintenance Fee Events |
Jul 29 2019 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Date | Maintenance Schedule |
Jul 20 2024 | 4 years fee payment window open |
Jan 20 2025 | 6 months grace period start (w surcharge) |
Jul 20 2025 | patent expiry (for year 4) |
Jul 20 2027 | 2 years to revive unintentionally abandoned end. (for year 4) |
Jul 20 2028 | 8 years fee payment window open |
Jan 20 2029 | 6 months grace period start (w surcharge) |
Jul 20 2029 | patent expiry (for year 8) |
Jul 20 2031 | 2 years to revive unintentionally abandoned end. (for year 8) |
Jul 20 2032 | 12 years fee payment window open |
Jan 20 2033 | 6 months grace period start (w surcharge) |
Jul 20 2033 | patent expiry (for year 12) |
Jul 20 2035 | 2 years to revive unintentionally abandoned end. (for year 12) |