A processor or control circuit of an apparatus receives data of an image based on sensing by one or more image sensors. The processor or control circuit also detects a region of interest (roi) in the image. The processor or control circuit then adaptively controls a light projector with respect to projecting light toward the roi.
|
1. A method, comprising:
receiving data of an image based on sensing by one or more image sensors;
detecting a region of interest (roi) in the image; and
adaptively controlling a light projector with respect to projecting light toward the roi,
wherein the detecting of the roi in the image comprises:
detecting a face in the image; and
detecting an eye on the face, and
wherein the adaptively controlling of the light projector with respect to projecting the light toward the roi comprises:
enabling an infrared (ir) projector to project an ir light toward the face responsive to the eye being closed or blinked; and
disabling the ir projector to cease projecting the ir light toward the face responsive to the eye being open.
3. An apparatus, comprising:
a processor or control circuit which, during operation, performs operations comprising:
receiving data of an image based on sensing by one or more image sensors;
detecting a region of interest (roi) in the image; and
adaptively controlling a light projector with respect to projecting light toward the roi,
wherein, in detecting the roi in the image, the processor or control circuit performs operations comprising:
detecting a face in the image; and
detecting an eye on the face, and
wherein, in adaptively controlling the light projector with respect to projecting the light toward the roi, the processor or control circuit performs operations comprising:
enabling an infrared (ir) projector to project an ir light toward the face responsive to the eye being closed or blinked;
controlling the ir projector to reduce an amount of power in projecting the ir light toward the face responsive to an enlargement of a pupil of the eye; and
disabling the ir projector to cease projecting the ir light toward the face responsive to the eye being open.
5. A method, comprising:
receiving data of an image based on sensing by one or more image sensors;
detecting a region of interest (roi) in the image; and
adaptively controlling a light projector with respect to projecting light toward the roi based on a predefined region within the roi by:
enabling the light projector when the one or more image sensors are reading a portion of the image where the roi is located responsive to a first action being performed in the predefined region within the roi, and
disabling the light projector when the one or more image sensors are reading the portion of the image where the roi is located responsive to a second action different from the first action being performed in the predefined region within the roi,
wherein the detecting of the roi in the image comprises detecting a face in the image,
wherein the predefined region within the roi comprises an eye,
wherein the first action performed in the predefined region within the roi comprises the eye being closed, and
wherein the second action performed in the predefined region within the roi comprises the eye being open.
2. The method of
4. The apparatus of
6. The method of
|
The present disclosure is part of a divisional of U.S. patent application Ser. No. 16/354,552, filed 15 Mar. 2019, which claims the priority benefit of U.S. Patent Application No. 62/651,815, filed 3 Apr. 2018. Contents of aforementioned applications are incorporated by reference in their entirety.
The present disclosure is generally related to computer vision and, more particularly, to adaptive infrared (IR) projection control for depth estimation in computer vision.
Unless otherwise indicated herein, approaches described in this section are not prior art to the claims listed below and are not admitted as prior art by inclusion in this section.
Depth estimation in computer vision utilizes images of a given scene captured by two cameras to triangulate and estimate distances. Typically, depth estimation can be achieved with either passive stereo vision or active stereo vision.
The following summary is illustrative only and is not intended to be limiting in any way. That is, the following summary is provided to introduce concepts, highlights, benefits and advantages of the novel and non-obvious techniques described herein. Select implementations are further described below in the detailed description. Thus, the following summary is not intended to identify essential features of the claimed subject matter, nor is it intended for use in determining the scope of the claimed subject matter.
An objective of the present disclosure is to propose schemes, solutions, concepts, designs, methods and apparatuses that enable adaptive IR projection control. It is believed that reduction in power consumption and IR projection time, as well as improved eye safety, may be achieved by implementing various proposed schemes in accordance with the present disclosure.
In one aspect, a method may involve receiving data of an image based on sensing by one or more image sensors. The method may also involve detecting a region of interest (ROI) in the image. The method may further involve adaptively controlling a light projector with respect to projecting light toward the ROI.
In one aspect, an apparatus may include a processor or control circuit which, during operation, may perform operations including: (a) receiving data of an image based on sensing by one or more image sensors; (b) detecting a region of interest (ROI) in the image; and (c) adaptively controlling a light projector with respect to projecting light toward the ROI.
It is noteworthy that, although description provided herein may be in the context of certain EM wave spectra and light-emitting topologies such as IR, the proposed concepts, schemes and any variation(s)/derivative(s) thereof may be implemented in, for and by other EM wave spectra and/or light-emitting technologies such as, for example and without limitation, light-emitting diode (LED), laser, light detection and ranging (LiDAR) and time-of-flight (TOF). Thus, the scope of the present disclosure is not limited to the examples described herein.
The accompanying drawings are included to provide a further understanding of the disclosure and are incorporated in and constitute a part of the present disclosure. The drawings illustrate implementations of the disclosure and, together with the description, serve to explain the principles of the disclosure. It is appreciable that the drawings are not necessarily in scale as some components may be shown to be out of proportion than the size in actual implementation in order to clearly illustrate the concept of the present disclosure.
Detailed embodiments and implementations of the claimed subject matters are disclosed herein. However, it shall be understood that the disclosed embodiments and implementations are merely illustrative of the claimed subject matters which may be embodied in various forms. The present disclosure may, however, be embodied in many different forms and should not be construed as limited to the exemplary embodiments and implementations set forth herein. Rather, these exemplary embodiments and implementations are provided so that description of the present disclosure is thorough and complete and will fully convey the scope of the present disclosure to those skilled in the art. In the description below, details of well-known features and techniques may be omitted to avoid unnecessarily obscuring the presented embodiments and implementations.
Overview
For passive stereo vision, an IR projected projects a structured IR light toward a scene and a single IR camera is utilized to capture an image of the scene. For active stereo vision, two IR cameras are utilized to capture images of the scene.
Part (C) of
Under a proposed scheme in accordance with the present disclosure, the IR projector may be adaptively controlled based on one or more conditions. That is, the IR projector may be adaptively enabled and disabled, or the power of the IR projector may be adaptively adjusted (e.g., reduced), according to whether at least one condition of one or more predefined conditions is met. Advantageously, compared to conventional approaches for type I and type II of IR exposure, the adaptively-controlled IP projection may result in reduction in power consumption and IR projection time as well as improved eye safety.
For illustrative purposes and without limiting the scope of the present disclosure, the face of a person may be a ROI in an image since facial recognition-related applications (e.g., three-dimensional (3D) face unlock, 3D face payment, 3D emoji, and so on) are gaining popularity in terms of usage. Accordingly, under the proposed scheme, timing and/or power of IR projection by an IR projector in active stereo vision and/or passive stereo vision may be adaptively controlled based on facial recognition or any ROI region in an image, which may be an IR image, RGB image or depth map. For instance, a face region in the image may be detected (e.g., by a control circuit or processor executing a facial detection/recognition algorithm) in an IR image, RGB image or depth map. Then, regardless of type I or type II of IR exposure, an amount of power used in projecting an IR light and/or an amount of time during which the IR light is projected may be reduced, thereby achieving reduction in power/time as well as improvement in eye safety.
Under the proposed scheme, the timing and/or power in IR projection by an IR projector (e.g., the IR projector in module 100 and/or module 200) may be adaptively controlled in one or more of a variety of ways. Description of some examples of adaptive control of IR projection is provided below.
In one example, a face (or another ROI) in an image may be used as a condition in adaptively controlling IR projection. Specifically, the IR projector may be enabled to project an IR light toward a face of a user (or another ROI in an image) during a first portion of a sensor readout time when one or more image sensors (e.g., IR camera and/or RGB camera of module 100/200) read(s) a portion of the image where the face/ROI is located. Moreover, the IR projector may be disabled to cease projecting the IR light toward the face during a second portion of the sensor readout time when the one or more image sensors read(s) another portion of the image where the ROI is not located. Accordingly, this may result in power saving. Furthermore, an eye region in an image may be used as a condition in adaptively controlling IR projection. Specifically, during the first portion of the sensor readout time, the IR projector may also be disabled to cease projecting the IR light when the one or more sensors read(s) a portion of the image where an eye/both eyes of the face is/are located so as to improve eye safety.
In one example, eye action(s) of the user may be used as a condition in adaptively controlling IR projection. Specifically, the IR projector may be enabled to project an IR light toward the face in response to eye(s) of the face being blinked or otherwise closed. Moreover, the IR projector may also be disabled to cease projecting the IR light toward the face in response to the eye(s) of the face being open. Accordingly, this may improve eye safety for the user.
In one example, pupil size may be used as a condition in adaptively controlling IR projection. Specifically, an amount of power in projecting the IR light toward the face may be reduced in response to an enlargement of the pupil of the eye(s) of the face. Again, this may improve eye safety for the user.
In one example, a distance of the face (or another ROI) may be used as a condition in adaptively controlling IR projection. Specifically, IR projection may be adaptively controlled based on a distance of the face/ROI from the one or more sensors (e.g., IR camera(s) and/or RGB camera). That is, as the distance of the face/ROI is inversely proportional to the size of the face/ROI in the image (i.e., larger size means the face/ROI is closer to the one or more sensors, and vice versa), an amount of power in projecting the IR light toward the face may be reduced in response to the distance of the face/ROI being lower than a distance threshold (e.g., reducing power in projecting the IR light when the face is relatively close). Accordingly, this may improve eye safety for the user.
In one example, environmental brightness (lux) may be used as a condition in adaptively controlling IR projection. Specifically, an amount of power in projecting the IR light may be reduced in response to the brightness level of an ambient light being lower than a brightness threshold.
In one example, environmental brightness (lux) may be used as a condition in adaptively controlling IR projection. Specifically, an amount of power in projecting the IR light may be reduced in response to the brightness level of an ambient light being lower than a brightness threshold.
In one example, a confidence level of a depth map may be used as a condition in adaptively controlling IR projection. Specifically, the IR projector may be enabled to project an IR light toward the ROI for each of one or more regions of a plurality of regions of a depth map having a respective confidence level lower than a confidence threshold.
In one example, an image texture of a ROI may be used as a condition in adaptively controlling IR projection. Specifically, the IR projector may be enabled to project an IR light toward the ROI for each of one or more regions of a plurality of regions of the ROI the respective amount of texture of which being lower than a texture threshold.
In one example, a power level of a battery that powers the IR projector may be used as a condition in adaptively controlling IR projection. Specifically, an amount of power in projecting the IR light may be reduced in response to the power level of the battery being lower than a power threshold.
Illustrative Implementations
Apparatus 300 may be a part of an electronic apparatus, a portable or mobile apparatus, a wearable apparatus, a wireless communication apparatus or a computing apparatus. For instance, apparatus 300 may be implemented in a smartphone, a smartwatch, a personal digital assistant, a digital camera, or a computing equipment such as a tablet computer, a laptop computer or a notebook computer. Moreover, apparatus 300 may also be a part of a machine type apparatus, which may be an IoT or NB-IoT apparatus such as an immobile or a stationary apparatus, a home apparatus, a wire communication apparatus or a computing apparatus. For instance, apparatus 300 may be implemented in a smart thermostat, a smart fridge, a smart door lock, a wireless speaker or a home control center. Alternatively, apparatus 300 may be implemented in the form of a system-on-chip (SoC) or one or more integrated-circuit (IC) chips such as, for example and without limitation, one or more single-core processors, one or more multi-core processors, one or more reduced-instruction-set-computing (RISC) processors or one or more complex-instruction-set-computing (CISC) processors.
Apparatus 300 may include at least some of those components shown in
In one aspect, control circuit 310 may be implemented in the form of an electronic circuit comprising various electronic components. Alternatively, control circuit 310 may be implemented as part of or in the form of one or more single-core processors, one or more multi-core processors, one or more RISC processors, or one or more CISC processors. That is, even though a singular term “a processor” is used herein to refer to control circuit 310, control circuit 310 may include multiple processors in some implementations and a single processor in other implementations in accordance with the present disclosure. In another aspect, control circuit 310 may be implemented in the form of hardware (and, optionally, firmware) with electronic components including, for example and without limitation, one or more transistors, one or more diodes, one or more capacitors, one or more resistors, one or more inductors, one or more memristors and/or one or more varactors that are configured and arranged to achieve specific purposes in accordance with the present disclosure. In other words, in at least some implementations, control circuit 310 is a special-purpose machine specifically designed, arranged and configured to perform specific tasks pertaining to adaptive IR projection control for depth estimation in computer vision in accordance with various implementations of the present disclosure. In some implementations, control circuit 310 may include an electronic circuit with hardware components implementing one or more of the various proposed schemes in accordance with the present disclosure. Alternatively, other than hardware components, control circuit 310 may also utilize software codes and/or instructions in addition to hardware components to implement adaptive IR projection control for depth estimation in computer vision in accordance with various implementations of the present disclosure.
Under various proposed schemes in accordance with the present disclosure, during operation, control circuit 310 may receive data of an image based on sensing by one or more of first sensor 330, second sensor 340 and third sensor 350. Additionally, control circuit 310 may detect a region of interest (ROI) in the image. Moreover, control circuit 310 may adaptively control light projector 320 with respect to projecting light toward the ROI.
In some implementations, in receiving the data of the image, control circuit 310 may receive data of an IR image, a RGB image or a depth map.
In some implementations, in detecting the ROI in the image, control circuit 310 may detect a face in the image. In such cases, in adaptively controlling the light projector with respect to projecting the light toward the ROI, control circuit 310 may enable IR projector 320 to project an IR light toward the face during a first portion of a sensor readout time when the one or more image sensors read a portion of the image where the ROI is located. Moreover, control circuit 310 may disable IR projector 320 to cease projecting the IR light toward the face during a second portion of the sensor readout time when the one or more image sensors read another portion of the image where the ROI is not located. In some implementations, in enabling IR projector 320 to project the IR light toward the face during the first portion of the sensor readout time when the one or more image sensors read the portion of the image where the ROI is located, control circuit 310 may disable the light projector to cease projecting the IR light when the one or more image sensors read a portion of the image where an eye of the face is located.
In some implementations, in detecting the ROI in the image, control circuit 310 may detect a face in the image and detecting an eye on the face. In such cases, in adaptively controlling the light projector with respect to projecting the light toward the ROI, control circuit 310 may enable IR projector 320 to project an IR light toward the face responsive to the eye being closed or blinked. Additionally, control circuit 310 may disable IR projector 320 to cease projecting the IR light toward the face responsive to the eye being open.
In some implementations, in detecting the ROI in the image, control circuit 310 may detect a face in the image, detecting an eye on the face, and monitoring a pupil of the eye. In such cases, in adaptively controlling the light projector with respect to projecting the light toward the ROI, control circuit 310 may control IR projector 320 to reduce an amount of power in projecting an IR light toward the face responsive to an enlargement of the pupil.
In some implementations, in detecting the ROI in the image, control circuit 310 may detect a face in the image and determining a distance of the face from the one or more image sensors based on a size of the face in the image. In such cases, in adaptively controlling the light projector with respect to projecting the light toward the ROI, control circuit 310 may control IR projector 320 to reduce an amount of power in projecting an IR light toward the face responsive to the distance of the face being lower than a distance threshold.
In some implementations, control circuit 310 may detect a brightness level of an ambient light (e.g., based on data received from any of first sensor 330, second sensor 340 and third sensor 350, or by receiving a sensing date from a light sensor (not shown)). Moreover, in adaptively controlling the light projector with respect to projecting the light toward the ROI, control circuit 310 may control IR projector 320 to reduce an amount of power in projecting an IR light responsive to the brightness level of the ambient light being lower than a brightness threshold.
In some implementations, in receiving the data of the image, control circuit 310 may receive data of a depth map of the ROI. In such cases, in adaptively controlling the light projector with respect to projecting the light toward the ROI, control circuit 310 may determine a respective confidence level for each of a plurality of regions of the depth map. Furthermore, control circuit 310 may enable IR projector to project 320 an IR light toward the ROI for each of one or more regions of the plurality of regions of the depth map the respective confidence level of which being lower than a confidence threshold.
In some implementations, in adaptively controlling the light projector with respect to projecting the light toward the ROI, control circuit 310 may perform some operations. For instance, control circuit 310 may determine a respective amount of texture for each of a plurality of regions of the ROI in the image. Moreover, control circuit 310 may enable IR projector 320 to project an IR light toward the ROI for each of one or more regions of the plurality of regions of the ROI the respective amount of texture of which being lower than a texture threshold.
In some implementations, in adaptively controlling the light projector with respect to projecting the light toward the ROI, control circuit 310 may perform some other operations. For instance, control circuit 310 may detect a power level of a battery that powers the light projector. Additionally, control circuit 310 may control IR projector 320 to reduce an amount of power in projecting an IR light toward the ROI responsive to the power level of the battery being lower than a power threshold.
Illustrative Processes
At 410, process 400 may involve control circuit 310 receiving data of an image based on sensing by one or more of first sensor 330, second sensor 340 and third sensor 350. Process 400 may proceed from 410 to 420.
At 420, process 400 may involve control circuit 310 detecting a region of interest (ROI) in the image. Process 400 may proceed from 420 to 430.
At 430, process 400 may involve control circuit 310 adaptively controlling light projector 320 with respect to projecting light toward the ROI.
In some implementations, in receiving the data of the image, process 400 may involve control circuit 310 receiving data of an IR image, a RGB image or a depth map.
In some implementations, in detecting the ROI in the image, process 400 may involve control circuit 310 detecting a face in the image. In such cases, in adaptively controlling the light projector with respect to projecting the light toward the ROI, process 400 may involve control circuit 310 enabling IR projector 320 to project an IR light toward the face during a first portion of a sensor readout time when the one or more image sensors read a portion of the image where the ROI is located. Moreover, process 400 may involve control circuit 310 disabling IR projector 320 to cease projecting the IR light toward the face during a second portion of the sensor readout time when the one or more image sensors read another portion of the image where the ROI is not located. In some implementations, in enabling IR projector 320 to project the IR light toward the face during the first portion of the sensor readout time when the one or more image sensors read the portion of the image where the ROI is located, process 400 may involve control circuit 310 disabling the light projector to cease projecting the IR light when the one or more image sensors read a portion of the image where an eye of the face is located.
In some implementations, in detecting the ROI in the image, process 400 may involve control circuit 310 detecting a face in the image and detecting an eye on the face. In such cases, in adaptively controlling the light projector with respect to projecting the light toward the ROI, process 400 may involve control circuit 310 enabling IR projector 320 to project an IR light toward the face responsive to the eye being closed or blinked. Additionally, process 400 may involve control circuit 310 disabling IR projector 320 to cease projecting the IR light toward the face responsive to the eye being open.
In some implementations, in detecting the ROI in the image, process 400 may involve control circuit 310 detecting a face in the image, detecting an eye on the face, and monitoring a pupil of the eye. In such cases, in adaptively controlling the light projector with respect to projecting the light toward the ROI, process 400 may involve control circuit 310 controlling IR projector 320 to reduce an amount of power in projecting an IR light toward the face responsive to an enlargement of the pupil.
In some implementations, in detecting the ROI in the image, process 400 may involve control circuit 310 detecting a face in the image and determining a distance of the face from the one or more image sensors based on a size of the face in the image. In such cases, in adaptively controlling the light projector with respect to projecting the light toward the ROI, process 400 may involve control circuit 310 controlling IR projector 320 to reduce an amount of power in projecting an IR light toward the face responsive to the distance of the face being lower than a distance threshold.
In some implementations, process 400 may involve control circuit 310 detecting a brightness level of an ambient light. Moreover, in adaptively controlling the light projector with respect to projecting the light toward the ROI, process 400 may involve control circuit 310 controlling IR projector 320 to reduce an amount of power in projecting an IR light responsive to the brightness level of the ambient light being lower than a brightness threshold.
In some implementations, in receiving the data of the image, process 400 may involve control circuit 310 receiving data of a depth map of the ROI. In such cases, in adaptively controlling the light projector with respect to projecting the light toward the ROI, process 400 may involve control circuit 310 determining a respective confidence level for each of a plurality of regions of the depth map. Furthermore, process 400 may involve control circuit 310 enabling IR projector to project 320 an IR light toward the ROI for each of one or more regions of the plurality of regions of the depth map the respective confidence level of which being lower than a confidence threshold.
In some implementations, in adaptively controlling the light projector with respect to projecting the light toward the ROI, process 400 may involve control circuit 310 performing some operations. For instance, process 400 may involve control circuit 310 determining a respective amount of texture for each of a plurality of regions of the ROI in the image. Moreover, process 400 may involve control circuit 310 enabling IR projector 320 to project an IR light toward the ROI for each of one or more regions of the plurality of regions of the ROI the respective amount of texture of which being lower than a texture threshold.
In some implementations, in adaptively controlling the light projector with respect to projecting the light toward the ROI, process 400 may involve control circuit 310 performing some other operations. For instance, process 400 may involve control circuit 310 detecting a power level of a battery that powers the light projector. Additionally, process 400 may involve control circuit 310 controlling IR projector 320 to reduce an amount of power in projecting an IR light toward the ROI responsive to the power level of the battery being lower than a power threshold.
Additional Notes
The herein-described subject matter sometimes illustrates different components contained within, or connected with, different other components. It is to be understood that such depicted architectures are merely examples, and that in fact many other architectures can be implemented which achieve the same functionality. In a conceptual sense, any arrangement of components to achieve the same functionality is effectively “associated” such that the desired functionality is achieved. Hence, any two components herein combined to achieve a particular functionality can be seen as “associated with” each other such that the desired functionality is achieved, irrespective of architectures or intermedial components. Likewise, any two components so associated can also be viewed as being “operably connected”, or “operably coupled”, to each other to achieve the desired functionality, and any two components capable of being so associated can also be viewed as being “operably couplable”, to each other to achieve the desired functionality. Specific examples of operably couplable include but are not limited to physically mateable and/or physically interacting components and/or wirelessly interactable and/or wirelessly interacting components and/or logically interacting and/or logically interactable components.
Further, with respect to the use of substantially any plural and/or singular terms herein, those having skill in the art can translate from the plural to the singular and/or from the singular to the plural as is appropriate to the context and/or application. The various singular/plural permutations may be expressly set forth herein for sake of clarity.
Moreover, it will be understood by those skilled in the art that, in general, terms used herein, and especially in the appended claims, e.g., bodies of the appended claims, are generally intended as “open” terms, e.g., the term “including” should be interpreted as “including but not limited to,” the term “having” should be interpreted as “having at least,” the term “includes” should be interpreted as “includes but is not limited to,” etc. It will be further understood by those within the art that if a specific number of an introduced claim recitation is intended, such an intent will be explicitly recited in the claim, and in the absence of such recitation no such intent is present. For example, as an aid to understanding, the following appended claims may contain usage of the introductory phrases “at least one” and “one or more” to introduce claim recitations. However, the use of such phrases should not be construed to imply that the introduction of a claim recitation by the indefinite articles “a” or “an” limits any particular claim containing such introduced claim recitation to implementations containing only one such recitation, even when the same claim includes the introductory phrases “one or more” or “at least one” and indefinite articles such as “a” or “an,” e.g., “a” and/or “an” should be interpreted to mean “at least one” or “one or more;” the same holds true for the use of definite articles used to introduce claim recitations. In addition, even if a specific number of an introduced claim recitation is explicitly recited, those skilled in the art will recognize that such recitation should be interpreted to mean at least the recited number, e.g., the bare recitation of “two recitations,” without other modifiers, means at least two recitations, or two or more recitations. Furthermore, in those instances where a convention analogous to “at least one of A, B, and C, etc.” is used, in general such a construction is intended in the sense one having skill in the art would understand the convention, e.g., “a system having at least one of A, B, and C” would include but not be limited to systems that have A alone, B alone, C alone, A and B together, A and C together, B and C together, and/or A, B, and C together, etc. In those instances where a convention analogous to “at least one of A, B, or C, etc.” is used, in general such a construction is intended in the sense one having skill in the art would understand the convention, e.g., “a system having at least one of A, B, or C” would include but not be limited to systems that have A alone, B alone, C alone, A and B together, A and C together, B and C together, and/or A, B, and C together, etc. It will be further understood by those within the art that virtually any disjunctive word and/or phrase presenting two or more alternative terms, whether in the description, claims, or drawings, should be understood to contemplate the possibilities of including one of the terms, either of the terms, or both terms. For example, the phrase “A or B” will be understood to include the possibilities of “A” or “B” or “A and B.”
From the foregoing, it will be appreciated that various implementations of the present disclosure have been described herein for purposes of illustration, and that various modifications may be made without departing from the scope and spirit of the present disclosure. Accordingly, the various implementations disclosed herein are not intended to be limiting, with the true scope and spirit being indicated by the following claims.
Ju, Chi-Cheng, Chang, Te-Hao, Chen, Ying-Jui, Li, Chun-Nan, Wang, Chi-Hui
Patent | Priority | Assignee | Title |
Patent | Priority | Assignee | Title |
10321055, | Sep 03 2013 | SEEING MACHINES LIMITED | Low power eye tracking system and method |
10740605, | Jul 05 2017 | GUANGDONG OPPO MOBILE TELECOMMUNICATIONS CORP., LTD. | Iris collection method, electronic device, and computer readable storage medium |
10776623, | Jun 26 2017 | GUANGDONG OPPO MOBILE TELECOMMUNICATIONS CORP., LTD. | Method for electronic device acquiring iris and electronic device |
11068712, | Sep 30 2014 | Qualcomm Incorporated | Low-power iris scan initialization |
7695138, | Mar 31 2005 | Wistron Corporation | Safe eye detection |
20080277601, | |||
20160275348, | |||
20160283789, | |||
20170061210, | |||
20180189547, | |||
20190279398, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Jul 26 2021 | MEDIATEK INC. | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Jul 26 2021 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Date | Maintenance Schedule |
Jan 31 2026 | 4 years fee payment window open |
Jul 31 2026 | 6 months grace period start (w surcharge) |
Jan 31 2027 | patent expiry (for year 4) |
Jan 31 2029 | 2 years to revive unintentionally abandoned end. (for year 4) |
Jan 31 2030 | 8 years fee payment window open |
Jul 31 2030 | 6 months grace period start (w surcharge) |
Jan 31 2031 | patent expiry (for year 8) |
Jan 31 2033 | 2 years to revive unintentionally abandoned end. (for year 8) |
Jan 31 2034 | 12 years fee payment window open |
Jul 31 2034 | 6 months grace period start (w surcharge) |
Jan 31 2035 | patent expiry (for year 12) |
Jan 31 2037 | 2 years to revive unintentionally abandoned end. (for year 12) |