A method and an image pick-up apparatus for calculating coordinates of an object being captured by using fisheye images are provided. In the method, a first and a second fisheye image containing the object are respectively captured by using a first and a second fisheye lens. Coordinates of the object in the first and the second fisheye image are detected. A first and a second azimuth angle of the object relative to fisheye centers of the first and the second fisheye image on an image sensor plane are accordingly calculated. Distances between the coordinates of the object and the fisheye centers are respectively transformed into a first and a second incident angle through lens curves of the fisheye lenses. Three-dimensional coordinates of the object are calculated according to the first and the second azimuth angle, the first and the second incident angle, and a baseline distance between the fisheye lenses.
|
9. A method for calculating coordinates of an object captured by using fisheye images, adapted for an image pick-up apparatus having a fisheye lens, the method comprising:
capturing a fisheye image containing an object by using the fisheye lens;
detecting coordinates of the object in the fisheye image;
calculating an azimuth angle of the object relative to a fisheye center of the fisheye image on an image sensor plane of the fisheye lens according to the coordinates;
transforming a distance between the coordinates and the fisheye center of the fisheye image into an incident angle through a lens curve of the fisheye lens;
calculating a distance between the object and the fisheye lens according to a width of the object in the fisheye image in a direction facing the fisheye center; and
calculating three-dimensional coordinates of the object by using a trigonometric function according to the azimuth angle, the incident angle, and the distance,
wherein the width of the object in the fisheye image is wT, an actual width of the object is WT, an image width of the fisheye image is WI, and a field of view (FOV) of the fisheye lens is F, and a distance r between the object and the fisheye lens is:
13. An image pick-up apparatus, comprising:
a fisheye lens;
a storage device, storing a plurality of modules;
a processor, coupled to the fisheye lens and the storage device and accessing and executing the modules stored in the storage device, the modules comprising:
an image pick-up module, capturing an image containing an object by using the fisheye lens;
an object detection module, detecting coordinates of the object in the fisheye image;
an azimuth angle calculation module, calculating an azimuth angle of the object relative to a fisheye center of the fisheye image on an image sensor plane of the fisheye lens according to the coordinates;
an incident angle calculation module, transforming a distance between the coordinates and the fisheye center of the fisheye image into an incident angle through a lens curve of the fisheye lens;
a distance calculation module, calculating a distance between the object and the fisheye lens according to a width of the object in the fisheye image in a direction facing the fisheye center; and
a coordinate calculation module, calculating three-dimensional coordinates of the object by using a trigonometric function according to the azimuth angle, the incident angle, and the distance,
wherein the width of the object in the fisheye image is wT, an actual width of the object is WT, an image width of the fisheye image is WI, a field of view (FOV) of the fisheye lens is F, and a distance r between the object and the fisheye lens is:
1. A method for calculating coordinates of an object being captured by using fisheye images, adapted for an image pick-up apparatus having a first fisheye lens and a second fisheye lens, wherein a baseline distance is provided between the first fisheye lens and the second fisheye lens, and the method comprises:
capturing a first fisheye image and a second fisheye image containing an object respectively by using the first fisheye lens and the second fisheye lens;
detecting first coordinates of the object in the first fisheye image and second coordinates of the object in the second fisheye image;
calculating a first azimuth angle and a second azimuth angle of the object relative to fisheye centers of the first fisheye image and the second fisheye image on an image sensor plane of the first fisheye lens and the second fisheye lens according to the first coordinates and the second coordinates;
transforming a first distance between the first coordinates and the fisheye center of the first fisheye image and a second distance between the second coordinates and the fisheye center of the second fisheye image into a first incident angle and a second incident angle respectively through lens curves of the first fisheye lens and the second fisheye lens; and
calculating three-dimensional coordinates of the object by using a trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance,
wherein the first azimuth angle is φl, the second azimuth angle is φr, the first incident angle is θl, the second incident angle is θr, the baseline distance is B, and the three-dimensional coordinates of the object are (x, y, z), wherein
5. An image pick-up apparatus, comprising:
a first fisheye lens;
a second fisheye lens, a baseline distance being provided between the first fisheye lens and the second fisheye lens;
a storage device, storing a plurality of modules;
a processor, coupled to the first fisheye lens, the second fisheye lens, and the storage device and accessing and executing the modules stored in the storage device, the modules comprising:
an image pick-up module, capturing a first fisheye image and a second fisheye image containing an object respectively by using the first fisheye lens and the second fisheye lens;
an object detection module, detecting first coordinates and second coordinates of the object in the first fisheye image and the second fisheye image;
an azimuth angle calculation module, calculating a first azimuth angle and a second azimuth angle of the object relative to fisheye centers of the first fisheye image and the second fisheye image on an image sensor plane of the first fisheye lens and the second fisheye lens according to the first coordinates and the second coordinates;
an incident angle calculation module, transforming a first distance between the first coordinates and the fisheye center of the first fisheye image and a second distance between the second coordinates and the fisheye center of the second fisheye image into a first incident angle and a second incident angle respectively through lens curves of the first fisheye lens and the second fisheye lens; and
a coordinate calculation module, calculating three-dimensional coordinates of the object by using a trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance,
wherein the first azimuth angle is φl, the second azimuth angle is φr, the first incident angle is θl, the second incident angle is θr, the baseline distance is B, and the three-dimensional coordinates of the object are (x, y, z), wherein
2. The method as claimed in
detecting a plurality of pixels having a brightness or a color component greater than a predetermined value in each of the first fisheye image and the second fisheye image, coordinates of centers or centroids of regions formed by the pixels in the first fisheye image and the second fisheye image acting as the first coordinates and the second coordinates.
3. The method as claimed in
4. The method as claimed in
the three-dimensional coordinates of the object are calculated by using the trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance when the object appears in the overlapping region; and
two-dimensional coordinates of the object are calculated by using the trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance when the object appears in the non-overlapping region.
6. The image pick-up apparatus as claimed in
7. The image pick-up apparatus as claimed in
8. The image pick-up apparatus as claimed in
the coordinate calculation module calculates the three-dimensional coordinates of the object by using the trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance when the object appears in the overlapping region; and
the coordinate calculation module calculates two-dimensional coordinates of the object by using the trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance when the object appears in the non-overlapping region.
10. The method as claimed in
detecting a plurality of pixels having a brightness or a color component greater than a predetermined value in the fisheye image, coordinates of a center or a centroid of a region formed by the pixels in the fisheye image acting as the coordinates.
11. The method as claimed in
12. The method as claimed in
x=r cos θ cos φ; y=r cos θ sin φ; and z=r sin θ. 14. The image pick-up apparatus as claimed in
15. The image pick-up apparatus as claimed in
16. The image pick-up apparatus as claimed in
x=r cos θ cos φ; y=r cos θ sin φ; and z=r sin θ. |
This application claims the priority benefits of Taiwan application serial no. 106136425, filed on Oct. 24, 2017, and Taiwan application serial no. 107100884, filed on Jan. 10, 2018. The entirety of each of the above-mentioned patent applications is hereby incorporated by reference herein and made a part of specification.
The invention relates to an apparatus and a method for image pick-up. More particularly, the invention relates to a method and an image pick-up apparatus for calculating coordinates of an object being captured by using fisheye images.
Game consoles are one of the indispensable electronic products for home entertainment nowadays. In the control modes of the game consoles, elements used for motion sensing are now added to most of the game consoles to replace the conventional handles, so as to bring greater interaction between the players and the game content. A user's movement or action in space is detected by a sensor such as an infrared sensor to control the game content, and entertainment of the game is thus considerably enhanced.
In addition to the conventional infrared sensor positioning technology, the light bulb detection technology is introduced to the game consoles recently. When the user holds and waves the light bulb joystick in the space, images of the light bulb are captured by the dual lenses disposed on the game console, and the position of the light bulb in the space is thus calculated according to the positions of the light bulb in the images.
Nevertheless, the conventional lenses have limited field of view (FOV), the capturing range of the dual lenses is thus limited, and the moving space of the user is limited as a result. The fisheye lenses with greater field of view may be used to replace the conventional lenses. Nevertheless, images picked up by the fisheye lenses are deformed, and fisheye calibration is thereby required before positioning. Geometry transformation is required during calibration, which may lead to decrease in pipeline duration of the video images, and the frame rates may thus need to be increased to solve such problem.
The invention provides a method and an image pick-up apparatus for calculating coordinates of an object being captured by using fisheye images, in which three-dimensional coordinates of the object in a space are calculated by using the fisheye images without performing geometry transformation.
In an embodiment of the invention, a method for calculating coordinates of an object being captured by using fisheye images is adapted for an image pick-up apparatus having a first fisheye lens and a second fisheye lens. A baseline distance is provided between the first fisheye lens and the second fisheye lens. In the method, a first fisheye image and a second fisheye image containing an object respectively are captured by using the first fisheye lens and the second fisheye lens. Next, first coordinates of the object in the first fisheye image and second coordinates of the object in the second fisheye image are detected. A first azimuth angle and a second azimuth angle of the object relative to fisheye centers of the first fisheye image and the second fisheye image on an image sensor plane of the first fisheye lens and the second fisheye lens are then calculated according to the first coordinates and the second coordinates. A first distance between the first coordinates and the fisheye center of the first fisheye image and a second distance between the second coordinates and the fisheye center of the second fisheye image are transformed into a first incident angle and a second incident angle respectively through lens curves of the first fisheye lens and the second fisheye lens. Finally, three-dimensional coordinates of the object are calculated by using a trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance.
In an embodiment of the invention, the object includes a light-emitting device. The step of detecting the first coordinates of the object in the first fisheye image and the second coordinates of the object in the second fisheye image includes detecting a plurality of pixels having a brightness or a color component greater than a predetermined value in each of the first fisheye image and the second fisheye image. Coordinates of centers or centroids of regions formed by the pixels in the first fisheye image and in the second fisheye image act as the first coordinates and the second coordinates.
In an embodiment of the invention, an incident angle of light emitted by the object on the first fisheye lens is proportional to a projection radius of the light projected on an image sensor of the first fisheye lens, and an incident angle of light emitted by the object on the second fisheye lens is proportional to a projection radius of the light projected on an image sensor of the second fisheye lens.
In an embodiment of the invention, the first azimuth angle is φl, the second azimuth angle is φr, the first incident angle is θl, the second incident angle is θr, the baseline distance is B, and the three-dimensional coordinates of the object are (x, y, z), wherein
In an embodiment of the invention, an included angle is provided between an optical axis of the first fisheye lens and an optical axis of the second fisheye lens and that a field of view of the first fisheye lens and a field of view of the second fisheye lens include an overlapping region and a non-overlapping region. The three-dimensional coordinates of the object are calculated by using the trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance when the object appears in the overlapping region; and two-dimensional coordinates of the object are calculated by using the trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance when the object appears in the non-overlapping region.
An image pick-up apparatus of an embodiment of the invention includes a first fisheye lens, a second fisheye lens, a storage device, and a processor. A baseline distance is provided between the first fisheye lens and the second fisheye lens. The storage device is configured for storing a plurality of modules. The processor is coupled to the first fisheye lens, the second fisheye lens, and the storage device and is configured for accessing and executing the modules stored in the storage device. The modules include an image capturing module, an object detection module, an azimuth angle calculation module, an incident angle calculation module, and a coordinate calculation module. The image pick-up module captures a first fisheye image and a second fisheye image containing an object respectively by using the first fisheye lens and the second fisheye lens. The object detection module detects first coordinates and second coordinates of the object in the first fisheye image and the second fisheye image. The azimuth angle calculation module calculates a first azimuth angle and a second azimuth angle of the object relative to fisheye centers of the first fisheye image and the second fisheye image on an image sensor plane of the first fisheye lens and the second fisheye lens according to the first coordinates and the second coordinates. The azimuth calculation module transforms a first distance between the first coordinates and the fisheye center of the first fisheye image and a second distance between the second coordinates and the fisheye center of the second fisheye image into a first incident angle and a second incident angle respectively through lens curves of the first fisheye lens and the second fisheye lens. The coordinate calculation module calculates three-dimensional coordinates of the object by using a trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance.
In an embodiment of the invention, the object includes a light-emitting device, and the object detection module includes a plurality of pixels having a brightness or a color component greater than a predetermined value in each of the first fisheye image and the second fisheye image, and coordinates of centers or centroids of regions formed by the pixels in the first fisheye image and in the second fisheye image act as the first coordinates and the second coordinates.
In an embodiment of the invention, an incident angle of light emitted by the object on the first fisheye lens is proportional to a projection radius of the light projected on an image sensor of the first fisheye lens, and an incident angle of light emitted by the object on the second fisheye lens is proportional to a projection radius of the light projected on an image sensor of the second fisheye lens.
In an embodiment of the invention, the first azimuth angle is φl, the second azimuth angle is φr, the first incident angle is θl, the second incident angle is θr, the baseline distance is B, and the three-dimensional coordinates of the object are (x, y, z), wherein
In an embodiment of the invention, an included angle is provided between an optical axis of the first fisheye lens and an optical axis of the second fisheye lens and that a field of view of the first fisheye lens and a field of view of the second fisheye lens include an overlapping region and a non-overlapping region. The coordinate calculation module calculates the three-dimensional coordinates of the object by using the trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance when the object appears in the overlapping region; and the coordinate calculation module calculates two-dimensional coordinates of the object by using the trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance when the object appears in the non-overlapping region.
In an embodiment of the invention, a method for calculating coordinates of an object being captured by using fisheye images is adapted for an image pick-up apparatus having a fisheye lens. In the method, a fisheye image containing an object is captured by using the fisheye lens. Next, coordinates of the object in the fisheye image are calculated. An azimuth angle of the object relative to a fisheye center of the fisheye image on an image sensor plane of the fisheye lens in calculated according to the coordinates. A distance between the coordinates and the fisheye center of the fisheye image is transformed into an incident angle through a lens curve of the fisheye lens. A distance between the object and the fisheye lens is calculated according to a width of the object in the fisheye image in a direction facing the fisheye center. Finally, the three-dimensional coordinates of the object are calculated by using a trigonometric function according to the azimuth angle, the incident angle, and the distance.
In an embodiment of the invention, the object includes a light-emitting device. The step of detecting coordinates of the object in a fisheye image includes detecting a plurality of pixels having a brightness or a color component greater than a predetermined value in the fisheye image. Coordinates of a center or a centroid of a region formed by the pixels in the fisheye image act as the coordinates.
An image pick-up apparatus of an embodiment of the invention includes a fisheye lens, a storage device, and a processor. The storage device is configured for storing a plurality of modules. The processor is coupled to the fisheye lens and the storage device and is configured for accessing and executing the modules stored in the storage device. The modules include an image capturing module, an object detection module, an azimuth angle calculation module, an incident angle calculation module, a distance calculation module, and a coordinate calculation module. The image pick-up module captures a fisheye image containing an object by using the fisheye lens. The object detection module detects coordinates of the object in the fisheye image. The azimuth angle calculation module calculates an azimuth angle of the object relative to a fisheye center of the fisheye image on an image sensor plane of the fisheye lens according to the coordinates. The incident angle calculation module transforms a distance between the coordinates and the fisheye center of the fisheye image into an incident angle through a lens curve of the fisheye lens. The distance calculation module calculates a distance between the object and the fisheye lens according to a width of the object in the fisheye image in a direction facing the fisheye center. The coordinate calculation module calculates three-dimensional coordinates of the object by using a trigonometric function according to the azimuth angle, the incident angle, and the distance.
In an embodiment of the invention, the object includes a light-emitting device. The object detection module detects a plurality of pixels having a brightness or a color component greater than a predetermined value in the fisheye image. Coordinates of a center or a centroid of a region formed by the pixels in the fisheye image act as the coordinates.
In an embodiment of the invention, the incident angle of light emitted by the object on the fisheye lens is proportional to a projection radius of the light projected on an image sensor of the fisheye lens.
In an embodiment of the invention, it is assumed that a width of the object in the fisheye image is wT, an actual width of the object is WT, an image width of the fisheye image is WI, and a field of view (FOV) of the fisheye lens is F, and a distance r between the object and the fisheye lens is:
In an embodiment of the invention, the azimuth angle is φ, the incident angle is θ, the distance is r, and the three-dimensional coordinates of the object are (x, y, z), wherein
x=r cos θ cos φ;
y=r cos θ sin φ; and
z=r sin θ.
To sum up, in the method and the image pick-up apparatus for calculating the coordinates of the object being captured by using the fisheye images provided by the embodiments of the invention, the image of the object is captured by using the fisheye lenses, and the coordinates of the object in the fisheye images are directly used to calculate the incident angles of the light emitted by the object on the fisheye lenses and the azimuth angles of the light emitted by the object on the lens plane without performing geometry transformation, and finally the three-dimensional coordinates of the object in the space are calculated by using a triangle positioning method.
To make the aforementioned and other features and advantages of the invention more comprehensible, several embodiments accompanied with drawings are described in detail as follows.
The accompanying drawings are included to provide a further understanding of the invention, and are incorporated in and constitute a part of this specification. The drawings illustrate embodiments of the invention and, together with the description, serve to explain the principles of the invention.
Reference will now be made in detail to the present preferred embodiments of the invention, examples of which are illustrated in the accompanying drawings. Wherever possible, the same reference numbers are used in the drawings and the description to refer to the same or like parts.
Since an equi-distance lens is used for a fisheye lens, an incident angle of light received from an object (e.g., a light bulb) by the fisheye lens is approximately linear with a projection radius of the light projected on an image sensor. Accordingly, in the embodiments of the invention, a position of the object in a fisheye image is detected, an incident angle of the light emitted by the object is inversely derived through the foregoing relationship, and three-dimensional coordinates of the object in a space may thus be calculated by using the incident angle combining with an azimuth angle of the object on a lens plane through a triangle positioning method. In another aspect, in the embodiments of the invention, a distance between the object and the fisheye lens is further calculated according to a width of the object in the fisheye image. Finally, the three-dimensional coordinates of the object in the space are calculated by using a trigonometric function according to the distance, the incident angle, and the azimuth angle calculated. In this way, in the embodiments of the invention, the three-dimensional coordinates of the object being captured may be calculated while increasing a field of view of image capturing without performing geometry transformation.
Each of the first fisheye lens 12 and the second fisheye lens 14 includes a lens and an image sensor. Fixed-focus lenses or variable-focus lenses with a view angle close to, equal to, or greater than 180 degrees are adopted for the lenses, such that, an object being captured within a field of view (FOV) of the lenses may be imaged on the image sensors. A charge coupled device (CCD), a complementary metal-oxide semiconductor (CMOS) device, or a photosensitive device of other types is disposed in the image sensor. Intensity of light entering into the lenses may thus be sensed, and an image signal is picked up to generate fisheye images. A baseline distance, for example, is provided between the first fisheye lens 12 and the second fisheye lens 14.
The storage device 16, for example, is a fixed or a movable random access memory (RAM) in any form, a read-only memory (ROM), a flash memory, any other similar device, or a combination of the foregoing devices. In this embodiment, the storage device 16 is configured to record an image capturing module 161, an object detection module 162, an azimuth angle calculation module 163, an incident angle calculation module 164, and a coordinate calculation module 165. The modules, for example, are stored in a program in the storage device 16.
The processor 18 may be, for example, a central processing unit (CPU), a programmable microprocessor for general or special use, a digital signal processor (DSP), a programmable controller, an application specific integrated circuit (ASIC), a programmable logic device (PLD), any other similar device, or a combination of the foregoing devices. The processor 18 is connected to the first fisheye lens 12, the second fisheye lens 14, and the storage device 16 and is configured for loading programs of the image capturing module 161, the object detection module 162, the azimuth angle calculation module 163, the incident angle calculation module 164, and the coordinate calculation module 165 from the storage device 16, so as to accordingly perform the method for calculating the coordinates of the object being captured by using the fisheye images in the embodiments of the invention.
To be specific,
First, the processor 18 executes the image pick-up module 161 for capturing a first fisheye image and a second fisheye image containing an object respectively by using the first fisheye lens 12 and the second fisheye lens 12 (step S202). The object is, for example, a white, blue, red, or green light bulb or a light bulb emitting any other color which may be recognized easily. The light bulb is disposed on, for example, a joystick, a controller, a head-mounted display or a helmet for virtual reality use, or a wearable device such as a wristband or a watch. The image pick-up apparatus 10 may thus identify actions of a user hand-holding or wearing the object through the white light or any other color of light emitted by the light bulb.
Next, the processor 18 executes the object detection module 162 for detecting first coordinates of the object in the first fisheye image and second coordinates of the object in the second fisheye image (step S204). The object detection module 162, for example, detects a plurality of pixels having a brightness or a specific color component greater than a predetermined value in each of the first fisheye image and the second fisheye image. Coordinates of centers or centroids of regions formed by the pixels in the first fisheye image and in the second fisheye image act as the first coordinates and the second coordinates.
To be specific, if the object is the light bulb, the object in the fisheye images is presented as circular (or oval) regions having a greater brightness or a greater specific color component. In this way, the object detection module 162 may detect an object (the white light bulb) by comparing a brightness value of each of the pixels in the fisheye images with the predetermined value, or detect an object with a specific color (e.g., the blue light bulb) by comparing a pixel value of the specific color component (e.g., R, G, B) (e.g., a pixel value of a blue component) of each of the pixels in the fisheye images with the predetermined value. In another aspect, the object captured by the fisheye lens has different degrees of deformation (e.g., a circular light bulb is changed to an oval light bulb) according to a distance of the object deviating from the fisheye center. When the coordinates of the object in the fisheye images are determined, a center (e.g., a center of a smallest rectangle with an area covering the object) or a centroid of the object is calculated by the object detection module 162, for example, and coordinates of the centers or the centroids in the fisheye images are used as the coordinates of the object.
The processor 18 then executes the azimuth angle calculation module 163 to calculate a first azimuth angle and a second azimuth angle of the object relative to the fisheye centers of the first fisheye image and the second fisheye image on an image sensor plane of the first fisheye lens and the second fisheye lens according to the first coordinates and the second coordinates (step S206). In addition, the processor 18 further executes the incident angle calculation module 164 to transform a first distance between the first coordinates and the fisheye center of the first fisheye image and a second distance between the second coordinates and the fisheye center of the second fisheye image into a first incident angle and a second incident angle through lens curves of the first fisheye lens 12 and the second fisheye lens 14 (step S208). Herein, the processor 18 executes step S206 and step S208 in an order not limited to the above. The processor 18 may execute step S208 before executing step S206 or may execute the two steps simultaneously.
Projection radiuses of the object captured by the first fisheye lens 12 and the second fisheye lens 14 on the image sensor plane may vary according to the lens curves. In an embodiment, if equi-distance lenses are adopted for the first fisheye lens 12 and the second fisheye lens 14, the incident angles of light received by the fisheye lenses are approximately linear with the projection radiuses of the light on the image sensors. In this way, an incident angle of the light emitted by the object on the first fisheye lens 12 is proportional to the projection radius of the light projected on the image sensor of the first fisheye lens 12. On the other hand, an incident angle of the light emitted by the object on the second fisheye lens 14 is proportional to the projection radius of the light projected on the image sensor of the second fisheye lens 14. In another embodiment, a polynomial function may also be used to present the relationship of the incident angles of the light received by the first fisheye lens 12 and the second fisheye lens 14 and the projection radiuses of the light on the image sensors. Such relationship may be obtained in advance through obtaining the lens curves in advance or measuring the projection radiuses of incident lights of different angles on the image sensors in advance.
For instance,
After the azimuth angle of the object relative to the fisheye center and an incident angle calculated through a distance between the object and the fisheye center, the processor 18 executes the coordinate calculation module 165, so as to calculate three-dimensional coordinates of the object by using a trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance (step S210).
To be specific,
According to the azimuth angle φl, the azimuth angle φr, the incident angle θl, the incident angle θr, and the baseline distance B, the three-dimensional coordinates of the object T is (x, y, z) may be derived from the formulas as follows:
Through the foregoing method, the image pick-up apparatus 10 of this embodiment may carry out calculation of the three-dimensional coordinates of the object being captured without performing fisheye calibration, and have a capturing range greater than that of conventional lenses.
Note that in the embodiment, optical axes of the two fisheye lenses of the image pick-up apparatus are parallel to each other by default. Nevertheless, in other embodiments, the optical axes of the two fisheye lenses are not required to be parallel to each other. That is, an included angle is provided between the optical axes of the two fisheye lenses. When the included angle increases, ranges covered by fields of view of the two fisheye lenses increase and an angle used to detect the location of the object increases as well.
The human eyes have a limited perception of depth, but have a better perception of objects located at the two sides. In an image pick-up apparatus of an embodiment of the invention, two fisheye lenses having non-parallel optical axes are adopted. As an included angle is provided between the two optical axes, an overlapping region and a non-overlapping region are included in the fields of view of the two fisheye lenses. When an object appears in the overlapping region, the image pick-up apparatus may calculate the three-dimensional coordinates of the object by using the trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance calculated in the foregoing embodiment. When an object appears in the non-overlapping region, the image pick-up apparatus may calculate two-dimensional coordinates of the object by using the trigonometric function according to the first azimuth angle, the second azimuth angle, the first incident angle, the second incident angle, and the baseline distance. Since the object located in the non-overlapping region is captured by only one fisheye lens, a depth of the object is unable to be calculated. Nevertheless, the two-dimensional coordinates of the object may still be calculated, the object located at an edge of the field of view is thus positioned.
First, the processor 56 executes an image pick-up module 541 for capturing a fisheye image containing an object by using the fisheye lens 52 (step S602). The object is, for example, a white, blue, red, or green light bulb or a light bulb emitting any other color which may be recognized easily. The light bulb is disposed on, for example, a joystick, a controller, a head-mounted display or a helmet for virtual reality use, or a wearable device such as a wristband or a watch. The image pick-up apparatus 50 may thus identify actions of a user hand-holding or wearing the object through the white light or any other color of light emitted by the light bulb.
Next, the processor 56 executes an object detection module 542 for detecting coordinates of the object in the fisheye image (step S604). The object detection module 542, for example, detects a plurality of pixels having a brightness or a specific color component greater than a predetermined value in the fisheye image. Coordinates of a center or a centroid of a region formed by the pixels in the fisheye image act as the coordinates.
To be specific, if the object is the light bulb, the object in the fisheye image is presented as a circular (or oval) region having a greater brightness or a greater specific color component. In this way, the object detection module 542 may detect an object (the white light bulb) by comparing a brightness value of each of the pixels in the fisheye image with the predetermined value, or detect an object with a specific color (e.g., the blue light bulb) by comparing a pixel value of the specific color component (e.g., R, G, B) (e.g., a pixel value of a blue component) of each of the pixels in the fisheye image with the predetermined value. In another aspect, the object captured by the fisheye lens has different degrees of deformation (e.g., a circular light bulb is changed to an oval light bulb) according to distances of the object deviating from the fisheye center. When the coordinates of the object in the fisheye image is determined, a center (e.g., a center of a smallest rectangle with an area covering the object) or a centroid of the object is calculated by the object detection module 542, for example, and coordinates of the center or the centroid in the fisheye image is used as the coordinates of the object.
The processor 56 than executes an azimuth angle calculation module 543 to calculate an azimuth angle of the object relative to a fisheye center of the fisheye image on an image sensor plane of the fisheye lens according to the coordinates (step S606). In addition, the processor 56 also executes an incident angle calculation module 544 to transform a distance between the coordinates and the fisheye center of the fisheye image into an incident angle through a lens curve of the fisheye lens 52 (step S608). Herein, the processor 56 executes step S606 and step S608 in an order not limited to the above. The processor 56 may execute step S608 before executing step S606 or may execute the two steps simultaneously.
A projection radius of the object captured by the fisheye lens 52 on the image sensor plane may vary according to the lens curve. In an embodiment, if an equi-distance lens is adopted for the fisheye lens 52, an incident angle of light received by the fisheye lens is approximately linear with a projection radius of the light on an image sensor. In this way, the incident angle of the light emitted by the object on the fisheye lens 52 is proportional to the projection radius of the light projected on the image sensor of the fisheye lens 52. In another embodiment, a polynomial function may also be used to present the relationship of the incident angle of the light received by the fisheye lens 52 and the projection radius of the light on the image sensor. Such relationship may be obtained in advance through obtaining the lens curve in advance or measuring the projection radius of incident lights of different angles on the image sensor in advance.
In addition, the processor 56 executes a distance calculation module 545 to calculate a distance between the object and the fisheye lens according to a width of the object in the fisheye image in a direction facing the fisheye center (step S610). To be specific, as described above, the object captured by the fisheye lens 52 has different degrees of deformation (e.g., a circular light bulb is changed to an oval light bulb) according to different distances of the captured object deviating from the fisheye center. The captured object may be subjected to greater degrees of deformation when the captured object is deviated farther away from the fisheye center. According to such characteristic, in the embodiments of the invention, a distance between the object and the fisheye lens 522 is calculated according to a width of the captured object after being deformed in the fisheye image.
For instance,
To be specific, in an embodiment, it is assumed that a width of the object in the fisheye image is wT, an actual width of the object is WT, an image width of the fisheye image is WI, and a field of view (FOV) of the fisheye lens is F, and a distance r between the object and the fisheye lens is:
For instance, when the fisheye lens is used to capture the fisheye images of the object being placed in different distances, Table 1 lists relationships of the widths (pixels) of the object in the fisheye images and the distances. Herein, an object 1 is a light bulb with a diameter of 1 centimeter, an object 2 is a light bulb with a diameter of 5 centimeters, an image width of a fisheye image captured by a fisheye lens is 864 pixels, and a field of view of the fisheye lens is 187 degrees.
TABLE 1
Distance
Object 1 Width
Object 2 Width
(centimeter)
(pixel)
(pixel)
100
2.64724885
13.23624425
90
2.941387611
14.70693806
80
3.309061063
16.54530531
70
3.781784072
18.90892036
60
4.412081417
22.06040709
50
5.2944977
26.4724885
40
6.618122126
33.09061063
30
8.824162834
44.12081417
20
13.23624425
66.18122126
10
26.4724885
132.3624425
5
52.944977
264.724885
Taking the object 2 with the capturing distance of 100 centimeters as an example, the width of the object 2 in the fisheye image captured is 13.23624425 pixels. After the parameters are introduced in formula (1), a distance r between the object 2 and the fisheye lens is calculated as:
After an azimuth angle, an incident angle, and a distance from the fisheye lens of the object are calculated, the processor 56 executes a coordinate calculation module 546, so as to calculate three-dimensional coordinates of the object by using a trigonometric function according to the azimuth angle, the incident angle, and the distance (step S612).
To be specific,
According to the azimuth angle φ, the incident angle θ, and the distance r, the three-dimensional coordinates of the object T is (x, y, z) may be derived from the formulas as follows:
x=
y=
z=
Through the foregoing method, the image pick-up apparatus 50 of this embodiment may carry out calculation of the three-dimensional coordinates of the object being captured without performing fisheye calibration, and have a capturing range greater than that of the conventional lenses.
In view of the foregoing, in the method and the image pick-up apparatus for calculating the coordinates of the object being captured by using the fisheye images, the incident angles of the light emitted by the object entering into the fisheye lenses and the azimuth angles of the object on the lens plane are calculated according to the locations of the object in the images captured by the fisheye lenses, and the three-dimensional coordinates of the object in the space are thus calculated. In another aspect, in the embodiments of the invention, the distances between the object and the fisheye lenses are calculated according to the widths of the object in the fisheye images. Finally, the three-dimensional coordinates of the object in the space are thereby calculated by using the trigonometric function. In this way, in the embodiments of the invention, the three-dimensional coordinates of the object being captured may be calculated without performing geometric conversion, and computation required to detect the object may be lowered while expanding the detection range.
It will be apparent to those skilled in the art that various modifications and variations can be made to the structure of the present invention without departing from the scope or spirit of the invention. In view of the foregoing, it is intended that the present invention cover modifications and variations of this invention provided they fall within the scope of the following claims and their equivalents.
Patent | Priority | Assignee | Title |
Patent | Priority | Assignee | Title |
10027883, | Jun 18 2014 | Amazon Technologies, Inc | Primary user selection for head tracking |
10063838, | Aug 11 2014 | SONY INTERACTIVE ENTERTAINMENT INC | Information processor, information processing method, and computer program |
10325339, | Apr 26 2016 | Qualcomm Incorporated | Method and device for capturing image of traffic sign |
10388027, | Jun 01 2016 | Kyocera Corporation | Detection method, display apparatus, and detection system |
10397524, | May 18 2016 | ULSEE INC | Three-dimensional around view monitoring system of vehicle and method thereof |
10469829, | May 22 2012 | SONY INTERACTIVE ENTERTAINMENT INC | Information processor and information processing method |
10551930, | Mar 25 2003 | Microsoft Technology Licensing, LLC | System and method for executing a process using accelerometer signals |
10552971, | May 15 2015 | HONOR DEVICE CO , LTD | Measurement method, and terminal |
10554945, | Jun 28 2012 | HITACHI ASTEMO, LTD | Stereo camera |
6005611, | May 27 1994 | B H IMAGE CO LLC | Wide-angle image dewarping method and apparatus |
9436973, | Jun 24 2013 | Mitsubishi Electric Corporation | Coordinate computation device and method, and an image processing device and method |
9532031, | Apr 08 2014 | The United States of America as represented by the Secretary of the Navy | Method for extrinsic camera calibration using a laser beam |
20010002936, | |||
20010014171, | |||
20020025154, | |||
20040167671, | |||
20040233280, | |||
20040234122, | |||
20060139619, | |||
20060165276, | |||
20070263924, | |||
20090167886, | |||
20100040279, | |||
20100150455, | |||
20110096319, | |||
20110102549, | |||
20110279653, | |||
20120154670, | |||
20130135439, | |||
20130210563, | |||
20140376780, | |||
20150077520, | |||
20150304612, | |||
CN103247024, | |||
CN103839227, | |||
CN104240236, | |||
CN104268867, | |||
TW200528945, | |||
TW201220828, | |||
TW201239807, | |||
TW356186, | |||
TW441514, | |||
TW558208, | |||
TW594453, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Sep 03 2018 | WU, PO-CHENG | Altek Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 047285 | /0768 | |
Oct 19 2018 | Altek Corporation | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Oct 19 2018 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Apr 22 2024 | REM: Maintenance Fee Reminder Mailed. |
Oct 07 2024 | EXP: Patent Expired for Failure to Pay Maintenance Fees. |
Date | Maintenance Schedule |
Sep 01 2023 | 4 years fee payment window open |
Mar 01 2024 | 6 months grace period start (w surcharge) |
Sep 01 2024 | patent expiry (for year 4) |
Sep 01 2026 | 2 years to revive unintentionally abandoned end. (for year 4) |
Sep 01 2027 | 8 years fee payment window open |
Mar 01 2028 | 6 months grace period start (w surcharge) |
Sep 01 2028 | patent expiry (for year 8) |
Sep 01 2030 | 2 years to revive unintentionally abandoned end. (for year 8) |
Sep 01 2031 | 12 years fee payment window open |
Mar 01 2032 | 6 months grace period start (w surcharge) |
Sep 01 2032 | patent expiry (for year 12) |
Sep 01 2034 | 2 years to revive unintentionally abandoned end. (for year 12) |