A method for mapping 2-dimensional point of interest to a 3-dimensional view. The disclosed method includes capturing a 3-dimensional (3d) image with an augmented reality device. Matching images of a 2d image database with the captured 3d-image, the 2d image database containing 2-dimensional (2d) images associated with points of interest having a 2-dimensional (2d) data set. When the 2d-image matches the 3d image, capturing 3d-data for the matched 2d-image and converting the 2d-data set of the point-of-interest to a 3d-data set. The disclosed method simplifies the configuration of points of interest for 3d applications, such as an augmented reality device.
|
1. A method for mapping a 2-dimensional (2d) point of interest in a 3-dimensional (3d) view, comprising:
in a learning phase, capturing 3d images that match 2d images previously stored in a 2d image database and associating the 3d images with the 2d images in the 2d image database;
outside of the learning phase:
capturing a 3d image with an augmented reality device, the captured 3d image comprising a 2d image captured by an image sensor and associated depth information captured by a laser 3d sensor based on a matrix of points corresponding to the captured 2d image, the captured 2d image and associated depth information being a 2d frame constructed as a vector field with a vector corresponding to depth associated with each point in the 2d image;
matching the 2d images of the 2d image database with the captured 3d image based on the 3d images associated with the 2d images in the 2d image database, the 2d image database comprising 2d images associated with points of interest having a 2d data set,
wherein when a 2d image of the 2d image database matches the captured 3d image:
capturing 3d data for the matched 2d image, and
converting the 2d data set to a 3d data set,
wherein the captured 3d image and the images in the 2d data base are in spherical coordinates.
10. An augmented reality device for capturing a 3-dimensional (3d) image, comprising:
an adjustable lens;
an image sensor;
laser 3d sensors;
an absolute positioning sensor;
a display; and
a processing unit configured to:
in a learning phase, capture 3d images that match 2d images previously stored in a 2d image database and associate the 3d images with the 2d images in the 2d image database;
outside of the learning phase:
match a 2d image from the 2d image database with a 3d image captured by the image sensor based on the 3d images associated with the 2d images in the 2d image database, the 2d image database comprising 2d images associated with points of interest having a 2d data set, the captured 3d image comprising a 2d image captured by the image sensor and associated depth information captured by the laser 3d sensors based on a matrix of points corresponding to the captured 2d image, the captured 2d image and associated depth information being a 2d frame constructed as a vector field with a vector corresponding to depth associated with each point in the 2d image,
capture 3d data of the captured 3d image from the 3d sensors and the adjustable lens when the 2d image from the 2d database matches the captured 3d image,
convert the 2d data set to a 3d data set, and
display the points of interest on the display in a 3d view as augmented reality,
wherein the captured 3d image and the images in the 2d data base are in spherical coordinates.
12. An augmented reality device for capturing a 3-dimensional (3d) image, comprising:
an adjustable lens;
an image sensor;
laser 3d sensors;
an absolute positioning sensor; and
a display;
a 2-dimensional (2d) database comprising 2d images with points of interest having 2d data sets;
a 3d database configured to store 3d images with points of interest having 3d data sets; and
a processing unit configured to:
in a learning phase, associate the 3d images in the 3d database with the 2d images in the 2d image database;
outside of the learning phase:
match a 2d image from the 2d database with a 3d image captured by the image sensor based on the 3d images associated with the 2d images in the 2d image database, the captured 3d image comprising a 2d image captured by the image sensor and associated depth information captured by the laser 3d sensors based on a matrix of points corresponding to the captured 2d image, the captured 2d image and associated depth information being a 2d frame constructed as a vector field with a vector corresponding to depth associated with each point in the 2d image,
capture 3d data of the captured 3d image from the 3d sensors and the adjustable lens when the 2d image from the 2d database matches the captured 3d image,
convert the 2d data set to a 3d data set, and
display the points of interest on the display in a 3d view as augmented reality,
wherein the captured 3d image and the images in the 2d data base are in spherical coordinates.
2. The method according to
providing the 2d images; and
for each image of the 2d images:
assigning an image identifier,
determining length and width of the image,
defining at least one point of interest having a respective poi-identifier,
determining first coordinates of the at least one point of interest in relation to a reference point of said each image,
associating each defined at least one point of interest with the image identifier, and
storing the image identifier, the length and width, the reference point, the poi-identifier, the first coordinates of the at least one point of interest, and the association with said each defined at least one point of interest.
3. The method according to
wherein capturing the 3d image comprises obtaining a 3d representation of the captured 3d image via 3d sensors of the augmented reality device,
wherein the 3d representation comprises a 2d frame consisting of points defined by 2d frame coordinates, and
wherein each point of said points is associated with a depth.
4. The method according to
5. The method according to
determining frame coordinates of a reference point of the matched 2d image in a 2d frame;
determining scale information from a size of the 2d image and a size of the 2d frame; and
obtaining a distance between the augmented reality device and the 2d image.
6. The method according to
applying scaling information to first coordinates of the points of interest; and
applying frame coordinates of a reference point to the first coordinates of the points of interest to obtain frame coordinates of the points of interest.
7. The method according to
a depth associated with the obtained frame coordinates of the points of interest; and
the scaling information.
8. The method according to
obtaining an absolute position of the augmented reality device;
determining an absolute position of the points of interest by applying a global position of the augmented reality device to the frame coordinates and the depth coordinates of the points of interest; and
storing the absolute position of the points of interest in the 3d data set.
9. A method for displaying a 2-dimensional (2d) point of interest in a 3-dimensional (3d) view, comprising:
providing a 3d image database comprising points of interest having a 3d data set obtained by the method according to
displaying the points of interest in the 3d view as augmented reality.
11. A system for providing an augmented reality, comprising:
an augmented reality device according to
an augmented reality image server comprising:
a 2-dimensional (2d) database comprising images with points of interest having 2d data sets, and
a 3-dimensional (3d) database comprising images with points of interest having 3d data sets.
|
The present invention relates to a method for 3D mapping of 2D Point of interest and an augmented reality device. In particular, the method relates to mapping a point of interest defined for a 2-Dimensional image in a 3-Dimensional image captured by an augmented reality device.
In mapping and navigation applications, points of interest are commonly applied to indicate where a particular object may be found. Such objects may be resources, such as gas stations, restaurants, or may be parts or components. A rough indication of the location where such objects are to be found is usually enough, but in certain conditions this may not be sufficient. In particular, in factory sites or other type of manufacturing plants, the precise location of components present in a machine, control cabinet or other industrial equipment, is required to prevent unintentional damage or even hazardous situations. The knowledge of where precisely to find and recognize parts in such circumstances, is of high importance. Prior art solutions include layout drawings and machine diagrams, pointing out the layout or structure of a particular machine. Enhanced versions thereof include actual pictures i.e. photo images of the particular machine or electrical control box. However, a configuration may be altered over time. For example, when parts are replaced or components are updated.
To address the above mentioned needs, operator and maintenance personnel may be instructed to re-take an image each time a particular machine, control cabinet or equipment is inspected or at least interacted with. Or, when new equipment is installed or equipment is updated, to take pictures of the newly installed configuration.
With these images Points of interest may be associated by assigning coordinates within the captured images. The operator may identify points-of-interest in the image at the moment of taking a picture, or it may also be performed by at a later stage. Or it may be performed by image-recognition software capable of identifying predefined components, equipment or other relevant parts that are of interest. Either way, these points of interest are defined in a 2D image that relate to a certain relevant part or piece of information. This allows creating a database of images having associated Points of interest that are defined in a 2-dimensional (2D) space. Hence, the captured image is treated as a 2D canvas i.e. surface space which may spanned by Cartesian coordinates or defined by any other suitable coordinate system.
It is an object of the invention to further alleviate the definition of points of interest for 3-dimensional applications.
According to one aspect, there is provided a method for mapping 2-dimensional point of interest to a 3-dimensional image. The method includes capturing a 3-dimensional (3D) with an augmented reality device. Matching images of a 2D image database with the 3D, the 2D image database containing 2-dimensional (2D) images associated with points of interest having a 2-dimensional (2D) data set. And when the 2D-image matches the 3D, capturing 3D-data for the matched 2D-image and converting the 2D-data set of the point-of-interest to a 3D-data set.
The disclosed method simplifies the configuration of Points of interest for 3D applications, such as an augmented reality device. As e.g. only X,Y coordinates need to be associated with a 2D image.
The invention further reduces the effort of defining Points of interest, as this may still be done in 2D images. Or may have been done previously, thus allowing the re-use of a prior created database with 2D images containing points of interest.
The invention further reduces the amount of data required to be stored as only 2D images and associated 3D absolute position coordinates are to be stored, and not full 3D images and associated content.
In this application, any 3D image has a 3D representation consisting of a 2D image i.e. 2D frame with associated depth information for a number of points or all points in the 2D frame.
By way of example only, the embodiments of the present disclosure will be described with reference to the accompanying drawing, wherein:
Referring to
In
The augmented reality i.e. 3D device 7 has an image sensor 10, an adjustable lens 11, and 3D-sensors 12 which are able to determine a depth i.e. distance between the 3D device and a visible object. The 3D device further has an absolute positioning sensor 13, a GPS sensor in this example, and a display 14. Further shown in
In this example the absolute position sensor is a GPS sensor, but other navigational sensors capable to determine the position of the device in relation to a reference point are also suited, such as e.g. the EU Galileo positioning system or China's BeiDou Navigation Satellite System. Other position sensors that allow to obtain an absolute position within a geographical space are also suited, such as position sensor and system that allows to define a fixed known reference point and determine the device position relative to that defined reference position.
Referring to
The above steps may be regarded as a learning phase wherein the 3D data base on the AR-server is populated with 3D data sets of the points of interest available for later retrieval. In this manner, 2D image recognition of one image is only required once, as the 3D data set of each point of interest will be stored with global position coordinates.
Turning to
One advantage of the above described methods is that this allows points of interest to be defined in 2D images in a very simple manner, while in the learning phase all points of interest may be instantiated in a 3D-image for the application by means of the 2D recognition.
Referring to
Thereto each image is assigned 302 an image identifier ID-im as a unique means for referring to the image. Further properties as length L and width W of the image are determined 303, as is a reference point. The reference point may be one of corners of the image, it may be the centre of the image, or it may be any other reference point suitable to be taken as an origin for a coordinate system, such as the origin of a Cartesian system. In the latter case, with the origin on one corner, maximum values for X and Y will be equal to L and W.
With the 2D image being processed, a point of interest having a respective poi-identifier ID-poi may be defined 304 within the image. Along with the defining of the point of the interest and its poi-identifier, first coordinates of the point of interest in relation to the reference point, i.e., origin, of the image are determined 305; for example, X-poi and Y-poi, for which purpose the 2D image may be regarded as a Cartesian coordinate space R. The above may be repeated for multiple points, as one 2D image may show multiple points of interest. In this manner, each point of interest has a 2-dimensional (2D) data set that at least contains the coordinates of the poi in the 2D image. Such a data set may for example take the form for each P.o.I. as (ID-poi, X-poi, Y-poi, ID-im).
Creating the 2D image database further requires associating 306 each defined point of interest with the image identifier. This may be done by creating a record for each image identifier and storing all poi-identifiers of the respective points of interest associated with the image and its image identifier. It may also be done by creating a record for each poi-identifier and storing the image identifier associated therewith.
All the above is stored 307 in the 2D image database, so the image identifier, the length and width of the image, the reference point in the image, the poi-identifier, the coordinates of the at least one point of interest, and the association with each defined point of interest. Accordingly, a 2D image database is provided containing 2-dimensional (2D) images associated with points of interest having a 2-dimensional (2D) data set.
Returning to
For example, the 3D device may use laser to apply a matrix projection to determine the deformation caused by objects in front of it. Such a projection will only obtain values for Z i.e. depth for a limited number of points depending on the mesh size or pitch of the matrix projection. In yet another way, the 2D frame may be construed as a projection of the 3D-image along the Z-axis onto the X-Y plane.
The matching of images of the 2D image database with the 3D-image includes comparing each 2D image of the 2D image database with the 2D frame of the 3D representation by scaling and translating the 2D image over the 2D frame. An example of the required processing movements is shown in a simplified manner in
When the 2D image matches the 3D image, so in the above example the 2D frame of the 3D representation, the 3D data for the corresponding match needs to be captured. This capturing of 3D data for the matched 2D image includes determining frame coordinates of the reference point of the 2D image in the 2D frame of the 3D representation. The capturing further includes determining scale information, and obtaining a distance D between the augmented reality device and the 2D image; both of which may be obtained from the lens of the 3D device. One common way of performing scaling is a homothetic projection, which allows resizing of a shape while maintaining its ratio aspects and with minimal deformation.
The frame coordinates, scale information, distance D and image identifier of the matched image may be stored in a matched image object as an intermediate result.
With the 3D-data captured, the 2D-data set of the point-of-interest can be converted to a 3D-data set. This converting of the 2D-data set to the 3D-data set includes applying the scaling information to the first coordinates of the point of interest. The converting further includes applying the frame coordinates of the reference point to the coordinates of the point of interest to obtain frame coordinates of the point of interest. And finally, to obtain a depth coordinate for the point of the interest, the depth associated with frame coordinates of the point of interest and the scaling information is processed to calculate the depth coordinate. In this manner, a set of 3-dimensional coordinates of the point of interest is obtained, defined in the coordinate system of the augmented reality device R′.
These frame coordinates may be converted to global position coordinates, such as e.g; known as GPS. Thereto, a global position of the augmented reality device is obtained. From this a global position of the point of interest is obtained by applying the global position of the augmented reality device to the frame and depth coordinates of the point of interest. These 3D data may then be stored as the global position of the point of interest in the 3D data set.
Referring to
The 2D images are processed one by one by the 2D recognition tool to be matched 401 with captured 3D image. When matched a match-image object is created containing the image identifier ID-im, updated coordinates X′ and Y′ defining the coordinates of the origin in the captures 3D image, and scale-information defining the scale of the 2D image in the captured 3D image. As the captured 3D image may show multiple points-of-interest, the 3D image is compared with further 2D images of the 2D image database; if further matches are obtained these are also stored in the match-image object. Therefore, in case of a match it is checked whether this is a new match 402 or not. As the location of the 3D device may be known the number of 2D images relevant to be compared may be limited based on the location.
When the 2D image is matched to the 3D image the 2D data X-poi, Y-poi are converted 403 by the 2D to 3D module to one 3D data set which is stored in the 3D database 9. The conversion is performed by processing the match-image objects for each 2D image associated with the 3D image. The 2D image matching may by itself also contain multiple points-of-interest, in which case the converting of the 2D data set is applied to all points-of-interest defined for that image. All the converted 3D data sets are stored in the 3D database 9.
When a 2D image was previously matched to 3D image, the converting may be skipped as prior converted 3D data set will be available. Hence, only when a new matched image is obtained the converting of the 2D to 3D data set will be performed.
Once the 3D data set is obtained, the 3D image will be shown on the display of the 3D device together with all identified points-of-interest.
Although the present invention has been described above with reference to specific embodiments, it is not intended to be limited to the specific form set forth herein. Rather, the invention is limited only by the accompanying claims and, other embodiments than the specific above are equally possible within the scope of these appended claims.
Furthermore, although exemplary embodiments have been described above in some exemplary combination of components and/or functions, it should be appreciated that, alternative embodiments may be provided by different combinations of members and/or functions without departing from the scope of the present disclosure. In addition, it is specifically contemplated that a particular feature described, either individually or as part of an embodiment, can be combined with other individually described features, or parts of other embodiments.
Gassion, Romain, Chiche, Thierry
Patent | Priority | Assignee | Title |
11770517, | Oct 28 2019 | SONY GROUP CORPORATION | Information processing apparatus and information processing method |
Patent | Priority | Assignee | Title |
20120026290, | |||
20120310098, | |||
20150071524, | |||
20150091905, | |||
20160296291, | |||
20170337690, | |||
WO2016026437, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Oct 11 2016 | CHICHE, THIERRY | Schneider Electric Industries SAS | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 043798 | /0252 | |
Oct 11 2016 | GASSION, ROMAIN | Schneider Electric Industries SAS | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 043798 | /0252 | |
Oct 05 2017 | Schneider Electric Industries SAS | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Oct 05 2017 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Date | Maintenance Schedule |
Jul 27 2024 | 4 years fee payment window open |
Jan 27 2025 | 6 months grace period start (w surcharge) |
Jul 27 2025 | patent expiry (for year 4) |
Jul 27 2027 | 2 years to revive unintentionally abandoned end. (for year 4) |
Jul 27 2028 | 8 years fee payment window open |
Jan 27 2029 | 6 months grace period start (w surcharge) |
Jul 27 2029 | patent expiry (for year 8) |
Jul 27 2031 | 2 years to revive unintentionally abandoned end. (for year 8) |
Jul 27 2032 | 12 years fee payment window open |
Jan 27 2033 | 6 months grace period start (w surcharge) |
Jul 27 2033 | patent expiry (for year 12) |
Jul 27 2035 | 2 years to revive unintentionally abandoned end. (for year 12) |