Embodiments of the present invention relate to unifying augmented reality technology and big data. An interactive operation element may be defined. The interactive operation element is associated with an event and a location on an augmented reality (AR) screen. An action may be performed based on the event using a predefined communication protocol. The action may be associated with an information artifact which is derived from big data.
|
1. An interactive method for unifying augmented reality (AR) or streaming video with big data, the method comprising the computer-implemented steps of:
obtaining an image of a set of real objects via a camera of a user device running an augmented reality application;
establishing a communications protocol between the augmented reality application and a semantic web knowledge based system;
submitting the image to the semantic web knowledge based system by the communication protocol;
receiving, in real-time, information on the set of real objects based on the image via the communication protocol from the semantic web knowledge based system and deriving from big data in the semantic web knowledge based system an information artifact based on the received information for each real object;
receiving a set of instructions over the communication protocol comprising a sequence performed on the set of real objects;
defining an interactive operation element in the augmented reality application associated with at least one of the set of real objects, wherein the interactive operation element is associated with a performance of at least one action of the sequence performed on the set of real objects;
associating the interactive operation element with a location on an augmented reality screen of the user device running the augmented reality application, the augmented reality screen showing a view of the set of real objects and the interactive operation element as an overlay over the view;
monitoring a performance of an action of the sequence on a real object of the set of real objects by a user; and
performing, in response to the action by the user, a subsequent action of the sequence as an augmented reality sequence on the view of the set of real objects in the augmented reality screen based on the information artifact derived from big data in the semantic web knowledge based system using the communication protocol.
11. A computer program product for unifying augmented reality (AR) and streaming video with big data, the computer program product comprising a computer readable hardware storage device, and program instructions stored on the computer readable hardware storage device, to:
obtain an image of a set of real objects via a camera of a user device running an augmented reality application;
establish a communications protocol between the augmented reality application and a semantic web knowledge based system;
submit the image to the semantic web knowledge based system by the communication protocol;
receive, in real-time, information on the set of real objects based on the image via the communication protocol from the semantic web knowledge based system and deriving from big data in the semantic web knowledge based system an information artifact based on the received information for each real object;
receive a set of instructions over the communication protocol comprising a sequence performed on the set of real objects;
define an interactive operation element in the augmented reality application associated with at least one of the set of real objects, wherein the interactive operation element is associated with a performance of at least one action of the sequence performed on the set of real objects;
associate the interactive operation element with a location on an augmented reality screen of the user device running the augmented reality application, the augmented reality screen showing a view of the set of real objects and the interactive operation element as an overlay over the view;
monitor a performance of an action of the sequence on a real object of the set of real objects by a user; and
perform, in response to the action by the user, a subsequent action of the sequence as an augmented reality sequence on the view of the set of real objects in the augmented reality screen based on the information artifact derived from big data in the semantic web knowledge based system using the communication protocol.
2. The interactive method of
4. The interactive method of
5. The interactive method of
8. The interactive method of
9. The interactive method of
10. The interactive method of
12. The computer program product of
14. The computer program product of
15. The computer program product of
16. The computer program product of
17. The computer program product of
18. The computer program product of
19. The computer program product of
20. The computer program product of
|
This application is the National Stage Entry of International Application No. PCT/KR2013/006014 filed on Jul. 5, 2013 which claims priority from and the benefit of Korean Patent Application No. 10-2013-0006014 filed on Aug. 28, 2012, both of which are herein incorporated by reference for all purposes as if fully set forth herein.
Embodiments of the present invention relate generally to unifying augmented reality technology and big data. Specifically, the present invention relates to the interaction between a user and virtual objects within an augmented reality environment.
Augmented reality (AR) is a live copy view of a physical, real-world environment whose elements are augmented (or supplemented) by computer-generated sensory input such as sound, video, graphics, or GPS data. It is related to a more general concept called mediated reality, in which a view of reality is modified by a computer. As a result, the technology functions by enhancing one's current perception of reality. By contrast, virtual reality replaces the real world with a simulated one. Augmentation is conventionally in real-time and in semantic context with environmental elements, such as sports scores on television (TV) during a sporting event. With the help of advanced AR technology (e.g., adding computer vision and object recognition), the information about the surrounding real world of the user becomes interactive and can be digitally manipulated. Artificial information about the environment and its objects can be overlaid on the real world.
Big data is the term for a collection of data sets so large and complex that it becomes difficult to process using on-hand database management tools or traditional data processing applications. The challenges include capture, curation, storage, search, sharing, transfer, analysis, and visualization. The trend to larger data sets is due to the additional information derivable from analysis of a single large set of related data, as compared to separate smaller sets with the same total amount of data, allowing correlations to be found to spot business trends, determine quality of research, prevent diseases, link legal citations, combat crime, and determine real-time roadway traffic conditions, and the like.
Embodiments of the present invention relate to unifying augmented reality technology and big data. An interactive operation element may be defined. The interactive operation element is associated with an event and a location on an augmented reality (AR) screen. An action may be performed based on the event using a predefined communication protocol. The action may be associated with an information artifact which is derived from big data.
A first aspect of the present invention provides an interactive method for unifying augmented reality (AR) and streaming video with big data, the method comprising the computer-implemented steps of: defining an interactive operation element, wherein the interactive operation element is associated with an event; associating the interactive operation element with a location on an augment reality screen; and performing an action based on the event using a communication protocol, wherein the action is associated with an information artifact derived from big data, wherein the big data is collected through the Semantic Web.
A second aspect of the present invention provides a computer program product for unifying AR and streaming video with big data, the computer program product comprising a computer readable storage media, and program instructions stored on the computer readable storage media, to: define an interactive operation element, wherein the interactive operation element is associated with an event; associate the interactive operation element with a location on an augment reality screen; and perform an action based on the event using a communication protocol, wherein the action is associated with an information artifact derived from big data, wherein the big data is collected through the Semantic Web.
These and other features of this invention will be more readily understood from the following detailed description of the various aspects of the invention taken in conjunction with the accompanying drawings in which:
The drawings are not necessarily to scale. The drawings are merely schematic representations, not intended to portray specific parameters of the invention. The drawings are intended to depict only typical embodiments of the invention, and therefore should not be considered as limiting the scope of the invention. In the drawings, like numbering represents like elements.
While the system and method of the present application is susceptible to various modifications and alternative forms, specific embodiments thereof have been shown by way of example in the drawings and are herein described in detail. It should be understood, however, that the description herein of specific embodiments is not intended to limit the invention to the particular embodiment disclosed, but on the contrary, the intention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the process of the present application as defined by the appended claims.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of this disclosure. As used herein, the singular forms “a”, “an”, and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. Furthermore, the use of the terms “a”, “an”, etc., do not denote a limitation of quantity, but rather denote the presence of at least one of the referenced items. It will be further understood that the terms “comprises” and/or “comprising”, or “includes” and/or “including”, when used in this specification, specify the presence of stated features, regions, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, regions, integers, steps, operations, elements, components, and/or groups thereof.
As indicated above, embodiments of the present invention relate to unifying augmented reality technology and big data. An interactive operation element may be defined. The interactive operation element is associated with an event and a location on an augmented reality (AR) screen. An action may be performed based on the event using a predefined communication protocol. The action may be associated with an information artifact which is derived from big data.
Augmented reality (AR) is divided into marker based technology and non-marker based technology. In marker based augmented reality, an image including a particular image of a marker such as a black and white pattern or a barcode is recognized and a relative coordinate of an area in which the virtual object is to be displayed is determined, and the virtual object is displayed based thereon; whereas, in non-maker based augmented reality, an object within the image is directly identified and related information is obtained. Since marker based approach can be included in the scope of non-marker based approach research, application of the present invention can be illustrated using “point of interest” based technology which is one of the sub-methods of non-marker based approach.
Augmented reality (AR) in the last decade has increased in popularity in various areas, such as education, advertising, maintenance, marketing, and entertainment. In the areas of maintenance and education specifically, the use of augmented reality can provide for the transfer of knowledge at a faster rate than other traditional methods. Additionally, the use of AR can help companies train their employees faster and better. The use of AR can also assist company employees in performing job tasks more competently and efficiently.
An area that can benefit from the use of AR is the maintenance of complex systems, such as aircraft maintenance. To that end, three-dimensional (3D) or two-dimensional (2D) graphics or images, text, or other media may be generated such that they are overlaid on and registered with surrounding objects in the environment. Applying AR to maintenance tasks could make it possible for users to be trained for those tasks, and actively assisted during their performance, without ever needing to refer to separate paper or electronic technical orders. Incorporating instruction and assistance directly within the task domain, and directly referencing the equipment at which the user is looking, may eliminate the need for maintenance employees to continually switch their focus of attention between the task and its separate documentation. Use of AR may decrease the overall time of the repair, reduce errors in the repair process, and increase the knowledge of maintenance personnel.
In step 104, images of the real-world objects are collected. In one example, images from various camera angles showing the front (representative image), left, right, top, and bottom of the real-world object are collected. The images are catalogued and stored in a database for easy retrieval.
In step 106, information artifacts are derived and assigned to each real-world object subject to recognition. These artifacts may include a label, circle, button, thumbnail, video, image, text, or the like. The information artifacts are virtual objects that can be overlaid on the actual equipment being maintained which can significantly improve the productivity of individuals performing maintenance tasks. As used herein, information artifacts are derived from big data.
In order for AR applications to merge virtual objects with real objects, the two sets of objects must be properly located and aligned in step 108. To that end, each representative image is associated with a homography matrix. In the field of computer vision, any two images of the same planar surface in space are related by a homography. Once camera rotation and translation have been extracted from an estimated homography matrix, this information can be used to insert a representative image of the real-world object into the scene, so that the representative image is rendered with the correct perspective and appear to have been part of the original scene.
In step 110, a search is performed for images of a real-world object when a camera is facing a real-world object that has been defined as subject to recognition. Real-time images generated by the camera are compared with images saved in the database to find a representative image from a similar angle.
In step 112, a homography matrix is calculated between the representative image and the real-world display. The homography matrix is used to alter the direction and location of the representative image in order to properly align the representative image and/or information artifacts in the augmented display.
In step 114, a user may interact with the overlaid information artifacts. Detailed information associated with information artifacts may be displayed based on the user's selected method of action (e.g., a single or double mouse click, mouse over, voice control, gesture, or the like).
Advancement of augmented reality technology has brought attempts of practical application of augmented reality technology to industries. However, in the case of large amount of artifact data, conventional technology for an interactive method and system for unifying augmented reality technology and big data has not been ideal. Until now, augmented reality took on relatively small amounts of data, making it easy to maintain artifact information. However, with an increasing amount of artifact data being needed, there is a need for an interactive system and method for unifying information artifacts with augmented reality.
Conventional AR technology is based on computer vision technology in which focus is placed on real-world object recognition, tracking, and information artifact composition technologies. However, as stated above, if a vast amount of information artifact data is needed for augmenting reality, artificial intelligence such as a knowledge based system (KBS) is needed. A KBS is a computer program that reasons and uses a knowledge base to solve complex problems. The term is broad and is used to refer to many different kinds of systems. The one common theme that unites all knowledge based systems is an attempt to represent knowledge explicitly via tools such as ontologies and rules rather than implicitly via code the way a conventional computer program does. A knowledge based system has at least one and usually two types of sub-systems: a knowledge base and an inference engine.
A KBS is independent from computer vision technology and needs to be maintained using a heterogeneous software (S/VV) platform. Particularly since KBS's vast data continues to be generated, revised, supplemented, and maintained through the medium of the Internet, the direction of the technology is progressing towards more structure where electronic documents are maintained by ontology technology based on Semantic Web.
Since augmented reality technology's S/VV platform is fundamentally different than KBS-based S/W, as well as programming language for both, the present invention defines an interaction method and protocol for two different platform systems, so that the offered solution displays information artifacts (defined in step 106) matching an AR's context (displayed in step 114). This is provided by KBS through an interaction method as well as a unifying protocol, described in detail below.
An interactive method is provided for unifying AR and big data. The method includes three steps. First, each element to offer interaction between a user and an information artifact to be displayed on an AR screen is defined. Second, the location where the information artifact is to be displayed on an AR screen is defined. Third, a communication protocol for the interaction of an AR engine and big data is defined.
Display of information artifacts (i.e., overlaid virtual objects) may be based on one or more predefined user actions. User actions may include, but are not limited to, single or double mouse click, mouse over, scroll, gesture, voice command, or other predefined user action. Additionally, display of information artifacts may be based on the type of information artifact, AR situation or context, and/or a user action.
However, as the data size increases, it becomes difficult for the conventional system shown in
For example, to develop an AR system for aircraft maintenance, initiation of simultaneous development of an AR engine module and large data processing module regarding aircraft manual is ideal for effective temporal reduction. As such, since conventional AR technology has been taking on relatively small amounts of data, it has been relatively easy to handle information artifacts. However, with increasing amount of data being used in AR, there is a need for a more effective system and method for unifying AR and information artifacts.
The present invention described herein focuses on two of the seven steps described in
Each information artifact may include any number of interactive operation elements that offer interaction between a user and an information artifact (i.e., virtual object) displayed on the AR screen. Interactive operation elements may be associated with user actions defined in interactive operation element definition 202. When a user performs a particular action, a predefined event may be triggered. Essentially, each interactive operation element is associated with a hotspot (or area on the screen) which, when acted upon, triggers a predefined event. The event may include providing additional information on the real-world object associated with the area selected by displaying an information artifact or linking the user to further related content (e.g., the next step in a maintenance sequence). Exemplary user actions include, but are not limited to, single or double mouse click, physical gestures, and voice commands.
Interactive operation element registration 204 defines the relationship between each interactive operation element and interactive artifact. In other words, interactive operation elements and associated user actions are defined, along with the events to be performed when the user actions are executed. Interactive operation element registration 204 may include, among others, circle registration 204A, label registration 204B, button registration 204C, menu registration 204D, and animation registration 204E.
Communication protocol definition 206 defines the communication protocol between AR and big data. Communication protocol definition 206 may include, among others, message ID definition 206A, source ID definition 206B, destination ID definition 206C, button name definition 206D, mouse over type definition 206E, instruction number definition 206F, event type definition 206G, and action type definition 206H. Communication protocol definition 206 defines the communication protocol used to facilitate the event to be performed based on the user action.
For illustrative purposes, two user actions associated with interactive operation element are discussed below. First, a “mouse over” event can take place. A mouse over event may be defined as a default value for the composition and display of an information artifact. In other words, if a user performs a mouse over event over an interactive operation element in the real world screen (invisible to the user), an associated information artifact may be displayed.
Performing an event may expose a large amount of data which may result in the user's loss of interest or discomfort. At the same time, augmenting multiple artifacts on a single screen may incur repetition issues. Mouse over is therefore a solution to these issues. When an augmented artifact is invisible and a user moves the mouse over an area of interest, such request is recognized and an associated information artifact becomes visible to the user.
Second, a user may perform a mouse click event on a visible information artifact. The user may perform this event when the user wants to display additional information or move to a subsequent screen. The information artifact may be displayed either in a circle or with a description label. The mouse click event may pop up a menu or link to an external site, altering the user's point of interest as well as transitioning to a subsequent screen.
Referring back to
Connection between an information artifact and interactive operation element may occur in one of the two ways: explicitly or implicitly. An explicit connection is made when direct display of an information artifact (or annotation) file (e.g. XML or RDF) [please fix this sentence and is there an explanation of “implicit connection”?. The Resource Description Framework (RDF) has come to be used as a general method for conceptual description or modeling of information that is implemented in web resources, using a variety of syntax notations and data serialization formats. An implicit connection can be made with a coding label.
In one example, fields names may include, but not be limited to, MESSAGE_ID, SOURCE, DESTINATION, SLIDE_ID, BUTTON_NAME, MOUSE_OVER_TYPE, INSTRUCTION_NUMBER, EVENT_TYPE, and ACTION_TYPE. Field names and field values may include alpha numeric characters and special characters.
The invention will be described in more detail with reference to examples described in
If the interactive operation element (‘Gear and Door’ button as shown in
A mouse over event can be used to display a partial amount of data from a large amount of data based on a user request. A mouse over event can be defined to display an information artifact (or annotation). As such, a user can mouse over to an interested part's location to display the information. When moving a mouse to a selected part from
When a mouse is moved over a particular location, a ‘Mouse Over Event Message (In)’ message can be transmitted from AR to KBS. When the message is received, the KBS transmits a response message to AR. Based on the real-time context, KBS sends a ‘View Change Message’ message as a response. To AR, receipt of the ‘View Change Message’ means to display the information artifact ‘pin-Main Fitting (Upper)’ on the AR screen. The information artifact is shown in
When KBS receives the ‘Mouse Over Event Message (Out)’ message, KBS transmits a response message to AR. KBS then transmits a ‘View Change Message’ to AR. This message says ‘move context to No. 8’ to AR. When in receipt of this message, AR removes the information artifact from the AR screen, as shown in
In a real-time AR screen like
Each information artifact may include any number of interactive operation elements that offer interaction between a user and an information artifact (i.e., virtual object) displayed on the AR screen. Interactive operation elements may be associated with user actions defined by element definition unit 1310. When a user performs a particular action, a predefined event may be triggered. Essentially, each interactive operation element is associated with a hotspot (or area on the screen) which, when acted upon, triggers an event defined by element definition unit 1310. The event may include providing additional information on the real-world object associated with the area selected by displaying an information artifact or linking the user to further related content (e.g., the next step in a maintenance sequence). Exemplary user actions include, but are not limited to, single or double mouse click, physical gestures, and voice commands.
Element registration unit 1304 defines the relationship between each interactive operation element and interactive artifact. In other words, interactive operation elements and associated user actions are defined, along with the events to be performed when the user actions are executed. Element registration may include, among others, circle registration, label registration, button registration, menu registration, and animation registration.
Communication protocol unit 1306 defines the communication protocol between AR and big data. Communication protocol definition may include, among others, message ID definition, source ID definition, destination ID definition, button name definition, mouse over type definition, instruction number definition, event type definition, and action type definition. Communication protocol unit 1306 defines the communication protocol used to facilitate the event to be performed based on the user action.
The embodiments of the invention may be implemented as a computer readable signal medium or storage device, which may include a propagated data signal with computer readable program code embodied therein (e.g., in baseband or as part of a carrier wave). Such a propagated signal may take any of a variety of forms including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on a computer readable medium may be transmitted using any appropriate medium including, but not limited to, wireless, wireline, optical fiber cable, radio-frequency (RF), etc., or any suitable combination of the foregoing.
As used herein, it is understood that the terms “program code” and “computer program code” are synonymous and mean any expression, in any language, code, or notation, of a set of instructions intended to cause a computing device having an information processing capability to perform a particular function either directly or after either or both of the following: (a) conversion to another language, code, or notation; and/or (b) reproduction in a different material form. To this extent, program code can be embodied as one or more of: an application/software program, component software/a library of functions, an operating system, a basic device system/driver for a particular computing device, and the like.
A data processing system suitable for storing and/or executing program code can be provided hereunder and can include at least one processor communicatively coupled, directly or indirectly, to memory elements through a system bus. The memory elements can include, but are not limited to, local memory employed during actual execution of the program code, bulk storage, and cache memories that provide temporary storage of at least some program code in order to reduce the number of times code must be retrieved from bulk storage during execution. Input/output and/or other external devices (including, but not limited to, keyboards, displays, pointing devices, etc.) can be coupled to the system either directly or through intervening device controllers.
Network adapters also may be coupled to the system to enable the data processing system to become coupled to other data processing systems, remote printers, storage devices, and/or the like, through any combination of intervening private or public networks. Illustrative network adapters include, but are not limited to, modems, cable modems, and Ethernet cards.
The foregoing description of various aspects of the invention has been presented for purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise form disclosed and, obviously, many modifications and variations are possible. Such modifications and variations that may be apparent to a person skilled in the art are intended to be included within the scope of the invention as defined by the accompanying claims.
Although presently preferred embodiments of the present invention have been described in detail herein above, it should be clearly understood that many variations and/or modifications of the basic inventive concepts herein taught, which may appear to those skilled in the pertinent art, will still fall within the spirit and scope of the present invention, as defined in the appended claims.
Patent | Priority | Assignee | Title |
10065750, | Feb 10 2015 | Honeywell International Inc. | Aircraft maintenance systems and methods using wearable device |
Patent | Priority | Assignee | Title |
6037936, | Mar 27 1995 | Criticom Corp. | Computer vision system with a graphic user interface and remote camera control |
8037108, | Jul 22 2009 | Adobe Inc | Conversion of relational databases into triplestores |
20100023878, | |||
20120075343, | |||
20120116728, | |||
20130114100, | |||
WO2014035041, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Apr 30 2014 | JO, GEUN SIK | JO, GUEN SIK | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 032858 | /0631 | |
Apr 30 2014 | JO, GEUN SIK | INHAU UNIVERSITY | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 032858 | /0631 | |
May 09 2014 | Geun Sik, Jo | (assignment on the face of the patent) | / | |||
Feb 11 2016 | INHA INDUSTRY PARTNERSHIP INSTITUTE | JO, GEUN SIK | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 037962 | /0857 |
Date | Maintenance Fee Events |
Feb 05 2021 | M2551: Payment of Maintenance Fee, 4th Yr, Small Entity. |
Date | Maintenance Schedule |
Aug 08 2020 | 4 years fee payment window open |
Feb 08 2021 | 6 months grace period start (w surcharge) |
Aug 08 2021 | patent expiry (for year 4) |
Aug 08 2023 | 2 years to revive unintentionally abandoned end. (for year 4) |
Aug 08 2024 | 8 years fee payment window open |
Feb 08 2025 | 6 months grace period start (w surcharge) |
Aug 08 2025 | patent expiry (for year 8) |
Aug 08 2027 | 2 years to revive unintentionally abandoned end. (for year 8) |
Aug 08 2028 | 12 years fee payment window open |
Feb 08 2029 | 6 months grace period start (w surcharge) |
Aug 08 2029 | patent expiry (for year 12) |
Aug 08 2031 | 2 years to revive unintentionally abandoned end. (for year 12) |