An image processing system including at least one processor configured to receive at least one first image of an inoperative product captured by a mobile device of a user; perform image analysis on the first image to identify a status of a functional element associated with the inoperative product; access memory to determine why the product is inoperative; cause visual guidance to be displayed by the mobile device, wherein the visual guidance is associated with a plurality of sequential actions for causing the inoperative product to become operative; receive at least one second image of the product, the second image being indicative of a completion of the plurality of sequential actions; perform image analysis on the at least one second image to determine that the completion of the plurality of sequential actions caused the inoperative product to become operative; and notify the user that the problem has been resolved.
|
20. A method for assisting a user during technical support, the method comprising:
receiving at least one first image of an inoperative product captured by a mobile device of a user;
performing image analysis on the at least one first image to identify in the at least one first image a status of a functional element associated with the inoperative product;
accessing memory to determine a reason why the product is inoperative based on the identified status of the functional element;
causing visual guidance to be displayed by the mobile device, wherein the visual guidance is associated with at least one action for causing the inoperative product to become operative;
receiving at least one second image of the product, the second image being indicative of a completion of the at least one action;
performing image analysis on the at least one second image to determine whether the at least one action was completed properly; and
notifying the user when the at least one action is completed properly.
11. An image processing system employing artificial intelligence for assisting a user during technical support, the image processing system, comprising:
at least one processor configured to:
receive at least one first image of an inoperative product captured by a mobile device of a user;
perform image analysis on the at least one first image to identify in the at least one first image a status of a functional element associated with the inoperative product;
access memory to determine a reason why the product is inoperative based on the identified status of the functional element;
cause visual guidance to be displayed by the mobile device, wherein the visual guidance is associated with a plurality of sequential actions for causing the inoperative product to become operative;
receive at least one second image of the product, the second image being indicative of a completion of the plurality of sequential actions;
perform image analysis on the at least one second image to determine that the completion of the plurality of sequential actions caused the inoperative product to become operative; and
provide a notification to the user indicating problem resolution.
1. A non-transitory computer readable medium for assisting a user during technical support, the computer readable medium including instructions that, when executed by at least one processor, cause the at least one processor to perform a method, the method comprising:
receiving at least one first image of an inoperative product captured by a mobile device of a user;
performing image analysis on the at least one first image to identify in the at least one first image a status of a functional element associated with the inoperative product;
accessing memory to determine a reason why the product is inoperative based on the identified status of the functional element;
causing visual guidance to be displayed by the mobile device, wherein the visual guidance is associated with a plurality of sequential actions for causing the inoperative product to become operative;
receiving at least one second image of the product, the second image being indicative of a completion of the plurality of sequential actions;
performing image analysis on the at least one second image to determine that the completion of the plurality of sequential actions caused the inoperative product to become operative; and
providing a notification to the user indicating problem resolution.
2. The computer readable medium of
3. The computer readable medium of
4. The computer readable medium of
5. The computer readable medium of
6. The computer readable medium of
7. The computer readable medium of
confirming that a product included in the at least one second image is a same product captured in the at least one first image.
8. The computer readable medium of
receiving at least one additional image of the inoperative product upon completion of at least one of the sequential actions;
analyzing the at least one additional image to identify in the additional image a change in an operational status of the inoperative product;
accessing memory to determine a different set of sequential actions for causing the inoperative product to become operative, based on the change in the operational status; and
causing visual guidance associated with the different set of sequential actions to display.
9. The computer readable medium of
10. The computer readable medium of
12. The image processing system of
13. The image processing system of
14. The image processing system of
detect that the inoperative product is outside a field of view of an image sensor of the mobile device; and
suspend the visual guidance associated with a plurality of sequential actions while the inoperative product is outside the field of view.
15. The image processing system of
16. The image processing system of
17. The image processing system of
18. The image processing system of
19. The image processing system of
|
This application is a continuation-in-part of U.S. patent application Ser. No. 16/392,972, filed Apr. 24, 2019, which is a continuation of U.S. patent application Ser. No. 16/196,818, filed Nov. 20, 2018, which is a continuation of U.S. patent application Ser. No. 15/366,483, filed Dec. 1, 2016 (now U.S. Pat. No. 10,182,153). This application also claims priority to U.S. Provisional Patent Application No. 62/669,407, filed on May 10, 2018, all of which are incorporated herein by reference.
The present disclosure generally relates to the field of image processing, and particularly relates to systems and methods for using image processing to assist a user with technical support.
Technical support systems utilized nowadays make it difficult for digital service providers (DSPs), and especially for service/technical support centers, to provide efficient (in terms of time and customer satisfaction) technical support services to their customers. Despite a recent push toward self-service schemes, customer have been slow to adopt self-service technologies. Today's customers support model and relevant technologies are subject to the numerous challenges, including increasingly complex customer needs, communication gaps, diagnosis challenges, limited problem solving rates, and customer frustration.
The techniques disclosed in this disclosure aim to provide remote efficient consumer support services and reduce the incidence of technician dispatch. These techniques are useful for shortening consumer wait time, improving installation and repair outcomes, and improving customer satisfaction and independence.
One aspect of the disclosure provides an image processing system for visually augmenting a real-time video stream, the image processing system, comprising: at least one processor configured to: receive the real-time video stream captured by an image sensor, the real-time video stream including images of at least one cable and an electronic appliance; analyze the real-time video stream to identify a plurality of ports in the electronic appliance; analyze the real-time video stream to identify a specific cable for association with a specific port of the plurality of ports; cause a movable augmented indicator to display on the real-time video stream, wherein the movable augmented indicator is configured to guide a user's connection of the specific cable to the specific port; monitor in the real-time video stream changing locations of the specific port as the image sensor moves relative to the electronic appliance; and adjust positions of the movable augmented indicator to account for the changing locations of the specific port in the real-time video stream
Another aspect of the disclosure provides a non-transitory computer readable medium including instructions for visually augmenting a real-time video stream, the instructions being executable by at least one processor to cause the at least one processor to perform a method, the method comprising: receiving the real-time video stream captured by an image sensor, the real-time video stream including images of at least one cable and an electronic appliance; analyzing the real-time video stream to identify a plurality of ports in the electronic appliance; analyzing the real-time video stream to identify a specific cable for association with a specific port of the plurality of ports; causing a movable augmented indicator to display on the real-time video stream, wherein the movable augmented indicator is configured to guide a user's connection of the specific cable to the specific port; monitoring in the real-time video stream changing locations of the specific port as the image sensor moves relative to the electronic appliance; and adjusting positions of the movable augmented indicator to account for the changing locations of the specific port in the real-time video stream.
Another aspect of the disclosure provides a method for visually augmenting a real-time video stream captured by an image sensor, the method comprising: receiving the real-time video stream captured by an image sensor, the real-time video stream including images of at least one cable and an electronic appliance; analyzing the real-time video stream to identify a plurality of ports in the electronic appliance; analyzing the real-time video stream to identify a specific cable for association with a specific port of the plurality of ports; causing a movable augmented indicator to display the real-time video stream, wherein the movable augmented indicator is configured to guide a user's connection of the specific cable to the specific port; monitoring in the real-time video stream changing locations of the specific port as the image sensor moves relative to the electronic appliance; and adjusting positions of the movable augmented indicator to account for the changing locations of the specific port in the real-time video stream.
Another aspect of the disclosure provides an image processing system employing artificial intelligence during technical support, the image processing system comprising: at least one processor configured to: receive image data captured by an image sensor of a mobile device, the image data including images of an inoperative appliance in an environment of a user; perform image recognition on the image data to identify the inoperative appliance and a likely cause of inoperability; retrieve a plurality of sequential instructions to be provided for enabling a user to complete a plurality of sequential actions in order to remedy the inoperability; cause the mobile device to sequentially display the plurality of sequential instructions; detect that the inoperative appliance is outside a field of view of the image sensor, based on the image data and during execution of the sequential actions; suspend display of additional sequential instructions when the inoperative appliance is outside of the field of view; detect when the inoperative appliance returns to the field of view after suspending display; and resume display of sequential instructions after the inoperative appliance is detected to return to the field of view.
Another aspect of the disclosure provides a non-transitory computer readable medium including instructions for employing artificial intelligence during technical support, the instructions being executable by at least one processor to cause the at least one processor to perform a method, the method comprising: receiving image data captured by an image sensor of a mobile device, the image data including images of an inoperative appliance in an environment of a user; performing image recognition on the image data to identify the inoperative appliance and a likely cause of inoperability; retrieving a plurality of sequential instructions to be provided for enabling a user to complete a plurality of sequential actions in order to remedy the inoperability; causing the mobile device to sequentially display the plurality of sequential instructions; detecting that the inoperative appliance is outside a field of view of the image sensor based on the image data and during execution of the sequential actions; suspending display of additional sequential instructions while the inoperative appliance is outside of the field of view; detecting when the inoperative appliance returns to the field of view after suspended display; and resuming display of the additional sequential instructions after detecting that the inoperative appliance has returned to the field of view.
Another aspect of the disclosure provides a method for employing artificial intelligence during technical support, the method comprising: receiving image data captured by an image sensor of a mobile device, the image data including images of an inoperative appliance in an environment of a user; performing image recognition on the image data to identify the inoperative appliance and a likely cause of inoperability; retrieving a plurality of sequential instructions to be provided for enabling a user to complete a plurality of sequential actions in order to remedy the inoperability; causing the mobile device to sequentially display the plurality of sequential instructions; detecting that the inoperative appliance is outside a field of view of the image sensor based on the image data and during execution of the sequential actions; suspending display of additional sequential instructions while the inoperative appliance is outside of the field of view; detecting when the inoperative appliance returns to the field of view after suspended display; and resuming display of the additional sequential instructions after detecting that the inoperative appliance has returned to the field of view.
Another aspect of the disclosure provides an image processing system employing artificial intelligence to electronically guide a technical support session, the image processing system comprising: at least one processor configured to: receive real-time image data captured by an image sensor of a mobile device at a location of an inoperative product, the real-time image data including at least one image of the inoperative product; perform image recognition on the real-time image data to identify a likely source of inoperability of the product; cause the mobile device to display a plurality of sequential instructions for mitigating inoperability of the product; determine that an error was made while performing the particular instruction, based on additional real-time image data captured following the display of a particular one of the plurality of sequential instructions; cause the mobile device to display an error notification when the particular instruction is not complied with, wherein the error notification being displayed before a subsequent instruction is displayed; determine that the particular instruction was subsequently complied with based on real-time image data captured following the notification; and cause the mobile device to display the subsequent instruction of the plurality of sequential instructions after the particular instruction is determined to have been complied with.
Another aspect of the disclosure provides a non-transitory computer readable medium including instructions for electronically guiding a technical support session, the instructions being executable by at least one processor to cause the at least one processor to perform a method, the method comprising: receiving real-time image data captured by an image sensor of a mobile device at a location of an inoperative product, the real-time image data including at least one image of the inoperative product; performing image recognition on the real-time image data to identify a likely source of inoperability of the product; causing the mobile device to display a plurality of sequential instructions for mitigating the inoperability; determining that an error was made in performing the particular instruction based on additional real-time image data acquired after the particular one of the plurality of sequential instructions is displayed; causing the mobile device to display an error notification when the particular instruction is not complied with, wherein the error notification being displayed before a subsequent instruction is displayed; determining that the particular instruction was subsequently complied with based on real-time image data captured following the notification; and causing the mobile device to display the subsequent instruction of the plurality of sequential instructions after determining that the particular instruction was complied with.
Yet another aspect of the disclosure provides a method for electronically guiding a technical support session, the method comprising: receiving real-time image data captured by an image sensor of a mobile device at a location of an inoperative product, the real-time image data including at least one image of the inoperative product; performing image recognition on the real-time image data to identify a likely source of inoperability of the product; causing the mobile device to display a plurality of sequential instructions for mitigating inoperability of the product; determining that an error was made in performing the particular instruction based on additional real-time image data captured following the display of a particular one of the plurality of sequential instructions; causing the mobile device to display an error notification when the particular instruction is not complied with, wherein the error notification being displayed before a subsequent instruction is displayed; determining that the particular instruction was subsequently complied with based on real-time image data captured following the notification; and causing the mobile device to display the subsequent instruction of the plurality of sequential instructions after determining that the particular instruction was subsequently complied with.
Yet another aspect of the disclosure provides a non-transitory computer readable medium for assisting a user during technical support, the computer readable medium including instructions that, when executed by at least one processor, cause the at least one processor to perform a method, the method comprising: receiving at least one first image of an inoperative product captured by a mobile device of a user; performing image analysis on the at least one first image to identify in the at least one first image a status of a functional element associated with the inoperative product; accessing memory to determine a reason why the product is inoperative based on the determined status of the functional element; causing visual guidance to be displayed by the mobile device, wherein the visual guidance is associated with a plurality of sequential actions for causing the inoperative product to become operative; receiving at least one second image of the product, the second image being indicative of a completion of the plurality of sequential actions; performing image analysis on the at least one second image to determine that the completion of the plurality of sequential actions caused the inoperative product to become operative; and providing a notification to the user indicating problem resolution.
Yet another aspect of the disclosure provides an image processing system employing artificial intelligence for assisting a user during technical support, the image processing system, comprising: at least one processor configured to: receive at least one first image of an inoperative product captured by a mobile device of a user; perform image analysis on the at least one first image to identify in the at least one first image a status of a functional element associated with the inoperative product; access memory to determine a reason why the product is inoperative based on the determined status of the functional element; cause visual guidance to be displayed by the mobile device, wherein the visual guidance is associated with a plurality of sequential actions for causing the inoperative product to become operative; receive at least one second image of the product, the second image being indicative of a completion of the plurality of sequential actions; perform image analysis on the at least one second image to determine that the completion of the plurality of sequential actions caused the inoperative product to become operative; and provide a notification to the user indicating problem resolution.
Yet another aspect of the disclosure provides a method for assisting a user during technical support, the method comprising: receiving at least one first image of an inoperative product captured by a mobile device of a user; performing image analysis on the at least one first image to identify in the at least one first image a status of a functional element associated with the inoperative product; accessing memory to determine a reason why the product is inoperative based on the determined status of the functional element; causing visual guidance to be displayed by the mobile device, wherein the visual guidance is associated with at least one action for causing the inoperative product to become operative; receiving at least one second image of the product, the second image being indicative of a completion of the at least one action; performing image analysis on the at least one second image to determine whether the at least one action was completed properly; and notifying the user when the at least one action is completed properly.
The foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the claims.
The accompanying drawings, which are incorporated in and constitute a part of this disclosure, illustrate various disclosed embodiments. In the drawings:
The following detailed description refers to the accompanying drawings. Wherever possible, the same reference numbers are used in the drawings and the following description to refer to the same or similar parts. While several illustrative embodiments are described herein, modifications, adaptations and other implementations are possible. For example, substitutions, additions, or modifications may be made to the components illustrated in the drawings, and the illustrative methods described herein may be modified by substituting, reordering, removing, or adding steps to the disclosed methods. Accordingly, the following detailed description is not limited to the disclosed embodiments and examples. Instead, the proper scope is defined by the appended claims.
The present disclosure provides remote assistance techniques for efficiently identifying technical problems, defects and improper equipment configurations, and determining a most likely solution to resolve them. In one non-limiting example, the techniques disclosed herein can be used in interactive applications to assist with the installation and troubleshooting of various items, including, furniture, consumer electronics, and appliances. Such technologies may utilize live, person-to-person video aided support and minimize consumer dissatisfaction and frustration.
In some embodiments, a processor of the system may receive image data captured by an image sensor of a mobile device. The term “mobile device” as used herein refers to any device configured to communicate with a wireless network, including, but not limited to a smartphone, tablet, mobile station, user equipment (UE), personal digital assistant, laptop, wearable sensor, e-Readers, dedicated terminals, smart glasses and any other movable device that enables communication with a remote server.
In some embodiments, the image data is captured by an image sensor. The term “image sensor” refers to a device capable of detecting and converting optical signals in the near-infrared, infrared, visible, and ultraviolet spectrums into electrical signals. The electrical signals may be used to form image data (e.g., an image or a video stream) based on the detected signal. Examples of image sensors may include semiconductor charge-coupled devices (CCD), active pixel sensors in complementary metal-oxide-semiconductor (CMOS), or N-type metal-oxide-semiconductors (NMOS, Live MOS). In some cases, the image sensor may be part of a camera included in the mobile device. The term “image data” refers to any form of data generated based on optical signals in the near-infrared, infrared, visible, and ultraviolet spectrums (or any other suitable radiation frequency range). Consistent with the present disclosure, the image data may include pixel data streams, digital images, digital video streams, data derived from captured images, and data that may be used to construct a 3D image. The image data acquired by the image sensor may be transmitted by wired or wireless transmission to a remote server.
In some embodiments, the received image data may be a “real-time” video stream. The term “real-time” pertains to on-going image data which closely approximates events as they are occurring. Such real-time video streams or “live” feeds allow, for example, a user to record an ongoing issue, installation, or repair at one location, and a remote customer service agent to visualize the ongoing issue, installation or repair.
In some embodiments the image data may be annotated. The term “annotation” pertain to symbols, text, images or other visual aids for guiding the user during a technical support session.
In some embodiments, the received image data may include images of an electronic appliance. The electronic appliance may include one or more functional elements. The term “functional element” means any component of an electrical appliance that aids in the function or operation of the appliance. Such elements may include, for example, jacks, ports, cables, buttons, triggers, indicator lights and switches.
In some embodiments, the electronic appliance may be inoperative. The term “inoperative appliance” pertains to electrical appliances that are non-operational because of non-assembly, defective assembly, defective installation or other defect. Such an appliance may be inoperative because it has not been plugged into a working electric source, because it has not been switched on or because a cable is in the wrong port (e.g., input instead of output).
In some embodiments, a processor of the system may cause the mobile device to display visual guidance for mitigating inoperability of a product (e.g. electrical appliance, furniture etc.). The term “visual guidance” pertains to any visual aid for directing a user to a particular component or group of components of the product (e.g., legs of a table or port of a cable box).
There is an ongoing demand for efficient customer service centers capable of quickly diagnosing and efficiently resolving problems encountered by their remote users. Traditional voice call support paradigms are rarely capable of establishing efficient and cost effective customer support centers. These conventional telephone voice call based support centers identify a limited number of user issues through tedious interrogation of their remote end users. In many instances an information gap exists between the center and the user. The end user may be incapable of correctly defining their problems/difficulties and thus unable to provide the support center with meaningful information for solving it. As a result, the center must dispatch a skilled technician to resolve the problem at the user's remote site (e.g., home, office, etc.).
The present application provides systems and methods for closing the information gap by enabling support centers to quickly identify product defects and other issues at the remote user's site, and for rendering fast and accurate working solutions. The amount of information exchanged between the remote end user and the support center is facilitated through use of a mobile device that allows for the exchange of image data and sound.
Upon establishing a video support session, the support center processes and analyzes the sounds and image data received from the remote end user. The support center provides tools for conveying instructions to the remote user. For example, where image data is comprised of one or more still images, the support center may add annotations, e.g., text, signs and/or symbols to the image data. Where image data is comprised of real-time video stream or video frames, the support center may superimpose a movable augmented indicator onto the image data. The annotated or superimposed image data is presented on the display of the user's mobile device. When the remote user successfully resolves the problem by following the annotated/superimposed instructions, the problem and solution is stored in a cloud or other database record system. By storing various problems and solutions, a database of working solutions is gradually established. The database may be used by the support center to more quickly and efficiently solve future problems. Alternatively, the database may form an artificial intelligence, whereby the artificial intelligence and not an agent of the support center solves the problem using the image data and relays the instructions to the user using annotations or a moveable augmented indicator as described above.
The disclosed methods and systems utilize computer vision tools having tracking capabilities for identifying complex objects/elements within a mobile device field of view, and for allowing tracking of such complex objects in sophisticated/challenging visioning conditions (e.g., poor lighting, poor resolution and noisy conditions). In some embodiments these capabilities are implemented by use of neural network tools. This way a multitude (e.g., thousands) of video streams can be analyzed by systems implementing the techniques disclosed herein to assist in the technical support sessions thereby conducted.
The present application thus provides systems and methods for conducting remote technical support sessions and real-time instructions. The methods and systems disclosed herein allow a technical support agent to see, in real-time, image data captured by the user's mobile device, and to provide real-time solutions.
Various tools continuously maintain the database records and discard database records that are not relevant or valid, inefficient, and/or rendered obsolete, and to facilitate real time matching of best working solutions from the database records to ongoing support sessions conducted by the support center. The techniques disclosed herein can be thus used to develop artificial intelligence (e.g., visual cellular chatbots) configured to provide automated customer support services.
Optionally, the video support session between the support center and the user's device may be activated by the user after receiving an activation link embedded in a text message (e.g., SMS, WhatsApp, email etc.) sent from the support center. By clicking on/accessing the embedded link the user opens the support session described herein and establishes the video support session communication with the support center. Optionally, the support session is achieved by means of an application installed on the user's device and configured to establish the video support session communication with the support center.
Support sessions techniques and systems are disclosed, wherein a bidirectional video communication is established between a technical support person and a remote user to provide additional layers of information exchange therebetween, thereby expanding the abilities of the system to detect issues/defects in appliances, objects and equipment, and of illustratively conveying solutions to the remote end-user. The bidirectional video communication can be achieved without requiring installation of a dedicated video support session application in the user's device. In some embodiments, after a telephone call is received from the remote user, the expert/agent verifies that the end-user owns a smartphone, and then sends a link (e.g., embedded in a text message) to the user's device. The bidirectional video stream is established once the customer clicks on the link received from the support center. The bidirectional video communication enables the expert/agent to see the environment at the remote site of the user, as captured by the back camera of the user's device, and thereby allows providing the remote user with substantially accurate instructions for resolving the encountered problem.
Once the code is entered/instructions followed, the support session begins. Agent 36p may request the user's permission to access and activate an image sensor, e.g. camera, on the mobile device 31 (step A5). Once user 33 approves the request (step A6), the camera is activated (step A7). User 33 may describe the problem and/or reasons for which the support is needed after which agent 36p may direct user 33 to direct the camera on the impaired equipment, appliance or other object of interest 33e. The session enables agent 36p to simultaneously view image data 33i received by mobile device 31 and speak with the user (audio-visual support) (step A8). Image data 33i includes the impaired equipment, appliance, or other object of interest and is displayed on a display device 36d at TSC 36 for agent 36p to see. Image data 33i may include real time video, still images, etc. and may or may not include audio support depending on agent/user preference or network bandwidth. Support session 20 may be conducted iteratively in real time until the user's problem is resolved. If the agent is unable to resolve the user's problem, a skilled technician might be sent to resolve the user's problem.
The techniques described herein enable agent 36p to superimpose annotations and/or moveable augmented indicators (markers) on the image data in real time. The markers and/or annotations may be created manually or selected from a pre-prepared library. These markers/annotations are superimposed on the object of interest and/or a functional element thereof within the image data during the video session. This may be achieved using a video tracker 36t so that the markers/annotations remain anchored to the object even if image sensor 31c moves relative to object 33e. Video tracker 36t may be implemented in a control unit 12, or configured to operate in a remote server/the cloud. Tracker 36t is configured and operable to track one or more objects 33e within image data 33i received from the remote user 33 and re-acquire the tracked object if it disappears from the image sensor's field of view and reappears due to movements of the image sensor 31c.
Video tracker 36t is configured to track the relevant objects/elements within the image data and to anchor annotations/markers when they move within the image data. Anchoring the annotations/marker to the object or particular element of the object allows the user to better track the object/element of interest, thereby facilitating assembly or repair thereof. In some embodiments, the object is annotated/marked using a computer vision tool. When the image data is received, the algorithm may identify its functional elements and brand. Once the identification is verified by the agent or through speech analysis, the data is marked accordingly.
Alternatively, snapshots of image data 33i may be taken and markers or annotations may be superimposed on the snapshots, then sent to the user. Image data 33i and data related to the support session, e.g., repair time, assembly time, solutions, annotations, markers, etc. may be saved to one or more databases 36r. Optionally, image data may be filtered so that background images, e.g., the room in which the object is located, are removed.
As show in
As illustrated in
Additionally, or alternatively, the agent 36p may select pre-defined annotations and/or markings and place them in the acquired image data 33i. In some embodiments an embedded video tracker is used to superimpose the annotations and/or markers 39 onto a desired object 33e within the image data. The tracker may move and/or resize the annotations and/or markers 39 whenever the image sensor 31c moves relative to object 33e. In this way, the position and/or size and of the annotations/markers may change in response to relative movement between image sensor 31c and object 33e. In some embodiments, AR tools may be used to create an associative connection between two or more objects in image data 33i. For example, an arrow may be superimposed between a particular cable and a particular functional element, e.g., port, of the object, to inform the user that the designated cable needs to be inserted into the designated port.
Using these techniques, a bidirectional video communication is established whereby the agent 36p is provided with image data 33i depicting the setup/configuration of the impaired appliance, equipment or object 33e, and the user 33 is provided with instructive annotated and/or augmented image data 33i′.
In some embodiments, TSC 36 is configured to record the video support sessions 21 in a database 36r. This way, the TSC 36 builds a continuously growing audio/visual database of user problems, and of their corresponding working solutions. The database may be used by computer vision tools to facilitate resolving of user's problems in future technical support sessions. Optionally, database 36r may be stored in a network computer/server of the TSC 36 (e.g., in the cloud).
After TSC 36 receives enough information from user 33, an audio/visual support session is established (step S4). Image data 33i received from user mobile device 31 is processed by the TSC 36 using a deep learning algorithm to detect the problematic appliance, equipment or object 33e in the image data (step S5), and to identify possible issues/defects causing the problem encountered by the remote user 33 (step S6).
Optionally, TSC 36 may utilize an embedded on-line computer vision tool configured and operable to automatically identify the relevant objects in the image data 33i, and/or identify codes, text and other symbols onto object 33e to thereby enable TSC 36 to identify the type, make, serial number etc. of the faulty object 33e. In some embodiments, agent 36p may guide the computer vision tools as to what objects to look for in image data 33i. For example, agent 36p may guide a cursor of a pointing device (e.g. a mouse) on/near the objects/elements thereof.
Additionally, or alternatively, speech analysis tools may be used to analyze the user's speech to identify keywords within the speech and aid the computer vision tool as it processes image data 33i for relevant objects/elements within the image sensor field of view. For example, if the speech recognition tool identifies words such as internet/network and communication/connectivity in the user' speech, it may guide the computer vision tool to look for LAN sockets or cables, Wi-Fi antennas and/or LEDs indications. Optionally, the keywords may be typed by agent 36p. Upon identifying the relevant objects in the image data 33i, the TSC system using the computer vision tool can analyze the object's setup/configuration and automatically identify potential issues/defects therein. Display device 36d may be used to present to the identified object to the agent.
Once the issue is identified, agent 36p may instruct the user on how to solve it. If the solution is relatively simple, (e.g., press the power switch), agent 36p may provide verbal instructions. If user 33 is unable to carry out the verbal instructions, or the instructions are relatively complex, agent 36p may generate an instructive augmented reality video stream using one or more markers 39 and trackers (step S8). The markers are superimposed onto the image data and displayed on the user's mobile device in real time to provide additional guidance. The agent may alternatively superimpose annotations as discussed and described above with respect to
Optionally, TSC 36 may query database 36r for a best working solution (step S7), based on the object's determined issues/defects, and transmits the best working solution to user 33. The instructions may comprise textual, auditory and/or annotated/augmented content. Agent 36p may provide user 33 with some (including one) or all types of instructive content, and/or limit the content to include some or the entire set of instructions.
After presenting proposed solution(s) to user 33 via mobile device 31 (step S9), the user performs the instructions received from the TSC 36. During this stage, image data 33i is continuously streamed from user's device 31, thereby allowing agent 36p to supervise and verify whether user 33 is following the instructions correctly, and provide corrective guidance if user 33 performs an action incorrectly. If the instructions provided do not resolve the problem, agent 36p may attempt to detect alternative issues/defects. If user 33 is able to resolve the problem using the provided solution, data relating to the support session 20 is recorded in the database 36r at the TSC 36 (step S11). The new record 51 may include data related to the resolved problem, e.g. keywords used by the system to identify the issues/defects, the objects/elements in which the issues/defects were found, and image data 33i received by the TSC 36 and/or annotated/augmented image data 33i′ conveyed to the remote user for resolving the problem.
The TSC is configured to use image data 33i to learn the nature of the issue/defect encountered, the best working solutions based on previous related sessions, and construct database records cataloguing the successful solutions. Database 36r is configured to continuously produce increasingly efficient solutions and improve customer satisfaction.
Periodic/intermittent maintenance procedures may be used to guarantee the effectiveness and validity of the records stored in the database. In some embodiments each database record is monitored during the maintenance procedures to determine its successful problem resolving percentage (rank) in real-time technical support sessions 20. For example, each solution may be ranked/scored according to the total number of times it was successfully used to resolve a specific problem or the total number of times it failed to resolve the problem. In some embodiments, maintaining the database further comprises discarding low-ranking solutions and only maintaining high-ranking solutions. Such database maintenance procedures increase the chances of successfully resolving user problems in future technical support sessions by using the good working examples used in the past to resolve the same problems.
Big data mining algorithms (dedicated for images and video i.e., video analytics tools) can be used to continuously monitor the database to sort and classify the problems and solutions that form the base line for the deep learning algorithms. Initially, mining may be done manually, whereby the agent scans the most relevant support sessions and classifies them according to problems they dealt with. Next, using the mining algorithm, the TSC may scan the support sessions and classify them automatically based on the keywords, objects/elements identified in the session. In some embodiments snapshots of the image data are automatically taken and the objects/elements therein are classified and added to the database for the computer vision tool discussed above. Optionally, image data 33i may be filtered so that background images, e.g., the room in which the object is located, are removed.
Deep learning algorithms can be used to analyze classified image data, and deliver the best working solution based on the “lessons” learned from all the past support sessions related to a certain class of problem.
Image recognition module 12i may scan the image data for certain objects/functional elements using keywords identified by the speech analysis module, and/or keywords 36i selected by the agent. In some embodiments, an optical character recognition (OCR) module 12c is used to identify letters/symbols within image data 33i, which can be used to guide the speech analysis module 12s and/or the image recognition module 12i.
Control unit 12 is configured and operable to use image recognition module 12i to identify an object/functional element's set up/configuration and detect potential problems or defects therein. Database 36r can be used to store a plurality of erroneous setups/configurations (also referred to herein as reference data) to be compared by a comparison module 12u of control unit 12. Whenever the comparison module 12u determines that newly acquired image data 33i contains the same objects, issues or defects as the reference data, control unit 12 generates a diagnosis 12d identifying the erroneous setup/configuration identified in the image data 33i.
Processing utility 12p may further include an image processing module 12g. Image processing module 12g is configured to superimpose annotations or markers onto the image data 33i based on keywords 36i, and generate annotated/augmented image data 33i′ for conveying to the remote user. Control unit video tracking module (tracker) 36t ensures that the annotations/markers superimposed onto the image data remain anchored to a desired object or functional element while the image sensor and/or object move.
Whenever a support session conducted by TSC 36 successfully resolves a problem encountered by a remote user 33, control unit 12 generates a new database record 51 including data related to the user's problem and the TSC's solution. The new database record is stored in the database 36r for use in future support sessions conducted by the TSC. In some embodiments, one or more databases 36r are used to store the reference data and the records 51.
Once object(s) 33e are identified as relevant, the image data undergoes deep image recognition 42 to identify the object's setup/configuration 43. A comparison step 45 is then used to compare the identified setup/configuration 43 to setup/configuration records 44 stored in database 36r of TSC 36. Based on the comparison results, the process proceeds to step 46, wherein a determination is made as to whether the assessed and recorded setups/configurations match.
If the setups/configurations do not match, the auditory and/or image data are reprocessed, to identify new objects/elements 33e. Any new auditory and/or image data obtained during the session 20 may also be processed, at which point all auditory/image data are processed to identify potential setup/configuration problems/defects. If a match is detected at step 46, possible problems/defects are determined accordingly in step 47. Comparing new problems against a database of potentially related problems allows for precise problem identification. Once problems/defects are determined, database 36r is queried in step 48 for the best past solution as discussed above with respect to
Although generally configured to match an identified setup/configuration 43 with known, defective/problematic configurations in the database, the comparison conducted in step 45 may also be configured to match the identified setup/configuration 43 with non-problematic/non-defective setup/configurations. If the identified configuration 43 is determined to be correct, the image data and/or auditory signals undergoes further processing using speech analysis module 12s and/or image recognition module 12i. If there is no match between the setups/configurations, the items/elements causing the mismatch are analyzed to determine potential problems/defects in step 47.
If it is determined in step 50 that the best past solution obtained in steps 48-49 resolved the user's problem, a new database record 51 is constructed in step 54, and then stored in the database of the system for use in future trouble shooting sessions. The new database record may include one or more annotated images, a video showing how to fix the problem (with or without AR markers), text and/or audible instructions. If the best past solution is unsuccessful, other high-ranking solutions are obtained from the database, and presented in attempt to resolve the problem. Steps 48 to 50 may be repeated for each solution until a successful solution is found. Alternatively, or concurrently, speech analysis 12s, image recognition 12i, and steps 41 to 46 can be carried out to determine alternative problems/defects in the object 33e. If the problem is not resolved using a predefined number of database solutions, agent 36p may provide supplemental instructions or send a professional expert to the user 33 in step 52.
Through repetition, the problem/defect detection process 40, image data analytics, and algorithms disclosed herein establish a self-service mechanism in which the computer vision tools are used to analyze objects/elements in image data 33i and identify problematic/defective setups/configurations. This automation allows TSC 36 to diagnose objects 33e in real-time and produce expedient solutions. In one non-limiting example, techniques and processing system disclosed herein are capable of identifying disconnected cable(s), cable(s) that are erroneously connected to the wrong port/sockets, errors indicated by certain LEDs, and/or error messages.
In some embodiments, once the problem/defect detection process 40 identifies an object's problems/defects, tracker 36t automatically tracks the relevant object as discussed above with respect to
The database generation and sorting process disclosed herein may be used in each support session 20 to improve the system's performance and facilitate customer service. Optionally, a machine learning process (e.g., employing any suitable state of the art machine learning algorithm having machine vision and deep learning capabilities) may be used to troubleshoot the technical support sessions. In one non-limiting example, the machine learning process logs and analyzes the users' interactions with the system during the support sessions to identify common users' errors. In this way, a dynamic database is constructed and problem solving sessions are optimized.
In some embodiments, the machine deep learning tool 52 is used to perform high-resolution, in depth, image recognition processes to identify the setup/configuration of object(s) 33e, as they appear in the image data. Identifying the setups/configurations allows the machine deep learning tool to accurately classify each support session in correct problem group. As will be described in detail below, the machine deep learning tool 52 may be used to find best matching solutions 55. In this way, machine deep learning tool 52 can be used to provide the system 50 a layer of automation, thereby allowing it to solve the user's problem without agent intervention.
Optionally, machine deep learning tool 52 may be configured and operable to carry out computer vision and video analysis algorithms for analyzing the image data and autonomously detecting problematic/defective objects therein. Machine deep learning tool 52, and/or processing utility 12p may then use the detected problems/defects to determine which past solutions from the database 36r may be used by agent 36p to resolve an ongoing user problem.
In some embodiments, machine deep learning tool 52 is further configured and operable to process and analyze data records 51 relating to previously conducted support, classify the database records according to the type of problem dealt with in each database record 51, identify keywords and/or objects/elements mentioned/acquired during the support session, and rank/weigh each database record 51 according to the number of times it was successfully used to resolve a problem of a particular type/classification.
At node 57a, support system 50 may identify the inoperative product (e.g., a router) using, for example, the router's model number. In step 58a the instruction may be to connect different cables of the router in a certain order. At node 57b, support system 50 may use additional image data to determine whether indicator lights of the router are red or green. If the indicator light is green, support system 50 may instruct the user to preform action steps 58b and 58c. If the indicator light is red, or otherwise not green, the support system 50 may instruct the user to perform other actions. Each line (i.e. action) may be associated with a weight representing the probability that the action will resolve the product's inoperability. Support system 50 may select a next sequential action based on the associated weight as will be discussed in further detail below. A person skilled in the art will recognize that the above example is simplified and that the actual process of remedy the inoperability of a product may involve substantially more steps.
It is to be appreciated that the systems and methods disclosed herein facilitate problem identification and resolution by substantially automating the process. Such automation reduces the agent training time and expense.
In some embodiments, machine deep leaning tool 52 is configured to scan database records 51 during each support session to match the problem presented in an ongoing session with a best matching solution 55. Machine deep leaning tool 52 identifies a set of database records 51 whose classification field 51b matches the classification field 51b of an ongoing support session 20. The machine deep leaning tool 52 then compares the keywords/objects field 51d of each database records 51 in classification field 51b to the keywords/objects identified/acquired in the ongoing support session 20, and selects a sub-set of best matching solutions 55. Thereafter, machine deep leaning tool 52 compares ranking fields 51c of the sub-set of best matching solutions 55 and selects at least one high-ranking solution 55 therefrom. The solution is then used by system 50 to resolve the problem presented in the ongoing support session 20.
In some embodiments, system 50 comprises a maintenance tool 56 configured and operable to operate in the background and continuously, periodically or intermittently, check the validity of each one of records 51 of database 36r. Maintenance tool 56 can determine whether certain types of database records 51 are no longer relevant (e.g., whether the record relates to obsolete/outdated technologies) and thus can be discarded. Maintenance tool 56 may also discard database records that are infrequently used, or which had little (or no) success in resolving user's problems.
In some embodiments maintenance tool 56 comprises a classification module configured and operable to classify the database records, and/or verify the classification determined for each record by machine deep learning tool 52. As illustrated in
As illustrated in
As explained hereinabove, the techniques disclosed herein allow for the gradual implementation of full self-service, wherein system 50 autonomously analyzes auditory/image data from the user's mobile device 31 and automatically determines the potential issues/defects causing the user's problem(s). System 50 is capable of automatically and remotely guiding the user to fix the problem, while monitoring the user's actions in real time.
System 50 can be thus configured to concurrently conduct a plurality of support sessions 20, without any human intervention, using combined speech and image/video recognition techniques, to extract the proper and relevant keywords from auditory signals and/or image data obtained from user 33 that describe the experienced problem, and to determine the setup/configuration of the user's object.
The combinations of the speech and image/video recognition techniques disclosed herein enable system 50 to assess the nature of the user's problem, and to generate a set of best working solutions from the system's database of relevant past working solutions. By using deep machine learning tools, the system gains enhanced problem solving capabilities, thereby guaranteeing that common user problems will get the best solutions.
In some embodiments, database records may be stored on user mobile device 31. This is especially useful when mobile device connectivity is poor or non-existent, or connectivity is unnecessary. These database records 51 are relevant to one or more of the user's problematic/defective objects/equipment. In such embodiments user mobile device 31 can be configured to automatically identify the object/equipment that needs to be serviced, using any of the techniques described herein, or alternatively let the user select the item/equipment that needs to be serviced from a list. Based on the user's selection, and/or automatic identification, user 33 will be provided with the best working solutions as provided in the maintained database records e.g., by playing/showing the recorded augmented reality based instructions. This way, different and specific self-service support modes can be implemented in a user's device, according to the specific items/equipment of the user.
If the user's device has temporary connectivity over a communication network (e.g., to the cloud), the best working solution can be downloaded to the user's device using the same user selection, and/or automatic identification procedures, either manually or by pattern recognition techniques.
One inventive aspect of the disclosure provides an image processing system for visually augmenting a real-time video stream. The image processing system comprises at least one processor configured to receive the real-time video stream captured by an image sensor, the real-time video stream including images of at least one cable and an electronic appliance. The at least one processor may analyze the real-time video stream to identify a plurality of ports in the electronic appliance. The at least one processor may also analyze the real-time video stream to identify a specific cable for association with a specific port of the plurality of ports; causing a display of a movable augmented indicator together with the live video stream, wherein the movable augmented indicator is configured to guide a user's connection of the specific cable to the specific port. The at least one processor may further monitor in the real-time video stream changing locations of the specific port as the image sensor moves relative to the electronic appliance. In some cases, the at least one processor may adjust positions of the movable augmented indicator to account for the changing locations of the specific port in the real-time video stream.
In accordance with some embodiments, the system is configured to display a movable augmented indicator together with the real-time (live) video stream to account for varying distances between two objects.
Mobile device 61 may transmit, and image processing system 60 may receive, image data 33i in the form of a real-time video stream. Once received, system 60 may analyze the video stream to identify the electrical appliance and functional elements thereof, using the database 36r and data processing methods discussed and described above with respect to
In one embodiment, image processing system 60 may utilize the tracker 36t, discussed and described above, to adjust a position of indicator 64 when the image sensor, appliance or functional elements move. For example, arrow 64a may become shorter or shift position, or circle 64c may become smaller as the user inserts the cable into the port. Image recognition and processing modules 12i, 12g of the processor may recognize functional elements when they are being manipulated. In this way, the processing system is able to identify and track the appliance and functional elements while they are being held by the user.
Once within the field of view, the real time video stream 33i may be analyzed using the data processing techniques described above to identify one or more functional elements, e.g., a plurality of ports in the electronic appliance and a specific cable for association with a specific port of the plurality of ports (step 72). The received video stream may be captured by a wearable image sensor or mobile device. In some embodiments, the at least one processor may be configured to analyze the video stream to identify a user's hand and one or more fingers, and/or an additional cable associated with an additional port of the plurality of ports when the specific cable is held by the user's hand or one or more fingers.
After identifying the relevant elements, the system 60 may cause a movable augmented indicator to display on the live video stream, the movable augmented indicator may be configured to guide the user's connection of the specific cable to the specific port (step 74). In some embodiments, the augmented indicator may include a boundary at least partially circumscribing the specific port. The boundary may form at least a portion of a circle, square or other polygon. Alternatively, the augmented indicator may be an arrow, bracket, finger or line. In some embodiments, the augmented indicator may change in appearance (e.g. transition from blinking to steady or arrow to check mark, or disappear) or color to alert the user that a step has been successfully completed. In other examples, the augmented indicator may change when the specific cable is successfully connected to the specific port
The real-time video stream may be monitored throughout the course of the repair/assembly (step 76). In this way, the system may move, resize or otherwise adjust the augmented movable indicators to reflect the user's progress (step 78). In one embodiment, the at least one processor may be further configured to adjust the positions of the movable augmented indicator by changing a pointing direction of the arrow to coincide with movement of the specific port in the live video stream. For example, system 60 may adjust the movable augmented indicator to account for the changing locations of the specific port in the real-time video stream. Consistent with one embodiment, after the specific cable is connected to the specific port, the video stream may be analyzed to confirm that the specific cable was properly connected to the specific port. The video stream may be analyzed, for example, by assessing a status light on the electronic appliance.
In some embodiments, the method further comprises retrieving from information about the electronic appliance from a database. Where, for example, the plurality of ports in the electronic appliance share a similar visual appearance, a specific port of the plurality of ports may be identified based on the retrieved information. Specifically, the database records may be stored on a remote server or on the user's mobile device. This may be useful when mobile device connectivity is poor or non-existent, or connectivity is unnecessary. If the user's device has temporary connectivity over a communication network (e.g., to the cloud), the best working solution can be downloaded to the user's device using the same user selection, and/or automatic identification procedures, either manually or by pattern recognition techniques. Methods for visually augmenting the real-time video stream may be stored on various non-transitory computer readable media (CRM). The instructions contained on the CRM may be executed by a processor to cause the processor to perform the method discussed above.
In accordance with some embodiments, the method may further include causing auditory feedback to be provided in conjunction with the moveable augmented indicator to guide the user's connection of the specific cable to the specific port. Optionally, the feedback may be adjusted to account for the changing locations of the specific port in the real-time video stream. The auditory feedback may include discrete “beeps” or “clicks”, wherein a frequency of auditory feedback correlates to the user's proximity to the specific port. Additionally, or alternatively, the auditory feedback may include a chime or other distinct sound when the specific cable is properly placed in the specific port.
Although reference is made to an electrical appliance and cable, it is to be understood that the system and methods disclosed herein may be used with other objects. Where two objects are concerned, for example, the two objects include may two distinct components of an electrical appliance (e.g. cable and port) or a user's hand and a component of the electrical appliance (e.g. button, switch, cable, port, trigger, indicator light etc.) The augmented indicator may direct the user to a particular component of an electrical appliance (e.g. button, switch, cable, port, trigger etc.) to guide manipulation or use thereof. The two objects may also include two distinct components of furniture (e.g. rail and drawer), a user's hand and a component of the furniture (e.g. leg, shelf, rail, drawer etc.), or a user's hand and a tool (e.g. screw driver, wrench, bolt etc.). The augmented indicator may direct the user to a particular component of the furniture (e.g. leg, shelf, rail, drawer etc.) to guide manipulation or use thereof.
In accordance with some embodiments, the video stream is captured by a wearable image sensor or mobile device, e.g. smart glasses, mobile phone, tablet, laptop, etc.
In accordance with some embodiments, information about the electronic device (e.g. port and cable orientation and configuration) may be stored in a cloud or other database.
In accordance with some embodiments, the moveable augmented indicator includes an arrow, finger, line or boundary (e.g. a portion of a circle, square or other polygon). The indicator may change in appearance (e.g. transition from blinking to steady or arrow to check mark) or color to alert the user that a step has been successfully completed.
In accordance with some embodiments, auditory feedback is provided in conjunction with the moveable augmented indicator to guide the user to a desired component of the electrical appliance,
In accordance with some embodiments the auditory feedback may include discrete “beeps” or “clicks”, wherein a frequency of auditory feedback correlates to the user's proximity to the desired component. Additionally, or alternatively, the auditory feedback may include a chime or other distinct sound when a component is properly placed and/or when an entire task is completed. Alternatively, the auditory feedback may include words.
In some embodiments, processing system 60 is configured to suspend further instruction when the electrical appliance or functional element are not within the image sensor's field of view. The processor may be configured to notify the user, via the mobile device, when instructions have been suspended and/or when they have resumed. The processer may further be configured to notify the user, via the mobile device, of one or more pending instructions once instructions have been suspended.
As seen in
During repair/assembly, the user may purposefully or accidently move the appliance/functional element out of the field of view (see
When the processor detects the move, it may suspend further instruction as illustrated in
Once the object/functional element returns to the field of view, the processor resumes real-time instruction. Additional instructions may display as the user progresses. As illustrated in
Mobile device 81 transmits and image processing system 60 receives image data including images of inoperative appliance 83e and various functional elements related thereto (step 90). The processor may identify the appliance using the appliance model number or brand (illustrated as “XXX” in
During repair/assembly of the inoperative appliance, the processor continuously scans the image data to determine whether the appliance/related functional elements are within the image sensor's field of view (FOV) (step 98). The processor may determine that the appliance/functional element is no longer within the field of view by comparing a previous image of the appliance/functional element to a live image of the appliance/functional element.
If the processor determines that the appliance/functional element has left the field of view, the processor may suspend display of any pending instructions (including moveable augmented indicators) and/or instruct the user to refocus the image sensor so that the appliance/functional elements return to the field of view (step 99). The processor may notify the user via a visual or audible message. Such notifications may occur automatically or after a preset time.
For example, the processor may allot three minutes to a particular step and only notify the user that the appliance/functional element is out of the field of view once three minutes have passed. Such a configuration allows the user to set the mobile device aside and proceed with repair/assembly without receiving distracting notifications. In some embodiments, the processor will only notify the user if a functional element related to a particular step is outside the field of view. For example, if step three of a set of sequential instructions directs the user to insert cable “b” into port “z”, and cable “c” or ports v-y are outside the field of view, the user will not receive a notification. If, however, step four requires the user to insert cable “c” into port “y”, a notification will occur. The user may continue to get notifications until the relevant functional elements/appliance are within the field of view. Alternatively, the processor may continue to provide audible instructions even though the appliance/functional element is outside the field of view.
Once the relevant functional elements/appliance return to the field of view, the processor will resume display of the pending instructions (step 100). Methods for employing A.I. during a technical support session may be stored on various CRM. The instructions contained on the CRM may be executed by a processor to cause the processor to perform the method discussed above.
To ensure successful repair/assembly and customer satisfaction, image processor 60 may be further configured to notify the user when an assembly/repair error occurs. The processor may be configured to identify the product and the likely source of inoperability of the product using the product's model number, brand, or other labeling.
The augmented live feed illustrated in
In certain situations (e.g., complex assembly), the processor may ascertain whether the user has all of the necessary functional elements in his possession.
To begin, the processor identifies an appliance 111 using the product name or other labeling (illustrated as “XXX” in
Once assembly/repair is complete a data record relating to the support session is generated and saved in the database as discussed above with respect to
Mobile device 101 transmits and image processing system 60 receives real-time image data including images of an inoperative appliance 103e and various functional elements related thereto (step 120). The image processing system then performs image recognition on the image data to identify the appliance's likely source of inoperability (step 122). As discussed above with respect to
For each step of the repair/assembly of the inoperative appliance, the processor continuously scans the image data to determine whether and why the user committed an error (step 126). If the processor determines that an error has occurred, e.g., the user connected the wrong cable to the wrong port, or connected the cable up-side down, an error notification is displayed on the real-time image data (step 127). The error notification may include steps for correcting the error. For example, if the user attempts to insert a cable into a port upside-down, the error notification may instruct the user to correct the orientation of the cable.
In some embodiments, the method further includes assessing the user's technical skill based on the user's error rate and revising the instructions to accommodate the user's skill level when the error rate exceed a predetermined threshold. If the user makes numerous errors, for example, or the repair/assembly process is complex/requires numerous steps, the processor may instruct the user to scan a functional element prior to connecting it to the appliance to prevent errors from occurring. If the processor then determines that the error was corrected and that a step has been completed successfully, subsequent instructions will appear (step 128). This error checking cycle may occur at each step or after a predetermined number of steps. In some embodiments, the processor determines whether the particular instruction was subsequently complied with by determining an operational or positional state of a functional element of the inoperative product. For example, the processor may determine that instructions were complied with because a functional element, e.g., a switch, is in the “On” position, or a series of buttons on an electrical appliance have illuminated.
Methods for guiding the technical support session may be stored on various CRM. The instructions contained on the CRM may be executed by a processor to cause the processor to perform the method discussed above.
While live video streams are beneficial to enable real-time feedback during a support session, in certain situations, the system may rely on discrete images.
Once retrieved, the processor annotates the image with one or more instructions for correcting the source of inoperability.
In some embodiments, the processor is further configured to check the inoperative product's operational status using a medium other than image analysis before providing the notice. For example, the processor may check for the presence or absence of Wi-fi to determine whether a router was properly repaired or installed.
Before generating a new data record 51 in database 36r, the processor may prompt the user to test the functionality of the configuration. For example, the processor may cause the mobile device to visually or audibly instruct the user to “access the internet” to determine whether the instructions caused an inoperative router to become operative, or to turn on the TV to determine whether the instructions caused an inoperative TV to become operative. As illustrated in
If the instructions are unsuccessful the user may be presented with an option to try again, or to receive human technical support. In the latter case, the processor may be configured to initiate a support session with a human operator when the processor determines that the completion of the plurality of sequential actions failed, or request that a technician be dispatched.
To begin, a mobile device transmits and image processing system 60 receives at least one first image including an inoperative object and various functional elements relevant thereto (step 140). As discussed above, the functional elements may include one of a button, a switch, a trigger, a port, a cable, and an indicator light.
The image processing system then performs image recognition on the image to identify the appliance's likely source of inoperability (step 142). As discussed above with respect to
Once the source of inoperability is identified, the processor retrieves at least one solution from database 36r related to the assembly, repair, or correction of the inoperative object and causes the mobile device to display one or more annotations relating to the solution (step 146). In some embodiments, the annotations may be anchored to at least one functional element of the inoperative product, such that as a camera angle of the mobile device changes, the annotation remains anchored to the functional element.
After one or more steps relating to the solution are complete, a second image of the object and relevant functional elements are transmitted to the processor. Next, system 50 performs additional image analysis on the second image to determine whether the one or more steps were completed (step 148). The processor may also confirm that the object in the second image is a same object that was captured in the at least one first image.
Once it is confirmed that the solution was properly carried out, the processor may notify the user using positive feedback (step 150). The notification may be visually provided to the user's mobile device or audibly provided to the user via a speaker of the mobile device. Notifications may be provided after each step is completed and/or after all steps have been completed.
It is to be understood that the processor may provide negative feedback in the form of the error notification discussed above, and provide further instruction to facilitate assembly/repair. In some embodiments, a different set of sequential actions may be displayed when the processor determines, based on additional images, that the object remains inoperative. Consistent with other disclosed embodiments, methods for guiding the technical support session may be stored on various CRM. The instructions contained on the CRM may be executed by a processor to cause the processor to perform the method discussed above.
Although the system and method described herein relate to an electrical appliance, it is to be understood that the same principles apply to any of a variety of equipment or objects including, e.g., furniture. The system may, for example, identify a plurality of rails, a head board and a foot board, and instruct the user on proper assembly using one or more moveable augmented indicators.
It should also be understood that in the processes or methods described above, the steps of the processes/methods may be performed in any order or simultaneously, unless it is clear from the context that one step depends on another being performed first.
The technology disclosed herein can be implemented by software that can be integrated into existing CRM systems of technical support centers or organizations, and that can replace it or work in parallel thereto. Such software implementations combine opening a voiceless bi-directional video channel, when the customer's smartphone transmits the video image to the agent/expert, and the agent/expert gives the costumer audiovisual instructions over the communication channel.
Accordingly, the systems and methods disclosed herein enable users to proactively diagnose faulty item/equipment for increasing the productivity and efficiency, and to resolve issues faster based on a maintained pool of past working solutions. The user's mobile device, e.g., 31 is thereby harnessed to conduct support sessions and improve customer satisfaction, decrease technician dispatch rates for resolving user's problems, substantially improve the first call resolution rates, and decrease the average handling time.
As will be appreciated by one skilled in the art, aspects of the present disclosure may be embodied as a system, method or computer program product. Accordingly, aspects of the present disclosure may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present disclosure may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
As described hereinabove and shown in the associated figures, the present disclosure provides support session techniques, systems and methods, for expeditiously identifying product defects/other issues and corresponding working solutions for resolving problems encountered by remote users. While particular embodiments of the disclosure have been described, it will be understood, however, that the disclosure is not limited thereto, since modifications may be made by those skilled in the art, particularly in light of the foregoing teachings. As will be appreciated by the skilled person, the disclosure can be carried out in a great variety of ways, employing more than one technique from those described above, all without exceeding the scope of the claims.
The foregoing description has been presented for purposes of illustration. It is not exhaustive and is not limited to the precise forms or embodiments disclosed. Modifications and adaptations will be apparent to those skilled in the art from consideration of the specification and practice of the disclosed embodiments. Additionally, although aspects of the disclosed embodiments are described as being stored in a database, one skilled in the art will appreciate that these aspects can also be stored on other types of computer readable media, such as secondary storage devices, for example, hard disks or CD ROM, or other forms of RAM or ROM, USB media, DVD, Blu-ray, or other optical drive media.
Computer programs based on the written description and disclosed methods are within the skill of an experienced developer. The various programs or program modules can be created using any of the techniques known to one skilled in the art or can be designed in connection with existing software. For example, program sections or program modules can be designed in or by means of .Net Framework, .Net Compact Framework (and related languages, such as Visual Basic, C, etc.), Java, C++, Objective-C, HTML, HTML/AJAX combinations, XML, or HTML with included Java applets.
Moreover, while illustrative embodiments have been described herein, the scope of any and all embodiments having equivalent elements, modifications, omissions, combinations (e.g., of aspects across various embodiments), adaptations and/or alterations as would be appreciated by those skilled in the art based on the present disclosure. The limitations in the claims are to be interpreted broadly based on the language employed in the claims and not limited to examples described in the present specification or during the prosecution of the application. The examples are to be construed as non-exclusive. Furthermore, the steps of the disclosed methods may be modified in any manner, including by reordering steps and/or inserting or deleting steps. It is intended, therefore, that the specification and examples be considered as illustrative only, with a true scope and spirit being indicated by the following claims and their full scope of equivalents.
Patent | Priority | Assignee | Title |
10785269, | Nov 20 2017 | STREEM, INC. | Augmented reality platform for professional services delivery |
11228624, | Nov 13 2019 | Amazon Technologies, Inc.; Amazon Technologies, Inc | Overlay data during communications session |
Patent | Priority | Assignee | Title |
10182153, | Dec 01 2016 | TECHSEE AUGMENTED VISION LTD | Remote distance assistance system and method |
10397404, | Dec 01 2016 | TechSee Augmented Vision Ltd. | Methods and systems for providing interactive support sessions |
20020044104, | |||
20080030575, | |||
20130278635, | |||
20170308751, | |||
20180159979, | |||
20190268470, | |||
20190268472, | |||
20190268571, | |||
20190278992, | |||
EP943972, | |||
WO2007066166, | |||
WO2009036782, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
May 09 2019 | TechSee Augmented Vision Ltd. | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
May 09 2019 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
May 20 2019 | SMAL: Entity status set to Small. |
Aug 02 2023 | M2551: Payment of Maintenance Fee, 4th Yr, Small Entity. |
Date | Maintenance Schedule |
Feb 18 2023 | 4 years fee payment window open |
Aug 18 2023 | 6 months grace period start (w surcharge) |
Feb 18 2024 | patent expiry (for year 4) |
Feb 18 2026 | 2 years to revive unintentionally abandoned end. (for year 4) |
Feb 18 2027 | 8 years fee payment window open |
Aug 18 2027 | 6 months grace period start (w surcharge) |
Feb 18 2028 | patent expiry (for year 8) |
Feb 18 2030 | 2 years to revive unintentionally abandoned end. (for year 8) |
Feb 18 2031 | 12 years fee payment window open |
Aug 18 2031 | 6 months grace period start (w surcharge) |
Feb 18 2032 | patent expiry (for year 12) |
Feb 18 2034 | 2 years to revive unintentionally abandoned end. (for year 12) |