A remote device infrastructure can be used to test and develop applications and websites. A user developer can select a number of remote devices at a remote location and test a programming application from a local machine. The remote devices run the programming application. The user interacts with mirrored displays of the remote devices on the local machine of the user. User inputs are transmitted to a remote device. The user can also enable a multisession mode, where the user can test a programming application on multiple remote devices and observe a display output of each remote device on the local machine of the user. The user can interact with any mirrored display of a remote devices in a multisession and observe a synced output in the other mirrored displays.

Patent
   11860771
Priority
Sep 26 2022
Filed
Sep 26 2022
Issued
Jan 02 2024
Expiry
Sep 26 2042
Assg.orig
Entity
Large
0
55
currently ok
1. A method comprising:
establishing a connection session between a browser and a first device and the browser and a second device, wherein the first and second devices are in one or more locations remote relative to the browser, and the connection session between the browser and the first device comprises a first device data channel and a first device video channel, and the connection session between the browser and the second device comprises a second device data channel and a second device video channel;
generating a first device display on the browser by broadcasting, via the first device video channel, a first device video feed from the first device to the browser;
generating a second device display on the browser by broadcasting, via the second device video channel, a second device video feed from the second device to the browser;
transmitting, via the first device data channel, one or more user interactions with a webpage in the generated first device display on the browser, to the first device;
transmitting the interactions to a sync server;
generating, by the sync server, a request to a host of the webpage, based at least in part on the interactions;
transmitting, by the sync server, the request to the host of the webpage;
receiving, by the sync server, a response from the host of the webpage;
serving, from the sync server, the response from the host of the webpage, comprising a synced version of the webpage to the first and second devices.
13. A system comprising a processor, the processor configured to perform operations comprising:
establishing a connection session between a browser and a first device and the browser and a second device, wherein the first and second devices are in one or more locations remote relative to the browser, and the connection session between the browser and the first device comprises a first device data channel and a first device video channel, and the connection session between the browser and the second device comprises a second device data channel and a second device video channel;
generating a first device display on the browser by broadcasting, via the first device video channel, a first device video feed from the first device to the browser;
generating a second device display on the browser by broadcasting, via the second device video channel, a second device video feed from the second device to the browser;
transmitting, via the first device data channel, one or more user interactions with a webpage in the generated first device display on the browser, to the first device;
transmitting the interactions to a sync server;
generating, by the sync server, a request to a host of the webpage, based at least in part on the interactions;
transmitting, by the sync server, the request to the host of the webpage;
receiving, by the sync server, a response from the host of the webpage;
serving, from the sync server, the response from the host of the webpage, comprising a synced version of the webpage to the first and second devices.
7. A non-transitory computer storage medium that stores executable program instructions that, when executed by one or more computing devices, configure the one or more computing devices to perform operations comprising:
establishing a connection session between a browser and a first device and the browser and a second device, wherein the first and second devices are in one or more locations remote relative to the browser, and the connection session between the browser and the first device comprises a first device data channel and a first device video channel, and the connection session between the browser and the second device comprises a second device data channel and a second device video channel;
generating a first device display on the browser by broadcasting, via the first device video channel, a first device video feed from the first device to the browser; generating a second device display on the browser by broadcasting, via the second device video channel, a second device video feed from the second device to the browser;
transmitting, via the first device data channel, one or more user interactions with a webpage in the generated first device display on the browser, to the first device;
transmitting the interactions to a sync server;
generating, by the sync server, a request to a host of the webpage, based at least in part on the interactions; transmitting, by the sync server, the request to the host of the webpage;
receiving, by the sync server, a response from the host of the webpage;
serving, from the sync server, the response from the host of the webpage, comprising a synced version of the webpage to the first and second devices.
2. The method of claim 1, further comprising establishing the connection session between a plurality of devices, wherein the serving, from the sync server, comprises serving the synced webpage to the plurality of the devices simultaneously.
3. The method of claim 1, wherein the interactions comprise commands for scrolling the webpage and the host performs a corresponding scrolling on the response comprising the synced webpage served to the first and second devices.
4. The method of claim 1, wherein the sync server determines webpage HTML elements corresponding to the user interactions.
5. The method of claim 1, wherein the first and second devices access the sync server via a shared URL corresponding to the webpage.
6. The method of claim 1, wherein transmitting the interactions to the sync server further comprises:
determining coordinates of the interactions on the generated first device displayed on the browser;
identifying corresponding HTML elements of the webpage; and
transmitting the interactions and the coordinates to the sync server.
8. The non-transitory computer storage medium of claim 7, wherein the operations further comprise establishing the connection session between a plurality of-devices, wherein the serving, from the sync server, comprises serving the synced webpage to plurality of the devices simultaneously.
9. The non-transitory computer storage medium of claim 7, wherein the interactions comprise commands for scrolling the webpage and the host performs a corresponding scrolling on the response comprising the synced webpage served to the first and second devices.
10. The non-transitory computer storage medium of claim 7, wherein the sync server determines webpage HTML elements corresponding to the user interactions.
11. The non-transitory computer storage medium of claim 7, wherein the first and second devices access the sync server via a shared URL corresponding to the webpage.
12. The non-transitory computer storage medium of claim 7, wherein transmitting the interactions to the sync server further comprises:
determining coordinates of the interactions on the generated first device displayed on the browser;
identifying corresponding HTML elements of the webpage; and transmitting the interactions and the coordinates to the sync server.
14. The system of claim 13, wherein the operations further comprise establishing the connection session between a plurality of devices, wherein the serving, from the sync server, comprises serving the synced webpage to the plurality of the devices simultaneously.
15. The system of claim 13, wherein the interactions comprise commands for scrolling the webpage and the host performs a corresponding scrolling on the response comprising the synced webpage served to the first and second devices.
16. The system of claim 13, wherein the sync server determines webpage HTML elements corresponding to the user interactions.
17. The system of claim 13, wherein the first and second devices access the sync server via a shared URL corresponding to the webpage.

This invention relates generally to the field of software development using multiple platforms and more particularly to enabling a remote device infrastructure for testing and development of software on multiple remote hardware and software platforms.

The approaches described in this section are approaches that could be pursued, but not necessarily approaches that have been previously conceived or pursued. Therefore, unless otherwise indicated, it should not be assumed that any of the approaches described in this section qualify as prior art merely by virtue of their inclusion in this section.

The multitude of computers, mobile devices and platforms have given businesses and consumers a vast array of options when they choose a device. The plethora of choices include both hardware and software. Naturally, software, application and website developers have a keen interest in ensuring their products work seamlessly across the existing hardware and platforms, including older devices on the market. This creates a challenge for the developers to properly test their products on the potential devices and platforms that their target consumer might use. On the one hand, acquiring and configuring multiple potential target devices can strain the resources of a developer. On the other hand, the developer may not want to risk losing a potential market segment by disregarding a particular platform in his typical development cycle. Even for prominent platforms, such as iOS® and Android®, at any given time, there are multiple generations and iterations of these devices on the market, further complicating the development and testing process across multiple platforms. Even in a given platform, a variety of software, operating systems and browser applications are used by a potential target audience of a developer. This dynamic illustrates a need for a robust infrastructure that enables developers to test their products across multiple devices and platforms, without having to purchase or configure multiple devices and platforms.

The appended claims may serve as a summary of this application.

These drawings and the associated description herein are provided to illustrate specific embodiments of the invention and are not intended to be limiting.

FIG. 1 illustrates an example remote test system.

FIG. 2 illustrates a diagram of an example data flow implementation of the embodiment of FIG. 1.

FIG. 3 illustrates a flow chart of a method of enabling a remote session at a first location using a remote device at a second location.

FIG. 4 illustrates a flowchart of a method of an example operation of a remote test system.

FIG. 5 illustrates another flowchart of a method of an example operation of the remote system.

FIG. 6 illustrates an example environment within which some described embodiments can be implemented.

FIG. 7 illustrates an example data flow diagram of the operations of an infrastructure enabling a remote session using a remote device, using a video capturing API.

FIG. 8 illustrates an environment of multisession operations of the remote test system.

FIG. 9 illustrates a block diagram of the operations of multisession mode of the remote test system.

FIG. 10 illustrates a flowchart of a method of performing a multisession mode.

FIG. 11 illustrates a flowchart of a method of multisession operations including a request/response type of operation.

The following detailed description of certain embodiments presents various descriptions of specific embodiments of the invention. However, the invention can be embodied in a multitude of different ways as defined and covered by the claims. In this description, reference is made to the drawings where like reference numerals may indicate identical or functionally similar elements.

Unless defined otherwise, all terms used herein have the same meaning as are commonly understood by one of skill in the art to which this invention belongs. All patents, patent applications and publications referred to throughout the disclosure herein are incorporated by reference in their entirety. In the event that there is a plurality of definitions for a term herein, those in this section prevail. When the terms “one”, “a” or “an” are used in the disclosure, they mean “at least one” or “one or more”, unless otherwise indicated.

Software developers, particularly website, web application and mobile device application developers have a desire to manually test their products on a multitude of hardware and software platforms that their target audience may use. A variety of mobile device manufacturers provide the hardware consumers and businesses use. Examples include, devices manufactured by Apple Inc., Google LLC, Samsung Electronics Co. Ltd., Huawei Technologies Co. Ltd. and others. Similarly, a variety of operating systems for consumer electronic devices exist. Examples include Apple iOS®, Android® operating system (OS), and Windows® Mobile, Windows® Phone and others. Furthermore, users have a variety of choices as far as the web browser application they can use. Examples include Safari®, Chrome®, FireFox®, Windows Explorer®, and others. This variety of choice presents a difficult challenge for a web/app developer to test products on potential target devices. Traditionally, the developer might have to acquire a test device and spend resources configuring it (for example, by installing a target OS, browser, etc.) as well as a secondary hardware device to connect the test device through the secondary hardware device to a local machine of the developer, in order to write code and conduct tests on the test device. The sheer variety of possible devices, operating systems, browsers and combinations of them are numerous and can present a logistical hurdle to the developer.

A testing provider can enable a remote test system (RTS), having a multitude of devices for a developer to connect to and conduct tests. The developer can connect to the test system, select a test device, select a configuration (e.g., a particular browser, etc.) and run tests using the selected remote device. The RTS can include a server powering a website or a desktop application, which the developer can use to launch a dashboard for connecting to the RTS and for conducting tests. The dashboard can include a display of the remote device presented to the developer. The RTS system can capture developer inputs and input them to the remote device. The RTS mirrors the display of the remote device on the developer's local machine and simultaneously captures the developer's interactions inputted onto the mirrored display and transfers those commands to the remote device. In a typical case, the developer can use a keyboard and mouse to input interactions onto the mirrored display. When the test device is a smart phone device, the RTS system translates those input interactions compatible with the smart phone. Examples of smart phone input interactions include gestures, pinches, swipes, taps, and others. The remote device display is mirrored on the developer's local machine. In this manner, the developer can experience a seamless interaction with the remote device using the developer's local machine. The RTS can be used both for development of launched and unlaunched products.

FIG. 1 illustrates an example RTS 100. Although some embodiments use the RTS 100 in the context of testing and software development, the RTS 100 can be used to enable a remote session for any purpose. Testing is merely provided as an example context of usage area of the system and infrastructure of the RTS 100. A user 102 uses a local machine 104 to launch a browser 106 to access a dashboard application to interact with the RTS 100, connect to a remote device and to conduct tests on the remote device. In some embodiments, the dashboard website/web application may be replaced by a desktop application, which the user 102 can install on the local machine 104. The user 102 may be a software developer, such as a website developer, web application developer or a mobile application developer. The local machine 104, in a typical case, may be a laptop or desktop computer, which the user 102 can use to write software code, debug, or run tests on a website/web application or mobile application. The user 102 can enter a uniform resource locator (URL) 108 in browser 106 to connect to the dashboard application powered by a server 110. The server 110 can enable the browser 106 and a remote device 114 to establish a connection. The RTS 100 can use the connection for streaming the display of a remote device 114 onto the browser 106 in order to mirror the display of the remote device 114 and present it to the user 102. The RTS 100 can also capture user inputs entered into the mirrored display and input them to the remote device 114.

The RTS 100 can include multiple datacenters 112 in various geographical locations. The datacenters 112 can include a variety of test devices for the users 102 to connect with and to conduct tests. In this description, the test devices in datacenters 112 are referred to as remote devices 114, as they are remote, relative to the user 102 and the user's local machine 104. A variety of communication networks 116 can be used to enable connection between the browser 106, the server 110 and the remote device 114. The remote devices 114 can include various hardware platforms, provided by various manufacturers, different versions of each brand (for example, old, midmarket, new) and optionally various copies of each brand, to enable availability for numerous users 102 to connect and conduct tests.

The RTS 100 can use a host 118 connected to one or more remote devices 114. In some embodiments, the browser 106 does not directly communicate with the remote device 114. The host 118 enables communication between the browser 106 and the remote device 114 through one or more private and/or public communication networks. The host 118 can be a desktop, laptop, or other hardware connected with a wired or wireless connection to the remote device 114. The hardware used for the host 118 can depend on the type of the remote device 114 that it hosts. Examples of host 118 hardware can include Apple Macintosh® computers for iPhone® and iOS® devices and Zotac® for Android® devices.

The RTS 100 mirrors the display of the remote device 114 on the browser 106, by generating a display 120 on the browser 106. In some embodiments, the display 120 can be a graphical, or pictorial replica representation of the remote device 114. For example, if an iPhone® 12 device is chosen, the display 120 can be an image of an iPhone® 12. The RTS 100 mirrors the display of the remote device 114 on the display 120 by streaming a video feed of the display of the remote device 114 on the display 120. In some embodiments, the video stream used to mirror the display of the remote device 114 is generated by capturing and encoding screenshots of the display of the remote device 114 into a video stream feed of high frames per second to give the user 102 a seamless interaction experience with the display 120. Using input devices of the local machine 104, the user 102 can interact with the display 120, in the same manner as if the remote device 114 were locally present.

The RTS 100 captures and translates the user interactions to input commands compatible with the remote device 114 and inputs the translated input commands to the remote device 114. The display responses of the remote device 114 are then streamed to the user 102, via display 120. In some embodiments, the user 102 has access to and can activate other displays and menu options, such as developer tools display 122. An example usage of the RTS 100, from the perspective of the user 102, includes, the user 102, opening a browser on the remote device 114, via menu options provided by the dashboard application. The user 102 can access the dashboard application via the browser 106 on the user's local machine 104. The RTS 100 opens the user's selected browser on the remote device 114 and generates a display of the remote device 114 and the remotely opened browser on the browser 106 on the user's local machine 104. The user 102 can then use a mouse to click on a URL field 124 in the display 120, which corresponds to the URL field in the browser on the remote device 114. The user 102 can subsequently enter a URL address in the URL field 124. Simultaneously, the user's interactions, such as mouse clicks and keyboard inputs are captured and translated to the input commands compatible with the remote device 114 at the datacenter 112. For example, the mouse click in the URL field 124 is translated to a tap on the corresponding location on the display of the remote device 114 and the keyboard inputs are translated to keyboard inputs of the remote device 114, causing the remote device 114 to open the user requested URL and download the user requested website. Simultaneously, a video stream of the display of the remote device 114 is sent to and generated on the display 120 on browser 106. In this manner, the user perceives entering a URL in the URL field 124 and seeing the display 120 (a replica of the remote device 114) open the requested URL. Additional interactions of the user 102 can continue in the same manner. The user 102 can use the RTS 100 in the manner described above to perform manual or automated testing.

The display 120 is a pictorial and graphical representation of the remote device 114. The RTS 100 does not open a copy of the browser opened on the remote device 114 or conduct simultaneous parallel processes between the remote device 114 and the local machine 106. Instead, the RTS 100 streams a video feed from the remote device 114 to generate the display 120. Consequently, the user's interactions is inputted to the display 120, appearing as if a functioning browser is receiving the interactions, while the RTS 100 captures, transfers and translates those interactions to the remote device 114, where the functioning browser is operating on the remote device 114.

FIG. 2 illustrates a diagram 200 of an example data flow implementation of the RTS 100. The example shown in FIG. 2 will be described in the context of the user 102 requesting to start a remote session. The remote session can be used for a variety of purposes. In one example, the remote session can be used to test a web application or a website. The user launches a dashboard application using the browser 106, running on the user's local machine 104. The dashboard application can provide menu options to the user 102 to choose initial test session parameters, including a type/brand of a test device, operating system, a browser brand, and an initial test URL to access. The browser 106, running the dashboard application, can generate and send a request 220 for starting a remote session to the server 110. The server 110 can be a central or a distributed server over several geographical locations, enabling access to the RTS 100 from various locations. The request 220 can include details, such as a type/brand of a test device, operating system, a browser brand, and an initial test URL to access. In response to the user's request 220, the RTS 100 can select a datacenter 112, a test device 114, and can dynamically generate a test session identifier (ID). In some embodiments, a communication network is used to enable communication between the browser 106 and the remote device 114. The RTS 100 can choose a communication initiation server (CIS) 202 and associate the test session ID with the CIS 202. The selected CIS 202 can be communicated to both the browser 106 and the remote device 114, using an identifier of the selected CIS 202 or a CIS ID. In some embodiments, the CIS 202 can help the browser 106 and the remote device 114 to establish a peer-to-peer (P2P) communication network to directly connect. Other communication networks can also be used.

The server 110 can provide initial handshake data to both the remote device 114 and the browser 106, in order to establish a communication network. For example, after choosing the CIS 202 and other initial parameters, the server 110 can issue a start session response 222 to the browser 106. The start session response 222 can include details, such as the test session ID and an identifier of the CIS 202 to be used for establishing communication. The server 110 can send a session parameter message (SPM) 224 to the host 118. The SPM 224 can include parameters of the test session, such as the CIS ID, selected device ID, test session ID, browser type, and the requested URL. The host 118 routes the SPM 224 via a message 226 to a communication module (CM) 204 of the remote device 114. The CM 204 can be a hardware, software or a combination component of the remote device 114, which can handle the communication with the browser 106. Depending on the type of communication network and protocol used, the structure and functioning of the CM 204 can be accordingly configured. For example, in some embodiments, the CM 204 can handle WebRTC messaging, encoding of the screenshots from the remote device 114, transmitting them to the browser 106 and handling the interactions received from the browser 106.

The browser 106, via the start session response 222 receives the CIS 202 ID and the test session ID. The CM 204, via the message 226, receives the same information. The CM 204 can send a device connection message (DCM) 228 to the CIS 202. The browser 106 can send a browser communication message (BCM) 230 to the CIS 202. Both DCM 228 and BCM 230 use the same test session ID. Therefore, the CIS 202 can authenticate both and connect them. Once connected, the browser 106 and the remote device 114 can exchange communication data and the routes via which they can communicate. For example, they can indicate one or more intermediary servers that may be used to carry on their communication.

In some embodiments, Web real-time communication (WebRTC) can be used to enable communication between the remote device 114 and the browser 106, for example, when the remote device 114 is a smartphone device. In this scenario, the CM 204 can include, in part, a libjingle module, which can implement the WebRTC protocol handshake mechanisms in the remote device 114. The handshake made available through the CIS 202 allows the remote device 114 and the browser 106 to exchange communication data routes and mechanisms, such as traversal using relays around NAT (TURN) servers, session traversal utilities for NAT (STUN) servers, interactive connectivity establishment (ICE) candidates, and other communication network needs. NAT stands for Network Address Translation.

Once the communication network between the browser 106 and the remote device 114 is established, a plurality of channels can be established between the two. Each channel can in turn include a plurality of connections. For example, the communication network between the browser 106 and the remote device 114 can include a video communication channel (VCC) 232. The VCC 232 can include a plurality of connections between the browser 106 and the remote device 114 and can be used to transmit a video stream of the display of the remote device 114 to the browser 106. The communication network between the browser 106 and the remote device 114 can also include a data communication channel (DCC) 234. The DCC 234 can include a plurality of connections between the browser 106 and the remote device 114 and be used to transmit the interactions the user 102 inputs into the mirrored display of the remote device generated on the browser 106. The mirrored display can alternatively be described as a replica display of the remote device 114.

To generate a mirrored display of the remote device 114 on the browser 106, the captured screenshots from a screen capturing application (SCA) 208 can be assembled into a video stream and transmitted to the browser 106. The process of assembling the screenshots from the SCA 208 to a video stream may include performing video encoding, using various encoding parameters. Encoding parameters may be dynamically modifiable or may be predetermined. As an example, the available bandwidth in VCC 232 can vary depending on network conditions. In some embodiments, a frames-per-second encoding parameter can be adjusted based in part on the available bandwidth in the VCC 232. For example, if a low bandwidth in VCC 232 is detected, the video stream constructed from the captured screenshots can be encoded with a downgraded frames-per-second parameter, reducing the size of the video stream, and allowing an interruption free (or reduced interruption) transmission of the live video stream from the remote device 114 to the browser 106.

Another example of dynamically modifying the encoding parameters include dynamically modifying, or modulating the encoding parameter, based on the availability of hardware resources of the remote device, or the capacity of the hardware resources of the remote device 114 that can be assigned to handle the encoding of the video stream. The CM 204 can use the hardware resources of the remote device 114 in order to encode and transmit the video stream to the browser 106. For example, CM 204 can use the central processing unit (CPU) of the remote device 114, a graphics processing unit (GPU) or both to encode the video stream. In some cases, these hardware resources can be in high usage, reducing their efficiency in encoding. The reduction in hardware resources availability or capacity can introduce interruptions in the encoding. In some embodiments, a frame rate sampling parameter of the encoding parameters can be modulated based on the availability or capacity of hardware resources, such as the CPU and/or the GPU of the remote device 114 that can be assigned to handle the encoding of the video stream. For example, if a high CPU usage is detected, when the CPU is to be tasked with encoding, the CM 204 can reduce the sampling rate parameter of the encoding, so the CPU is not overburdened and interruptions in the video feed are reduced or minimized.

The CM 204 can also configure the encoding parameters, based on selected parameters at the browser 106. The browser 106 receives the video stream via the VCC 232, decodes the video stream and displays in the video stream in a replica display of the remote device 114 on the browser 106. In some embodiments, a predetermined threshold frames-per-second parameter of the video stream at the browser 106 can be selected. The predetermined threshold frames-per-second parameter can be based on a preselected level of quality of the video stream displayed on the replica display. For example, in some embodiments, the predetermined threshold frames-per-second parameter at the browser can be set to a value above 25 frames-per-second to generate a seamless and smooth mirroring of the display of the remote device 114 on the browser 106. The CM 204 can configure the encoding parameters at the remote device 114 based on the predetermined threshold frames-per-second parameter set at the browser 106. For example, the CM 204 can encode the video stream with a frame rate above 30 fps, so the decoded video stream at the browser 106 has a frames-per-second parameter above 25 fps.

In some embodiments, the screen capturing application (SCA) 208 can be used to capture screenshots from the remote device 114. The SCA 208 can differ from device to device and its implementation and configuration can depend on the processing power of the device and the mandates of the operating system of the device regarding usage of the CPU/GPU in capturing and generating screenshots. For example, in Android® environment, the Android® screen capture application programming interface (APIs) can be used. In iOS® devices, iOS® screen capture APIs can be used. Depending on the processing power of the selected remote device 114, the SCA 208 can be configured to capture screenshots at a predefined frames per second (fps) rate. Additionally, the SCA 208 can be configured to capture more screenshots at the remote device 114 than the screenshots that are ultimately used at the browser 106. This is true in scenarios where some captured screenshots are dropped due to various conditions, such as network delays and other factors. For example, in some embodiments, the SCA 208 can capture more than 30 fps from the display of the remote device 114, while at least 20 fps or more are able to make it to the browser 106 and shown to the user 102. In the context of packaging and assembling the captured screenshots into a video stream transmitted to the browser 106, screenshots that are received out of order may need to be dropped to maintain a fluid experience of the remote device 114 to the user 102. For example, the captured screenshots are streamed over a communication network to the browser 106, using various protocols, including internet protocol suite (TCP/IP), user datagram protocol, and/or others. When unreliable transmission protocols are used, it is possible that some screenshots arrive at browser 106 out of order. Out of order screenshots can be dropped to maintain chronology at the video stream displayed on browser 106. Some captured screenshots might simply drop as a result of other processing involved. For example, some screenshots may be dropped, due to lack of encoding capacity, if heavy animation on the remote device 114 is streamed to the browser 106. Consequently, in some embodiments, more screenshots are captured at the remote device 114 than are ultimately shown to the user 102.

The upper threshold for the number of screenshots captured at the remote device 114 can depend, in part, on the processing power of the remote device 114. For example, newer remote devices 114 can capture more screenshots than older or midmarket devices. The upper threshold for the number of screenshots can also depend on an expected bandwidth of a communication network between the remote device 114 and the browser 106.

The SCA 208 can be a part of or make use of various hardware components of the remote device 114, depending on the type of the selected remote device 114, its hardware capabilities and its operating system requirements. For example, some Android® devices allow usage of the device's graphical processing unit (GPU), while some iOS® devices limit the usage of GPU. For remote devices 114, where the operating system limits the use of GPU, the SCA 208 can utilize the central processing unit (CPU) of the remote device 114, alone or in combination with the GPU to capture and process the screenshots. The SCA 208 can be implemented via the screen capture APIs of the remote device 114 or can be independently implemented. Compared to command line screen capture tools, such as screencap command in Android®, the SCA 208 can be configured to capture screenshots in a manner that increases efficiency and reliability of the RTS 100. For example, command line screenshot tools, may capture high resolution screenshots, which can be unnecessary for the application of the RTS 100, and can slow down the encoding and transmission of the video stream constructed from the screenshots. Consequently, the RTS 100 can be implemented via modified native screenshot applications, APIs or independently developed and configured to capture screenshots of a resolution suitable for efficient encoding and transmission. As an example, using command line screen capture tools, a frames-per-second rate of only 4-5 can be achieved, which is unsuitable for mirroring the display of the remote device 114 on the browser 106 in a seamless manner. On the other hand, the described embodiments achieve frames-per-second rates of above 20 frames per second. In some embodiments, the CM 204 can down-sample the video stream obtained from the captured screenshots, from for example, a 4K resolution to a 1080P resolution. Still, in older devices, the down-sampling may be unnecessary, as the original resolution may be low enough for efficient encoding and transmission.

In some embodiments, the remote device 114 and the browser 106 can connect via a P2P network, powered by WebRTC. The CM 204 can then include a modified libjingle module. In the context of the RTS 100, the relationship between the browser 106 and the remote device 114 is more of a client-server type relationship than a pure P2P relationship. An example of a pure P2P relationship is video teleconferencing, where both parties transmit video to one another in equal and substantial size. In the context of the RTS 100, the transfer of video is from CM 204 to the browser 106, and no video is transmitted from the browser 106 to the CM 204. Therefore, compared to a P2P libjingle, the CM 204 and its libjingle module, as well as communication network parameters between the browser 106 and the remote device 114, can be modified to optimize for the transfer of video from the remote device 114 to the browser 106. An example modification of libjingle includes modifying the frames-per-second rate in favor of video transfer from the remote device 114. Other aspects of encoding performed by libjingle module of the CM 204 can include adding encryptions and/or other security measures to the video stream. When WebRTC is used to implement the communication network between the remote device 114 and the browser 106, libjingle module of the CM 204 can encode the video stream in WebRTC format.

While FIG. 2 illustrates messaging lines directly to the CM 204, this is not necessarily the case in all embodiments. In some implementations, the DCM 28, BCM 230, VCC 232, and DCC 234 can be routed through the host 118. The communication network between the remote device 114 and the browser 106, having channels, VCC 232 and DCC 234 can be implemented over the internet via a WiFi connection at the datacenter 112 where the remote device 114 is located, or can be via an internet over universal serial bus (USB) via the host 118, or a combination of wired or wireless communication to the internet. In some cases, one or more methods of connecting to the internet is used as a backup to a primary mode of connection to the internet and establishing the communication network between the remote device 114 and the browser 106.

The CM 204 can receive, via the DCC 234, user interactions inputted to the replica display on the browser 106. The CM 204 can route the received user interactions to an interaction server 206 for translation to a format compatible with the remote device 114. In a typical case, the user 102 runs the browser 106 on a laptop or desktop machine and inputs commands and interacts with the replica display on the browser 106, using the input devices of the local machine 104. Input devices of the local machine 104 generate mouse or keyboard user interactions, which are captured and transferred to the CM 204. In some embodiments, JavaScripts® can be used to capture user interactions inputted in the replica display on the browser 106. The captured user interactions are then encoded in a format, compatible with the format of the communication network established between the browser 106 and the remote device 114. For example, if WebRTC is used, the user interactions are formatted in the WebRTC format and sent over the DCC 234 to the CM 204.

The CM 204 decodes and transfers the user interactions to the interaction server 206. The interactions server 208 translates the mouse and keyboard user interactions to inputs compatible with the remote device 114. For example, when the remote device 114 is a mobile device, such as a smartphone or tablet having a touch screen as an input device, the interaction server 206 can translate keyboard and mouse inputs to gestures, swipes, pinches, and other commands compatible with the remote device 114. The translation of user interactions to remote device inputs also takes advantage of the coordinates of the inputs. For example, a meta data file accompanying the user interactions can note the coordinates of the user interactions on the replica display on the browser 106. The meta data can also include additional display and input device information of the user local machine 104 and the replica display on the browser 106.

The interaction server 206 also maintains or has access to the resolution and setup of the display of the remote device 114 and can make a conversion of a coordinate of an input on the replica display versus a corresponding coordinate on the real display of the remote device 114. For example, in some embodiments, the interaction server 206 can generate coordinate multipliers to map a coordinate in the replica display on the browser 106 to a corresponding coordinate in the real display of the remote device 114. The coordinate multipliers can be generated based on the resolutions of the replica display and the real display. The interaction server 206 then inputs the translated user interactions to the remote device 114. The display output of the remote device 114 responding to the input of the translated user inputs are captured via the SCA 208, sent to the CM 204, encoded in a format compatible with the communication network between the remote device 114 and the browser 106 (e.g., WebRTC) and sent to the browser 106. The browser 106 decodes the received video stream, displaying the video stream in the replica display on the browser 106. The data flow over the DCC 234 and the VCC 232 happen simultaneously or near simultaneously, as far as the perception of the user 102, allowing for a seamless interaction of the user 102 with the replica display, as if the remote device 114 were present at the location of the user 102.

FIG. 3 illustrates a flow chart of a method 300 of enabling a remote session at a first location using a remote device at a second location. The method 300 utilizes the RTS 100 as described above. The method 300 starts at step 302. At step 304, the browser 106 at a first location issues a request 220 to start a remote session at the first location, using a remote device at the second location. The request 220 can include a type/brand of a remote device, a browser to be opened on the remote device and a test URL to be accessed on the remote device. At step 306, the request 220 is received at a dashboard application of the RTS 100. The dashboard application may be locally installed, as a desktop application or may be a web application, accessible via a URL entered in the browser 106. The dashboard application can be powered by a server 110. At step 308, the server 110 can select a remote device 114 from a plurality of remote devices at the second location. The selection of the remote device is based on the user choice in the request 220. The selected remote device 114 can launch the browser type/brand, as indicated in the request 220. The selected remote device 114 can access the test URL, as indicated in the request 220.

At step 310, the server 110 selects a communication initiation server (CIS) 202 to allow the browser 106 and the selected remote device 114 to establish a connection. At step 312, both the browser 106 and the remote device 114 connect to the CIS 202, using the same test session ID. At step 314, the browser 106 and the remote device 114, via the CIS 202, exchange parameters of a communication network between the two. At step 316, the browser 106 and the remote device 114 establish the communication network, using the exchanged parameters. The exchanged parameters can include the routes, ports, gateways, and other data via which the browser 106 and the remote device 114 can connect. The communication network between the two includes a video channel, VCC 232 and a data channel, DCC 234.

At step 318, a replica display of the selected remote device 114 is generated in the browser 106. The browser 106 can receive, via the video channel, a video stream of the display output of the remote device 114 and use that to generate the replica display. At step 320, user interactions with the replica display are captured and transmitted, via the data channel DCC 234 to the remote device 114. At step 322, the SCA 208 captures screenshots of the display screen of the remote device 114. The CM 204 uses the captured screenshots to generate a video stream of the screen of the remote device 114. The CM 204 transmits, via the video channel VCC 232, the video stream to the browser 106, which uses the video stream to generate the replica display. The method 300 ends at step 324.

FIG. 4 illustrates a flowchart of a method 400 of an example operation of the RTS 100. The method 400 starts at step 402. At step 404, a request to start a remote session using a remote device is received at a dashboard application, powered by a server 110. The server 110 selects a CIS 202, a remote device 114 and issues a response to the browser 106. The response includes an identifier of the CIS 202 and an identifier of the test session. At step 406, the browser 106 and the remote device 114 establish a communication network and connect to one another using the communication network. The remote device 114 connects to the communication network via a host 118.

At step 408, the CM 204 generates a video stream from the screenshots captured by the SCA 208, based on one or more encoding parameters. An example of the encoding parameters includes a frames-per-second parameter of the encoding. At step 410, the CM 204 modules the encoding parameters based on one or more factors, including bandwidth of the VCC 232, and available capacity of hardware resources of the remote device 114 for encoding operations, including capacity of CPU and/or GPU of the remote device 114. The CM 204 can also modulate the encoding parameters based on a predetermined minimum threshold of frames per second video stream decoded and displayed at the browser 106. At step 412, the CM 204 transmits the video stream to the browser 106 to display. The method 400 ends at step 414.

FIG. 5 illustrates a flowchart of a method 500 of an example operation of the RTS 100. The method starts at step 502. At step 504, a communication network is established between the browser 106 and the remote device 114. In some embodiments, the communication network can be a P2P, WebRTC. The CM 204 in the remote device 114 can handle the translation, encoding and data packaging for transmission over the communication network. At step 506, a data channel DCC 234 is established using the communication network. The data channel can be used to transmit user interactions entered into a replica display in browser 106 to the remote device 114. At step 508, the user interactions with the replica display on browser 106 are captured and transmitted to the CM 204. In some embodiments, the DCC 234 is through the host 118 and in other embodiments, a WiFi network at datacenter 112 where the remote device 114 is located can be used to connect the CM 204 and the browser 106. The CM 204 transfers the user interactions to the interaction server 206.

At step 510, the interaction server 206 translates the user interactions to user inputs compatible with the remote device 114. For example, if the remote device 114 is a mobile computing device, such as a smartphone or smart tablet, the interaction server 206 translates keyboard and mouse inputs to touch screen type inputs, such as taps, swipes, pinches, double tap, etc. The interaction server 206 may use coordinate multipliers to translate the location of a user interaction to a location on the display of the remote device 114. The coordinate multipliers are derived from the ratio of the resolution and/or size difference between the replica display on the browser 106 and the display screen of the remote device 114. At step 512, the user inputs are inputted into the remote device 114 at the corresponding coordinates. The remote devices's display output response to the user inputs are captured via SCA 208, turned into a video stream and transmitted to the browser 106. The browser 106 displays the video stream in the replica display. The method 500 ends at the step 514.

Example Implementation Mechanism—Hardware Overview

Some embodiments are implemented by a computer system or a network of computer systems. A computer system may include a processor, a memory, and a non-transitory computer-readable medium. The memory and non-transitory medium may store instructions for performing methods, steps and techniques described herein.

According to one embodiment, the techniques described herein are implemented by one or more special-purpose computing devices. The special-purpose computing devices may be hard-wired to perform the techniques, or may include digital electronic devices such as one or more application-specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs) that are persistently programmed to perform the techniques, or may include one or more general purpose hardware processors programmed to perform the techniques pursuant to program instructions in firmware, memory, other storage, or a combination. Such special-purpose computing devices may also combine custom hard-wired logic, ASICs, or FPGAs with custom programming to accomplish the techniques. The special-purpose computing devices may be server computers, cloud computing computers, desktop computer systems, portable computer systems, handheld devices, networking devices or any other device that incorporates hard-wired and/or program logic to implement the techniques.

For example, FIG. 6 is a block diagram that illustrates a computer system 600 upon which an embodiment of can be implemented. Computer system 600 includes a bus 602 or other communication mechanism for communicating information, and a hardware processor 604 coupled with bus 602 for processing information. Hardware processor 604 may be, for example, special-purpose microprocessor optimized for handling audio and video streams generated, transmitted or received in video conferencing architectures.

Computer system 600 also includes a main memory 606, such as a random access memory (RAM) or other dynamic storage device, coupled to bus 602 for storing information and instructions to be executed by processor 604. Main memory 606 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 604. Such instructions, when stored in non-transitory storage media accessible to processor 604, render computer system 600 into a special-purpose machine that is customized to perform the operations specified in the instructions.

Computer system 600 further includes a read only memory (ROM) 608 or other static storage device coupled to bus 602 for storing static information and instructions for processor 604. A storage device 610, such as a magnetic disk, optical disk, or solid state disk is provided and coupled to bus 602 for storing information and instructions.

Computer system 600 may be coupled via bus 602 to a display 612, such as a cathode ray tube (CRT), liquid crystal display (LCD), organic light-emitting diode (OLED), or a touchscreen for displaying information to a computer user. An input device 614, including alphanumeric and other keys (e.g., in a touch screen display) is coupled to bus 602 for communicating information and command selections to processor 604. Another type of user input device is cursor control 616, such as a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 604 and for controlling cursor movement on display 612. This input device typically has two degrees of freedom in two axes, a first axis (e.g., x) and a second axis (e.g., y), that allows the device to specify positions in a plane. In some embodiments, the user input device 614 and/or the cursor control 616 can be implemented in the display 612 for example, via a touch-screen interface that serves as both output display and input device.

Computer system 600 may implement the techniques described herein using customized hard-wired logic, one or more ASICs or FPGAs, firmware and/or program logic which in combination with the computer system causes or programs computer system 600 to be a special-purpose machine. According to one embodiment, the techniques herein are performed by computer system 600 in response to processor 604 executing one or more sequences of one or more instructions contained in main memory 606. Such instructions may be read into main memory 606 from another storage medium, such as storage device 610. Execution of the sequences of instructions contained in main memory 606 causes processor 604 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions.

The term “storage media” as used herein refers to any non-transitory media that store data and/or instructions that cause a machine to operation in a specific fashion. Such storage media may comprise non-volatile media and/or volatile media. Non-volatile media includes, for example, optical, magnetic, and/or solid-state disks, such as storage device 610. Volatile media includes dynamic memory, such as main memory 606. Common forms of storage media include, for example, a floppy disk, a flexible disk, hard disk, solid state drive, magnetic tape, or any other magnetic data storage medium, a CD-ROM, any other optical data storage medium, any physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, NVRAM, any other memory chip or cartridge.

Storage media is distinct from but may be used in conjunction with transmission media. Transmission media participates in transferring information between storage media. For example, transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise bus 602. Transmission media can also take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.

Various forms of media may be involved in carrying one or more sequences of one or more instructions to processor 604 for execution. For example, the instructions may initially be carried on a magnetic disk or solid state drive of a remote computer. The remote computer can load the instructions into its dynamic memory and send the instructions over a telephone line using a modem. A modem local to computer system 600 can receive the data on the telephone line and use an infra-red transmitter to convert the data to an infra-red signal. An infra-red detector can receive the data carried in the infra-red signal and appropriate circuitry can place the data on bus 602. Bus 602 carries the data to main memory 606, from which processor 604 retrieves and executes the instructions. The instructions received by main memory 606 may optionally be stored on storage device 610 either before or after execution by processor 604.

Computer system 600 also includes a communication interface 618 coupled to bus 602. Communication interface 618 provides a two-way data communication coupling to a network link 620 that is connected to a local network 622. For example, communication interface 618 may be an integrated services digital network (ISDN) card, cable modem, satellite modem, or a modem to provide a data communication connection to a corresponding type of telephone line. As another example, communication interface 618 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN. Wireless links may also be implemented. In any such implementation, communication interface 618 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.

Network link 620 typically provides data communication through one or more networks to other data devices. For example, network link 620 may provide a connection through local network 622 to a host computer 624 or to data equipment operated by an Internet Service Provider (ISP) 626. ISP 626 in turn provides data communication services through the world wide packet data communication network now commonly referred to as the “Internet” 628. Local network 622 and Internet 628 both use electrical, electromagnetic or optical signals that carry digital data streams. The signals through the various networks and the signals on network link 620 and through communication interface 618, which carry the digital data to and from computer system 600, are example forms of transmission media.

Computer system 600 can send messages and receive data, including program code, through the network(s), network link 620 and communication interface 618. In the Internet example, a server 630 might transmit a requested code for an application program through Internet 628, ISP 626, local network 622 and communication interface 618. The received code may be executed by processor 604 as it is received, and/or stored in storage device 610, or other non-volatile storage for later execution.

Video Feed for Generating a Mirrored or Replica Display

Some remote devices 114 do not provide a high-performance screenshot capturing API, suitable for efficient operations of the RTS 100. On the other hand, some operating systems of the remote devices 114 can support a video capturing API for the purposes of recording and/or broadcasting the display of the remote device 114 in real time. In these scenarios, the SCA 208 can be implemented using a video capturing API of the operating system of the remote device 114. As an example, for some iOS® devices, when the SCA 208 is implemented, using a native screenshot application, the FPS achieved on browser 106 can drop to as low as 5 FPS in some cases. At the same time, iOS® in some versions, provides a video capturing facility, such as ReplayKit, which can be used to implement the operations of the SCA 208. When a video capturing API is used, corresponding modifications to the data flow and operations of the RTS 100 are also implemented as will be described below.

FIG. 7 illustrates an example data flow diagram 700 of the operations of the RTS 100, using a video capturing API for implementing the SCA 208. The diagram 700 is provided as an example. Persons of ordinary skill in the art can modify the diagram 700, without departing from the spirit of the disclosed technology. Some platforms and operating systems may provide an API for capturing a video stream of the remote device 114. For example, iOS® provides such an API in ReplayKit. The captured video stream can be used to replicate the display of the remote device 114 in lieu of using static screenshots to generate the video stream. In some cases, the SCA 208 can be implemented using the video capturing API provided by the remote device 114. For example, a launcher application can include a broadcaster extension, which can output a video stream of the display of the remote device 114. In other embodiments, a broadcast extension, broadcasting the video stream, can be an extension to a launcher application, which the host 118 uses to control the operations of the remote device 114. Various implementations are possible. Some are described below.

At step 702, the browser 106 can send a request 220 to start a remote session to the server 110. At step 704, the server 110 can respond by sending a response 222 to the browser 106. At step 706, the server 110 can send a SPM 224 to the host 118. At step 708, the host 118 can send a message 226 to the CM 204. The steps 702-708 enable the remote device 114 and the browser 106 to login to a communication initiation server (CIS) 202 with the same credentials, such as a common remote session identifier, thereafter, exchange communication network parameters, and establish communication using the communication network.

At step 710, the CM 204 can signal a broadcaster 712 to launch and begin capturing a video stream of the display of the remote device 114. As described earlier, the broadcaster 712 can be a stand-alone application or can be an extension to a launcher application that the host 118 runs on the remote device 114 to perform the operations of the RTS 100. For example, when ReplayKit is used, the ReplayKit API provides a broadcaster extension which can run as an extension of an application and provide a video stream of the display of the remote device 114 to that application.

At this stage, the DCM 228 and the BCM 230 have already occurred between the browser 106 and the CM 204, allowing the browser 106 and the CM 204 to exchange network communication parameters via the CIS 202. The network communication parameters can include network pathways, servers, and routes via which the two can establish one or more future communication networks. The browser 106 and the CM 204 establish a communication network and connect using these network communication parameters. At step 714, the CM 204 can establish a DCC 234 with the browser 106. The DCC 234 can be used in the future operations of the RTS 100 to capture user interactions on the replica display generated on the browser 106 and transmit them to the remote device 114. At step 716, the host can extract a requested URL and a type of browser from the user's initial request (at step 720) and launch the chosen browser on the remote device 114, with a request for the remote device browser to access the user requested URL.

At step 718, the broadcaster 712 can query the host 118 for session and user data to determine where and how to establish a video channel to broadcast the video stream feed of the display of the remote device 114. At step 720, the host 118 responds to the broadcaster 712 with session and user data. The session and user data can include an identifier of the session, a user identifier, network details, gates and ports, pathways or other information related to the remote session and/or the communication network established between the CM 204 and the browser 106. At step 722, the broadcaster 712 can use the session and/or user data, received at step 720, to establish the VCC 232 and begin broadcasting the video stream of the display of the remote device 114 to the browser 106. A dashboard application, executable on and/or by the browser 106, can generate a replica display of the remote device 114 on the browser 106 and use the video stream received on the VCC 232 to populate the replica display with a live video feed of the display of the remote device 114. In some implementations, the CM 204 can set up or modify the encoding parameters of the video from the broadcaster 712. For example, the CM 204 can be configured to determine the bandwidth of the VCC 232 and modify the FPS encoding parameter of the video stream to increase the likelihood of an efficient, stable and/or performant video stream on the browser-end. The dashboard application executable on the browser 106 can decode the video received on the VCC 232 and use the decoded video to generate the replica display on the browser 106. Other examples of the CM 204 modifying the encoding parameters of the video sent on the VCC 232 are described above in relation to the pervious embodiments. The CM 204 can apply the same techniques to the embodiments, where a broadcaster 712 is used. As described earlier, having the VCC 232 consume a video stream, via the broadcaster 712, can offer advantages, such as more efficient encoding, and having a higher and more stable FPS performance.

Multisession Mode

In some applications of the RTS 100, the user 102 may desire to test and/or develop an application, for example a web application, and/or a mobile application on multiple remote devices 114, simultaneously, and observe the display of each remote device 114 on the browser 106 simultaneously. In this scenario, the user 102 can test and develop an application on multiple platforms and devices at the same time, receiving a display of each device on the user's browser 106. This can substantially increase the efficiency of test and development using the RTS 100. As an example, in the context of users developing financial applications, the ability to test multiple devices simultaneously offers substantial efficiencies. Many governments impose manual testing requirements on financial applications that operate in financial markets and institutions. In this scenario, the user 102 may be a developer tasked with certifying the compliance of an application with financial and governmental regulations on multiple devices and platforms. Such a user can benefit from receiving a display feed of multiple remote devices 114 and observing the behavior of the application in those devices simultaneously. For example, as part of testing, the user may have to enter data into the application and observe the behavior of the application in response to the entered data. Testing and development would be more efficient if the user entered the test in one application and on one remote device 114 and the same input was synced across multiple remote devices 114 that the user may wish to simultaneously test. Therefore, the user 102 can benefit and realize efficiencies if the remote devices 114 can receive and respond to the interactions with the user 102 in a synchronized manner. Examples include the user 102 scrolling a webpage, clicking on an element, and entering a login credential. The ability to test multiple remote devices simultaneously in a synchronized manner can be referred to as using the RTS 100 in “multisession mode,” or enabling the “multisession” mode in the RTS 100.

FIG. 8 illustrates an environment 800 of multisession operations of the RTS 100. The user 102 is selecting to test or develop the web application 802 in multisession mode. The user can select any number of remote devices 114 to conduct test and development activities. Choosing two or more remote devices 114 can trigger the multisession mode. In the example shown, the user 102 has selected three remote devices 114. The remote devices 114 can be from any datacenter 112. The RTS 100 can assign remote devices 114 from the same datacenter 112 or from different datacenters 112 and/or a combination of one or more datacenters 112. In some cases, the remote devices assigned to user 102 for multisession can be from datacenters in different countries and/or even different continents. The user 102 can connect with the selected remote devices 114, via the embodiments described above. For example, a data channel and a video channel can be established between each remote device 114 and the local computer of the user 102. In some embodiments, an audio channel can also be established between the remote device 114 and the local computer of the user 102. For ease of discussion, the embodiments of multisession are described where the web application 802 is a webpage and the user 102 utilizes the browser 106 to conduct multisession testing. However, the user can also test a mobile or other programming application in the same or similar manner as described herein with respect to the web application 802. Furthermore, in some embodiments, the user 102 can utilize a desktop application on the user's local machine to access the RTS 100 and conduct multisession testing.

A video feed of each selected remote device 114 is used to generate a mirrored display of each selected remote device on the browser 106. In this manner, whatever is displayed on the screen of the remote device 114 is also displayed on the browser 106. Furthermore, the RTS 100 can generate the mirrored display in a shell or skin or perimeter graphics resembling the selected remote device 114 to provide a visual aid to the user 102 as to which remote device the user 102 is testing and viewing on the browser 106. For example, if the selected remote device 114 is an iPhone®, the mirrored display generated on the browser 106 can be presented to the user 102 in a shell of the same iPhone®. Other methods of identification, such as texts, graphics or other visual identifiers can also be used, in addition to or in lieu of replicating the selected remote device 114, to distinguish the selected remote devices in a multisession mode.

In the example shown, the user 102 has chosen three remote devices 114. The display of each remote device 114 is mirrored on the browser 106, generating mirrored displays 804, 806, 808. For ease of illustration three remote device mirrored displays are shown, but a multisession mode can be enabled with only two remote devices 114 or with more than three remote devices 114.

The user 102 can interact with the mirrored displays 804, 806, 808, via any input devices or hardware available at local machine of the user. Example input devices can include keyboard, mouse, and touch screen. The user interactions with the mirrored displays are captured and transmitted to the remote devices 114. As described earlier, the user interactions are translated to gestures and/or inputs compatible with the remote device 114 and inputted to the remote device 114. In multisession mode, the user interactions with one mirrored display are synced with the other mirrored displays in the multisession. For example, if the user 102 scrolls the web application 802 in the mirrored display 806, the web application 802 also scrolls in the mirrored displays 804 and 806. If the user 102 enters a login credential on the mirrored display 804, those credentials are also synced and displayed simultaneously on the other mirrored displays 806, 808, as the user 102 types them in an appropriate webpage field of the web application 802.

FIG. 9 illustrates a block diagram 900 of operations of multisession mode of the RTS 100. For ease of discussion, only two remote devices 114 in the multisession of the diagram 900 are shown, but the same technology can be used to enable three or more remote devices 114 in a multisession mode. The user 102 selects a webpage 902 to test in the browser 106, using a multisession mode with two remote devices 904 and 906. In this scenario, a connection session between the remote devices 904, 906 and the browser 106 is established. The connection session includes establishing a data channel 912 and a video channel 914 between the remote device 904 and the browser 106. The connection session also includes a data channel 916 and a video channel 918 between the remote device 906 and the browser 106. The video channel 914 broadcasts a video feed of the display of the remote device 904 to the browser 106, generating the mirrored display 908. The video channel 918 broadcasts a video feed of the display of the remote device 906 to the browser 106, generating the mirrored display 910.

In some embodiments, a multisession mode can use coordinate mapping to perform simultaneous testing and synced interactions between multiple remote devices. In this approach, when the user 102 enters an interaction in a mirrored display, the coordinates of the location of the interactions on the corresponding remote device are determined. Then corresponding locations in the displays of the other remote devices in the multisession are determined and the same interaction is entered in each remote device, via their respective data channels. In other words, the coordinates and the type of interactions are transmitted to each remote device via their respective data channels.

In some cases, the coordinate mapping approach can be challenging, as the remote devices in a multisession can vary in specification, display resolution, size and other characteristics that can make coordinate mapping between multiple devices challenging. Consequently, in some embodiments, coordinate mapping can be replaced or augmented by use of a sync server. The sync server receives or otherwise detects user interactions on one of the remote devices and serves a synced output to all remote devices in a multisession.

The user 102 enters user interactions to one of the mirrored displays 908, 910. For example, the user 102 can enter interactions with the mirrored display 908. The user interactions entered in the mirrored display 908 are transmitted via the data channel 912 to the remote device 904. The RTS 100 can determine the coordinates and/or other indicator of the location of user interactions with the mirrored display 908. The RTS 100 further can determine the user interface (UI) element upon which the user interactions are entered. For example, in the case of the webpage 902, the RTS 100 can determine an HTML element upon which the user interactions are entered. The user interactions sent to the remote device 904 can be detected and/or otherwise transmitted to the sync server 920, along with the identity of the UI element upon which the user interactions are entered. In other embodiments, the functionality of determining the UI element corresponding to the coordinates and/or location of a user interaction can be implemented as part of the sync server 920 functionality and/or in other parts of the RTS 100. The sync server 920 can generate a synced version of the output and/or the result of the user interaction on the webpage 902 and serve the synced version to all remote devices in the multisession (the remote device 906 in this example). If the user 102 starts interacting with the mirrored display 910, the same operations are performed with the roles of the remote devices 904, 906 reversed, where the user interactions are sent to the remote device 906. The sync server 920 detects and/or otherwise receives these interactions, coordinates and/or UI elements upon which the user interactions are entered and serves a synced version of the output, in this case the webpage 902, to all remote devices in the multisession, including the remote device 904. The sync server 920 serves a synced version of the output to all remote devices in a multisession and all remote devices broadcast their displays via their respective video channels back to the browser 106. In this manner, the user 102 observes a synced display in each mirrored display 908, 910.

In the case of a webpage 902 as the test application, a browser local to each remote device and running on each remote device can render the output as received from the sync server 920, suitable to the remote device in which the browser is running. Therefore, each remote device receiving an input from the sync server 920 renders the webpage 902 on the remote device correctly. In this manner, the need for coordinate mapping from one remote device to another is alleviated. Each remote device, receiving the same copy of the webpage 902 from the sync server 920 renders the webpage 902 on its display correctly. The same dynamic can be applicable when testing programming applications other than websites. The sync server can provide the same copy of the output to all remote devices, and each remote device can render the output correctly according to the internal facilities and specifications of each remote device.

In some embodiments, the sync server 920 serving the webpage 902 to each remote device can perform maintaining and updating operations, where the sync server 920 maintains and updates the same copy of the webpage 902 on each remote device in real time. In this manner, changes in one remote device reflect, in real time, in the other remote devices in the multisession.

While the above embodiments are described in terms of testing a webpage 902 in a multisession environment, the same or similar technology can be applied to testing other programming applications. For example, the sync server 920 and/or components or processes of the RTS 100 can determine the UI elements upon which the user interactions are operating and serve a synced version of the programming application, with the user interactions implemented, to all remote devices in a multisession.

Some user interactions can include a request that can be routed to a host 924 hosting the webpage 902. The sync server 920 can route the request via a network 922 to the host 924. The network 922 can be the Internet, a public, private network and/or a combination of various networks. Similarly, the host 924 can be a public and/or private host. The sync server 920 receives a response from the host 924 and uses the response to serve a synced copy of the webpage 920 to all remote devices in a multisession. An example of a user interaction triggering a request/response type operation in the multisession is “entering login credentials” or “clicking on a link” in the webpage 902. Not every user interaction triggers a request/response type of operation requiring communication with the host 924. In some embodiments, a user interaction can be a command for scrolling action. In this scenario, the sync server 920 can perform a corresponding scrolling action on the synced output, where the browser of each remote device performs a corresponding scrolling action on the webpage 902 and displays the result on its respective screen. Since the remote devices in a multisession can be of different sizes, orientation, and/or resolution, the displayed result of the scrolling can be different on each remote device.

In the case of testing a webpage using multisession, such as the webpage 902, the webpage can be loaded by a browser local to each remote device. The browser in each remote device can access the sync server 920 via the same uniform resource locator (URL) address. As the sync server 920 detects, captures, or receives user interactions from a remote device, the sync server 920 updates the webpage 902. Each browser in each remote device in a multisession, accessing the sync server 920 via the same URL, receives the same copy of the webpage 902 updated with the user interactions received from the other remote devices in the multisession. The display of each remote device is broadcast to the browser 106, thereby displaying a synced version of the webpage 902 on the browser 106.

FIG. 10 illustrates a flowchart of a method 1000 of performing a multisession mode. The method starts at step 1002. At step 1004, a connection session between the browser 106 and two or more remote devices 114 is established. Establishing the connection session can include establishing various communication channels between the remote devices and the browser 106, including data channels and video channels. The video channels can be used to broadcast or stream a video feed of the display of the remote devices 114 to the browser 106. The data channels can be used to transmit user interactions from the browser 106 to the remote devices 114. At step 1006, mirrored displays of the remote devices 114 are generated on the browser 106 by streaming a video feed of each remote device 114 to the browser 106. The mirrored displays can be generated in a manner to replicate an image of the remote device 114 from which they are originated.

At step 1008, user interactions with a mirrored display are captured and transmitted to the corresponding remote device 114 via the data channel of that remote device 114. At step 1010, a UI element upon which the user interactions are performed is determined. In some embodiments, determining the UI element can include determining the coordinates of where the user interactions are received on a mirrored display and mapping the mirrored display coordinates to a coordinate on the corresponding remote device 114 and to a UI element known to be at those coordinates in the remote device 114. At step 1012, a synced output of the user interactions are generated, based on the user interactions and the UI element upon which the user interactions are performed. In some embodiments, a sync server 920 can generate or update the synced output by performing the user interaction upon the determined UI element and triggering the operations corresponding to the determined UI element. The synced output can depend on the programming application and/or the webpage that is the subject of testing and development in the multisession. For example, a synced output can be a version of the webpage 902 updated with the user interactions inputted on the determined UI elements. In the case of a programming application tested and developed in a multisession using the method 1000, the synced output can be the output display of the programming application during and after the user interactions are inputted via the determined UI element.

At step 1014, the synced output is received by the other remote devices 114 in the multisession. In some embodiments, the sync server 920 can maintain and update a synced output in each remote device 114 in the multisession. In some embodiments, each remote device 114 in the multisession can receive the synced output and display the synced output according to its specification and internal facilities. In this manner, the task of rendering the display of the synced output in each remote device 114 is performed by the remote device 114 and the sync server is relieved from having to provide a compatible and unique synced output to each remote device. At step 1016, the display of each remote device 114 in the multisession is transmitted to the browser 106 via the video channel of each remote device 114. The user 102 viewing the mirrored displays on the browser 106 observes a synced version of the application or webpage in each mirrored display. The method ends at step 1018.

FIG. 11 illustrates a flowchart of a method 1100 of multisession operations including a request/response type of operation that can occur when developing and testing a webpage using the RTS 100. The method starts at step 1102. At step 1104, user interactions with one of the mirrored displays are received. The user interactions in this case can be a user interaction, which can be associated with generating a request to a host of a webpage that is being tested or developed using the RTS 100. At step 1106, the user interaction is transmitted as a request to the host which accepts and responds to requests relating to the webpage. In some embodiments, an intermediary, such as the sync server 920 can receive the user interactions and send the user interactions as a request to the host, such as the host 924 providing services for the webpage 902. At step 1108, a response to the request is received from the host. For example, the sync server 920 can receive the response. At step 1110, the sync server 920 can generate a synced output using the response. For example, the sync server 920 can generate or update a version of the webpage based on the received response from the host. In some embodiments, the updated version of the webpage is the response received from the host. The synced output can be provided to the remote devices in the multisession, from which mirrored displays on the browser 106 can be generated. The method ends at step 1112.

It will be appreciated that the present disclosure may include any one and up to all of the following examples.

Example 1: A method comprising: establishing a connection session between a browser and a first device and the browser and a second device, wherein the first and second devices are in one or more locations remote relative to the browser, and the connection session between the browser and the first device comprises a first device data channel and a first device video channel, and the connection session between the browser and the second device comprises a second device data channel and a second device video channel; generating a first device display on the browser by broadcasting, via the first device video channel, a first device video feed from the first device to the browser; generating a second device display on the browser by broadcasting, via the second device video channel, a second device video feed from the second device to the browser; transmitting, via the first device data channel, one or more user interactions with a webpage in the generated first device display on the browser, to the first device; transmitting the interactions to a sync server; and serving, from the sync server, a synced version of the webpage to the first and second devices.

Example 2: The method of Example 1, further comprising establishing the connection session further between a plurality of second devices, wherein the serving, from the sync server, comprises serving the synced webpage to the first device and the plurality of the second devices simultaneously.

Example 3: The method of some or all of Examples 1 and 2, wherein serving, from the sync server, further comprises: transmitting the interactions from the sync server to a host server hosting the webpage; receiving a response from the host server; and serving the response to the first and second devices.

Example 4: The method of some or all of Examples 1-3, wherein the interactions comprise commands for scrolling the webpage and the sync server performs a corresponding scrolling on the synced webpage served to the first and second devices.

Example 5: The method of some or all of Examples 1-4, wherein the sync server performs syncing between the first and second devices by determining webpage HTML elements corresponding to the user interactions.

Example 6: The method of some or all of Examples 1-5, wherein the first and second devices access the sync server via a shared URL corresponding to the webpage.

Example 7: The method of some or all of Examples 1-6, wherein transmitting the interactions to the sync server further comprises determining coordinates of the interactions on the generated first device display on the browser and identifying corresponding HTML elements of the webpage and transmitting the interactions and the coordinates to the sync server.

Example 8: A non-transitory computer storage that stores executable program instructions that, when executed by one or more computing devices, configure the one or more computing devices to perform operations comprising: establishing a connection session between a browser and a first device and the browser and a second device, wherein the first and second devices are in one or more locations remote relative to the browser, and the connection session between the browser and the first device comprises a first device data channel and a first device video channel, and the connection session between the browser and the second device comprises a second device data channel and a second device video channel; generating a first device display on the browser by broadcasting, via the first device video channel, a first device video feed from the first device to the browser; generating a second device display on the browser by broadcasting, via the second device video channel, a second device video feed from the second device to the browser; transmitting, via the first device data channel, one or more user interactions with a webpage in the generated first device display on the browser, to the first device; transmitting the interactions to a sync server; and serving, from the sync server, a synced version of the webpage to the first and second devices.

Example 9: The non-transitory computer storage of Example 8, wherein the operations further comprise establishing the connection session further between a plurality of second devices, wherein the serving, from the sync server, comprises serving the synced webpage to the first device and the plurality of the second devices simultaneously.

Example 10: The non-transitory computer storage of some or all of Examples 8 and 9, wherein serving, from the sync server, further comprises: transmitting the interactions from the sync server to a host server hosting the webpage; receiving a response from the host server; and serving the response to the first and second devices.

Example 11: The non-transitory computer storage of some or all of Examples 8-10, wherein the interactions comprise commands for scrolling the webpage and the sync server performs a corresponding scrolling on the synced webpage served to the first and second devices.

Example 12: The non-transitory computer storage of some or all of Examples 8-11, wherein the sync server performs syncing between the first and second devices by determining webpage HTML elements corresponding to the user interactions.

Example 13: The non-transitory computer storage of some or all of Examples 8-12, wherein the first and second devices access the sync server via a shared URL corresponding to the webpage.

Example 14: The non-transitory computer storage of some or all of Examples 8-13, wherein transmitting the interactions to the sync server further comprises determining coordinates of the interactions on the generated first device display on the browser and identifying corresponding HTML elements of the webpage and transmitting the interactions and the coordinates to the sync server.

Example 15: A system comprising a processor, the processor configured to perform operations comprising: establishing a connection session between a browser and a first device and the browser and a second device, wherein the first and second devices are in one or more locations remote relative to the browser, and the connection session between the browser and the first device comprises a first device data channel and a first device video channel, and the connection session between the browser and the second device comprises a second device data channel and a second device video channel; generating a first device display on the browser by broadcasting, via the first device video channel, a first device video feed from the first device to the browser; generating a second device display on the browser by broadcasting, via the second device video channel, a second device video feed from the second device to the browser; transmitting, via the first device data channel, one or more user interactions with a webpage in the generated first device display on the browser, to the first device; transmitting the interactions to a sync server; and serving, from the sync server, a synced version of the webpage to the first and second devices.

Example 16: The system of Example 15, wherein the operations further comprise establishing the connection session further between a plurality of second devices, wherein the serving, from the sync server, comprises serving the synced webpage to the first device and the plurality of the second devices simultaneously.

Example 17: The system of some or all of Examples 15 and 16, wherein serving, from the sync server, further comprises: transmitting the interactions from the sync server to a host server hosting the webpage; receiving a response from the host server; and serving the response to the first and second devices.

Example 18: The system of some or all of Examples 15-17, wherein the interactions comprise commands for scrolling the webpage and the sync server performs a corresponding scrolling on the synced webpage served to the first and second devices.

Example 19: The system of some or all of Examples 15-18, wherein the sync server performs syncing between the first and second devices by determining webpage HTML elements corresponding to the user interactions.

Example 20: The system of some or all of Examples 15-19, wherein the first and second devices access the sync server via a shared URL corresponding to the webpage.

While the invention has been particularly shown and described with reference to specific embodiments thereof, it should be understood that changes in the form and details of the disclosed embodiments may be made without departing from the scope of the invention. Although various advantages, aspects, and objects of the present invention have been discussed herein with reference to various embodiments, it will be understood that the scope of the invention should not be limited by reference to such advantages, aspects, and objects. Rather, the scope of the invention should be determined with reference to patent claims.

Jain, Ritik, Dube, Abhinav, Sonawane, Suyash Yogeshwar

Patent Priority Assignee Title
Patent Priority Assignee Title
10264215, Dec 18 2017 MasterCard International Incorporated Server and method enabling augmented reality (AR) object sharing
10523898, Apr 11 2018 Securus Technologies, LLC One-way video conferencing
10963138, Jul 22 2014 Allstate Insurance Company Webpage testing tool
11080177, May 04 2020 SAP SE Test controller for cloud-based applications
11151025, May 15 2020 Dell Products L.P. Generating software test plans based at least in part on monitored traffic of a production application
11470141, Feb 01 2022 BrowserStack Limited Remote device infrastructure
11470183, Jan 27 2022 BrowserStack Limited Deriving proxy stability without network inspection
11579595, Oct 13 2015 Schneider Electric Industries SAS Software defined automation system and architecture
11606390, Mar 30 2021 Juniper Networks, Inc Rerouting network traffic based on detecting offline connection
6662312, Jun 30 2000 Qwest Communications International Inc Software-testing automation system
8225191, Nov 07 2011 GOOGLE LLC Synchronizing web browsers
8881111, Sep 17 2013 Microsoft Technology Licensing, LLC Testing user interface responsiveness for mobile applications
20020143859,
20030098879,
20030140138,
20030229900,
20060259553,
20080247308,
20090007074,
20090019315,
20100235321,
20110107238,
20120221629,
20140052853,
20150074659,
20150268936,
20150304379,
20150324276,
20160261475,
20160330266,
20170300286,
20180109434,
20180129519,
20180167426,
20180210822,
20180329812,
20190129773,
20190190908,
20200106699,
20200111156,
20200137159,
20200278920,
20200296171,
20200341888,
20200379885,
20210400093,
20220070151,
20220150447,
20220345311,
20220374191,
20230195491,
CA2936503,
CN112799891,
JP2016212596,
KR20190036671,
////
Executed onAssignorAssigneeConveyanceFrameReelDoc
Sep 20 2022SONAWANE, SUYASH YOGESHWARBrowserStack LimitedASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0656310633 pdf
Sep 21 2022JAIN, RITIKBrowserStack LimitedASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0656310633 pdf
Sep 22 2022DUBE, ABHINAVBrowserStack LimitedASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0656310633 pdf
Sep 26 2022BrowserStack Limited(assignment on the face of the patent)
Date Maintenance Fee Events
Sep 26 2022BIG: Entity status set to Undiscounted (note the period is included in the code).


Date Maintenance Schedule
Jan 02 20274 years fee payment window open
Jul 02 20276 months grace period start (w surcharge)
Jan 02 2028patent expiry (for year 4)
Jan 02 20302 years to revive unintentionally abandoned end. (for year 4)
Jan 02 20318 years fee payment window open
Jul 02 20316 months grace period start (w surcharge)
Jan 02 2032patent expiry (for year 8)
Jan 02 20342 years to revive unintentionally abandoned end. (for year 8)
Jan 02 203512 years fee payment window open
Jul 02 20356 months grace period start (w surcharge)
Jan 02 2036patent expiry (for year 12)
Jan 02 20382 years to revive unintentionally abandoned end. (for year 12)