A method comprising determining an input axis that extends between an apparatus and a separate apparatus, receiving information indicative of an input object, the input object being proximate to the input axis, determining an input axis position on the input axis based, at least in part, on the input object, and determining an operational directive based, at least in part, on the input axis position is disclosed.
|
1. A method comprising:
determining an input axis that extends between a point associated with a sensor of an apparatus and a point associated with a separate apparatus;
receiving information indicative of an input object, the input object being within a threshold distance of the input axis;
determining an input position relative to the input axis extending between the point associated with the sensor of the apparatus and the point associated with the separate apparatus based, at least in part, on the input object and data detected with at least the sensor;
causing performance of an operational directive based, at least in part, on the input position relative to the input axis;
receiving different information indicative of the input object, the input object failing to be within the threshold distance of the input axis; and
in response to determining that the input object fails to be within the threshold distance of the input axis, precluding performance of the operational directive.
9. An apparatus comprising at least one processor and at least one memory, the memory comprising machine-readable instructions, that when executed cause the apparatus to:
determine an input axis that extends between a point associated with a sensor of a device and a point associated with a separate apparatus;
receive information indicative of an input object, the input object being within a threshold distance of the input axis;
determine an input position relative to the input axis extending between the point associated with the sensor of the device and the point associated with the separate apparatus based, at least in part, on the input object and data detected with at least the sensor;
cause performance of an operational directive based, at least in part, on the input position relative to the input axis;
receive different information indicative of the input object, the input object failing to be within the threshold distance of the input axis; and
in response to determining that the input object fails to be within the threshold distance of the input axis, preclude performance of the operational directive.
2. The method of
3. The method of
4. The method of
5. The method of
receiving information indicative of movement of the input object, the input object remaining within the threshold distance of the input axis;
determining a different input position relative to the input axis based, at least in part, on the input object; and
determining a different operational directive based, at least in part, on the different input position.
6. The method of
determining an apparatus authentication parameter based, at least in part, on the input position;
sending an authentication request to the separate apparatus;
receiving an authentication response from the separate apparatus that is indicative of a separate apparatus authentication parameter;
determining that the separate apparatus authentication parameter corresponds with the apparatus authentication parameter; and
causing establishment of a pairing association between the apparatus and the separate apparatus.
7. The method of
receiving information indicative of movement of the input object, the input object remaining within the threshold distance of the input axis;
determining a different input position on the input axis based, at least in part, on the input object; and
determining that the different input position is further from the apparatus than the input position, wherein determining the operational directive is based, at least in part, on determining that the different input position is further from the apparatus than the input position.
8. The method of
determining an apparatus authentication parameter based, at least in part, on the input position; and
sending an authentication response to the separate apparatus that is indicative of the apparatus authentication parameter.
10. The apparatus of
11. The apparatus of
12. The apparatus of
perform one or more operation in conformance with the operational directive.
13. The apparatus of
receive information indicative of movement of the input object, the input object remaining within the threshold distance of the input axis;
determine a different input position relative to the input axis based, at least in part, on the input object; and
determine a different operational directive based, at least in part, on the different input position.
14. The apparatus of
determine an apparatus authentication parameter based, at least in part, on the input position;
send an authentication request to the separate apparatus;
receive an authentication response from the separate apparatus that is indicative of a separate apparatus authentication parameter;
determine that the separate apparatus authentication parameter corresponds with the apparatus authentication parameter; and
cause establishment of a pairing association between the device and the separate apparatus.
15. The apparatus of
receive information indicative of movement of the input object, the input object remaining within the threshold distance of the input axis;
determine a different input position on the input axis based, at least in part, on the input object; and
determine that the different input position is further from the device than the input position, wherein determination the operational directive is based, at least in part, on determination that the different input position is further from the device than the input position.
16. The apparatus of
17. The method of
18. The method of
|
This application was originally filed as Patent Cooperation Treaty Application No. PCT/FI2015/050119 filed Feb. 27, 2015 which claims priority benefit to European Patent Application No. 14157426.9, Mar. 3, 2014.
The present application relates generally to an input axis between an apparatus and a separate apparatus.
Electronic apparatuses are becoming increasingly accessible and increasingly powerful. Users of electronic apparatus utilize electronic apparatuses in relation to many aspects of their lives. As such, users are constantly interacting with their electronic apparatuses. In some circumstances, it may be desirable to configure an electronic apparatus such that a user of the electronic apparatus may interact with the electronic apparatus.
Various aspects of examples of the invention are set out in the claims.
One or more embodiments may provide an apparatus, a computer readable medium, a non-transitory computer readable medium, a computer program product, and a method for determining an input axis that extends between an apparatus and a separate apparatus, receiving information indicative of an input object, the input object being proximate to the input axis, determining an input axis position on the input axis based, at least in part, on the input object, and determining an operational directive based, at least in part, on the input axis position.
One or more embodiments may provide an apparatus, a computer readable medium, a computer program product, and a non-transitory computer readable medium having means for determining an input axis that extends between an apparatus and a separate apparatus, means for receiving information indicative of an input object, the input object being proximate to the input axis, means for determining an input axis position on the input axis based, at least in part, on the input object, and means for determining an operational directive based, at least in part, on the input axis position.
An apparatus comprising at least one processor and at least one memory, the memory comprising machine-readable instructions, that when executed cause the apparatus to perform determination of an input axis that extends between an apparatus and a separate apparatus, receipt of information indicative of an input object, the input object being proximate to the input axis, determination of an input axis position on the input axis based, at least in part, on the input object, and determination of an operational directive based, at least in part, on the input axis position.
In at least one example embodiment, the operational directive corresponds with an operational directive type.
In at least one example embodiment, the operational directive type corresponds with a parameter adjustment operational directive that causes a parameter to be set to a value.
In at least one example embodiment, the value is based, at least in part, on the input axis position.
In at least one example embodiment, the parameter is a video playback position parameter, and the value is a video playback position.
In at least one example embodiment, the parameter is a video scene parameter, and the value is a video scene.
In at least one example embodiment, the parameter is an audio playback position parameter, and the value is an audio playback position.
In at least one example embodiment, the parameter is an audio item parameter, and the value is an audio item.
In at least one example embodiment, the parameter is a game input position parameter, and the value is a game input position.
In at least one example embodiment, the parameter is a tag parameter, and the value is a tag.
In at least one example embodiment, the parameter is an output volume parameter, and the value is an output volume.
In at least one example embodiment, the parameter is an input volume parameter, and the value is an input volume.
In at least one example embodiment, the parameter is a file transfer parameter, and the value is indicative of a file.
In at least one example embodiment, the parameter is an instrument playback parameter, and the value is a pitch.
In at least one example embodiment, the operational directive type corresponds with a separate apparatus authentication operational directive that causes establishment of a pairing association between the apparatus and the separate apparatus.
In at least one example embodiment, the input axis is a cross apparatus delineation that extends between a point on the apparatus and a point on the separate apparatus.
In at least one example embodiment, the input object is an object having a predetermined characteristic that indicates that the object is intended to invoke the determination of the operational directive.
In at least one example embodiment, the predetermined characteristic is a characteristic that is identifiable by at least one sensor that detects the input object.
In at least one example embodiment, the input object is at least one of a finger, a single finger, two fingers, a hand, an arm, a stylus, or a wearable apparatus.
In at least one example embodiment, the input object being proximate to the input axis corresponds with the input object being within a threshold distance from the input axis.
In at least one example embodiment, the threshold distance is a distance from the input axis within which the input object is determined to be proximate to the input axis.
One or more example embodiments further perform determination that the input object is proximate to the input axis, wherein the determination of the position on the input axis is based, at least in part, on the determination that the input object is proximate to the input axis.
One or more example embodiments further perform receipt of different information indicative of the input object, the input object failing to be proximate to the input axis, determination that the input object fails to be proximate to the input axis, and preclusion of determination of a different position on the input axis based, at least in part, on the determination that the input object fails to be proximate to the input axis.
In at least one example embodiment, the determination of the input axis position on the input axis comprises determination of an input object correspondence delineation perpendicular to the input axis that intersects at least a part of the input object and determination of the input axis position to be a position on the input axis where the input object correspondence delineation intersects the input axis.
In at least one example embodiment, the determination of the input axis position on the input axis comprises determination of the input axis position to be a position on the input axis where the input object intersects the input axis.
In at least one example embodiment, the determination of the operational directive comprises correlation of the input axis position with the operational directive.
In at least one example embodiment, the determination of the operational directive comprises correlation of the input axis position with a predetermined input axis position that is associated with the operational directive.
In at least one example embodiment, the determination of the operational directive comprises correlation of the input axis position with a predetermined input axis position range that is associated with the operational directive.
In at least one example embodiment, the operational directive is a directive that instructs the apparatus to perform a particular set of operations that are associated with the directive.
In at least one example embodiment, the operational directive identifies at least one operation of the set of operations.
One or more example embodiments further perform causation of performance of one or more operation in conformance with the operational directive.
In at least one example embodiment, the causation of performance of the operation in conformance with the operational directive comprises performance of the operation in conformance with the operational directive.
In at least one example embodiment, the causation of performance of the operation in conformance with the operational directive comprises sending of the operational directive to a separate apparatus such that the separate apparatus is caused to perform the operation in conformance with the operational directive.
One or more example embodiments further perform determination of an input object classification based, at least in part, on the information indicative of the input object, wherein the determination of the operational directive is based, at least in part, on the input object classification.
In at least one example embodiment, the input object classification identifies a characteristic of the input object.
In at least one example embodiment, the input object classification is indicative of at least one of a finger, a single finger, two fingers, a hand, an arm, a stylus, or a wearable apparatus.
In at least one example embodiment, the input axis position is indicative of a position relative to a distance between the apparatus and the separate apparatus.
In at least one example embodiment, the input axis position is indicative of a percentage of the distance between the apparatus and the separate apparatus.
In at least one example embodiment, the input axis position is indicative of a percentage of the distance from the apparatus to separate apparatus.
In at least one example embodiment, the input axis position is indicative of a percentage of the distance from the separate apparatus to apparatus.
One or more example embodiments further perform determination of the distance between the apparatus and the separate apparatus.
In at least one example embodiment, the input axis position corresponds with an input axis segment, and the determination of the operational directive is based, at least in part, on the input axis segment.
In at least one example embodiment, the input axis segment identifies a range of input axis positions that correspond with a particular operational directive.
In at least one example embodiment, the input axis position corresponds with an input axis partition, and the determination of the operational directive is based, at least in part, on the input axis partition.
In at least one example embodiment, the input axis partition identifies a range of input axis positions that correspond with an operational directive type.
One or more example embodiments further perform receipt of information indicative of movement of the input object, the input object remaining proximate to the input axis, determination of a different input axis position on the input axis based, at least in part, on the input object, and determination of a different operational directive based, at least in part, on the different input axis position.
In at least one example embodiment, the input axis position corresponds with an input axis segment and the different input axis position corresponds with a different input axis segment, the determination of the operational directive is based, at least in part, on the input axis segment, and the determination of the different operational directive is based, at least in part, on the different input axis segment.
In at least one example embodiment, the input axis position corresponds with an input axis segment and the different input axis position corresponds with the input axis segment, the determination of the operational directive is based, at least in part, on the input axis segment, and the determination of the different operational directive is based, at least in part, on the input axis segment.
In at least one example embodiment, the input axis position corresponds with an input axis partition and the different input axis position corresponds with a different input axis partition, the determination of the operational directive is based, at least in part, on the input axis partition, and the determination of the different operational directive is based, at least in part, on the different input axis partition.
In at least one example embodiment, the input axis position corresponds with an input axis partition and the different input axis position corresponds with the input axis partition, the determination of the operational directive is based, at least in part, on the input axis partition, and the determination of the different operational directive is based, at least in part, on the input axis partition.
In at least one example embodiment, the operational directive causes the apparatus to perform determination of an apparatus authentication parameter based, at least in part, on the input axis position, sending of an authentication request to the separate apparatus, receipt of an authentication response from the separate apparatus that is indicative of a separate apparatus authentication parameter, determination that the separate apparatus authentication parameter corresponds with the apparatus authentication parameter, and causation of an establishment of a pairing association between the apparatus and the separate apparatus.
One or more example embodiments further perform receipt of information indicative of movement of the input object, the input object remaining proximate to the input axis, determination of a different input axis position on the input axis based, at least in part, on the input object, and determination that the different input axis position is further from the apparatus than the input axis position, wherein the determination of the operational directive is based, at least in part, on the determination that the different input axis position is further from the apparatus than the input axis position.
In at least one example embodiment, the operational directive is a separate apparatus authentication directive.
One or more example embodiments further perform receipt of an authentication request from the separate apparatus, wherein the operational directive causes the apparatus to perform determination of an apparatus authentication parameter based, at least in part, on the input axis position, and sending of an authentication response to the separate apparatus that is indicative of the apparatus authentication parameter.
One or more example embodiments further perform causation of display of a slider interface element by the separate apparatus, wherein the operational directive causes setting of a slider interface element selection point to correspond with the input axis position.
In at least one example embodiment, the slider interface element is associated with a parameter, and the slider interface element selection point is associated with a value of the parameter.
One or more example embodiments further perform causation of display of the slider interface element selection point in relation to the slider interface element.
For a more complete understanding of embodiments of the invention, reference is now made to the following descriptions taken in connection with the accompanying drawings in which:
An embodiment of the invention and its potential advantages are understood by referring to
Some embodiments will now be described more fully hereinafter with reference to the accompanying drawings, in which some, but not all, embodiments are shown. Various embodiments of the invention may be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure will satisfy applicable legal requirements. Like reference numerals refer to like elements throughout. As used herein, the terms “data,” “content,” “information,” and similar terms may be used interchangeably to refer to data capable of being transmitted, received and/or stored in accordance with embodiments of the present invention. Thus, use of any such terms should not be taken to limit the spirit and scope of embodiments of the present invention.
Additionally, as used herein, the term ‘circuitry’ refers to (a) hardware-only circuit implementations (e.g., implementations in analog circuitry and/or digital circuitry); (b) combinations of circuits and computer program product(s) comprising software and/or firmware instructions stored on one or more computer readable memories that work together to cause an apparatus to perform one or more functions described herein; and (c) circuits, such as, for example, a microprocessor(s) or a portion of a microprocessor(s), that require software or firmware for operation even if the software or firmware is not physically present. This definition of ‘circuitry’ applies to all uses of this term herein, including in any claims. As a further example, as used herein, the term ‘circuitry’ also includes an implementation comprising one or more processors and/or portion(s) thereof and accompanying software and/or firmware. As another example, the term ‘circuitry’ as used herein also includes, for example, a baseband integrated circuit or applications processor integrated circuit for a mobile phone or a similar integrated circuit in a server, a cellular network apparatus, other network apparatus, and/or other computing apparatus.
As defined herein, a “non-transitory computer-readable medium,” which refers to a physical medium (e.g., volatile or non-volatile memory device), can be differentiated from a “transitory computer-readable medium,” which refers to an electromagnetic signal.
Furthermore, apparatuses may readily employ embodiments of the invention regardless of their intent to provide mobility. In this regard, even though embodiments of the invention may be described in conjunction with mobile applications, it should be understood that embodiments of the invention may be utilized in conjunction with a variety of other applications, both in the mobile communications industries and outside of the mobile communications industries. For example, the apparatus may be, at least part of, a non-carryable apparatus, such as a large screen television, an electronic table, a kiosk, an automobile, and/or the like.
In at least one example embodiment, electronic apparatus 10 comprises processor 11 and memory 12. Processor 11 may be any type of processor, controller, embedded controller, processor core, and/or the like. In at least one example embodiment, processor 11 utilizes computer program code to cause an apparatus to perform one or more actions. Memory 12 may comprise volatile memory, such as volatile Random Access Memory (RAM) including a cache area for the temporary storage of data and/or other memory, for example, non-volatile memory, which may be embedded and/or may be removable. The non-volatile memory may comprise an EEPROM, flash memory and/or the like. Memory 12 may store any of a number of pieces of information, and data. The information and data may be used by the electronic apparatus 10 to implement one or more functions of the electronic apparatus 10, such as the functions described herein. In at least one example embodiment, memory 12 includes computer program code such that the memory and the computer program code are configured to, working with the processor, cause the apparatus to perform one or more actions described herein.
The electronic apparatus 10 may further comprise a communication device 15. In at least one example embodiment, communication device 15 comprises an antenna, (or multiple antennae), a wired connector, and/or the like in operable communication with a transmitter and/or a receiver. In at least one example embodiment, processor 11 provides signals to a transmitter and/or receives signals from a receiver. The signals may comprise signaling information in accordance with a communications interface standard, user speech, received data, user generated data, and/or the like. Communication device 15 may operate with one or more air interface standards, communication protocols, modulation types, and access types. By way of illustration, the electronic communication device 15 may operate in accordance with second-generation (2G) wireless communication protocols IS-136 (time division multiple access (TDMA)), Global System for Mobile communications (GSM), and IS-95 (code division multiple access (CDMA)), with third-generation (3G) wireless communication protocols, such as Universal Mobile Telecommunications System (UMTS), CDMA2000, wideband CDMA (WCDMA) and time division-synchronous CDMA (TD-SCDMA), and/or with fourth-generation (4G) wireless communication protocols, wireless networking protocols, such as 802.11, short-range wireless protocols, such as Bluetooth, and/or the like. Communication device 15 may operate in accordance with wireline protocols, such as Ethernet, digital subscriber line (DSL), asynchronous transfer mode (ATM), and/or the like.
Processor 11 may comprise means, such as circuitry, for implementing audio, video, communication, navigation, logic functions, and/or the like, as well as for implementing embodiments of the invention including, for example, one or more of the functions described herein. For example, processor 11 may comprise means, such as a digital signal processor device, a microprocessor device, various analog to digital converters, digital to analog converters, processing circuitry and other support circuits, for performing various functions including, for example, one or more of the functions described herein. The apparatus may perform control and signal processing functions of the electronic apparatus 10 among these devices according to their respective capabilities. The processor 11 thus may comprise the functionality to encode and interleave message and data prior to modulation and transmission. The processor 1 may additionally comprise an internal voice coder, and may comprise an internal data modem. Further, the processor 11 may comprise functionality to operate one or more software programs, which may be stored in memory and which may, among other things, cause the processor 11 to implement at least one embodiment including, for example, one or more of the functions described herein. For example, the processor 11 may operate a connectivity program, such as a conventional internet browser. The connectivity program may allow the electronic apparatus 10 to transmit and receive internet content, such as location-based content and/or other web page content, according to a Transmission Control Protocol (TCP), Internet Protocol (IP), User Datagram Protocol (UDP), Internet Message Access Protocol (IMAP), Post Office Protocol (POP), Simple Mail Transfer Protocol (SMTP), Wireless Application Protocol (WAP), Hypertext Transfer Protocol (HTTP), and/or the like, for example.
The electronic apparatus 10 may comprise a user interface for providing output and/or receiving input. The electronic apparatus 10 may comprise an output device 14. Output device 14 may comprise an audio output device, such as a ringer, an earphone, a speaker, and/or the like. Output device 14 may comprise a tactile output device, such as a vibration transducer, an electronically deformable surface, an electronically deformable structure, and/or the like. Output device 14 may comprise a visual output device, such as a display, a light, and/or the like. In at least one example embodiment, the apparatus causes display of information, the causation of display may comprise displaying the information on a display comprised by the apparatus, sending the information to a separate apparatus that comprises a display, and/or the like. The electronic apparatus may comprise an input device 13. Input device 13 may comprise a light sensor, a proximity sensor, a microphone, a touch sensor, a force sensor, a button, a keypad, a motion sensor, a magnetic field sensor, a camera, and/or the like. A touch sensor and a display may be characterized as a touch display. In an embodiment comprising a touch display, the touch display may be configured to receive input from a single point of contact, multiple points of contact, and/or the like. In such an embodiment, the touch display and/or the processor may determine input based, at least in part, on position, motion, speed, contact area, and/or the like. In at least one example embodiment, the apparatus receives an indication of an input. The apparatus may receive the indication from a sensor, a driver, a separate apparatus, and/or the like. The information indicative of the input may comprise information that conveys information indicative of the input, indicative of an aspect of the input indicative of occurrence of the input, and/or the like.
The electronic apparatus 10 may include any of a variety of touch displays including those that are configured to enable touch recognition by any of resistive, capacitive, infrared, strain gauge, surface wave, optical imaging, dispersive signal technology, acoustic pulse recognition or other techniques, and to then provide signals indicative of the location and other parameters associated with the touch. Additionally, the touch display may be configured to receive an indication of an input in the form of a touch event which may be defined as an actual physical contact between a selection object (e.g., a finger, stylus, pen, pencil, or other pointing device) and the touch display. Alternatively, a touch event may be defined as bringing the selection object in proximity to the touch display, hovering over a displayed object or approaching an object within a predefined distance, even though physical contact is not made with the touch display. As such, a touch input may comprise any input that is detected by a touch display including touch events that involve actual physical contact and touch events that do not involve physical contact but that are otherwise detected by the touch display, such as a result of the proximity of the selection object to the touch display. A touch display may be capable of receiving information associated with force applied to the touch screen in relation to the touch input. For example, the touch screen may differentiate between a heavy press touch input and a light press touch input. In at least one example embodiment, a display may display two-dimensional information, three-dimensional information and/or the like.
In embodiments including a keypad, the keypad may comprise numeric (for example, 0-9) keys, symbol keys (for example, #, *), alphabetic keys, and/or the like for operating the electronic apparatus 10. For example, the keypad may comprise a conventional QWERTY keypad arrangement. The keypad may also comprise various soft keys with associated functions. In addition, or alternatively, the electronic apparatus 10 may comprise an interface device such as a joystick or other user input interface.
Input device 13 may comprise a media capturing element. The media capturing element may be any means for capturing an image, video, and/or audio for storage, display or transmission. For example, in at least one example embodiment in which the media capturing element is a camera module, the camera module may comprise a digital camera which may form a digital image file from a captured image. As such, the camera module may comprise hardware, such as a lens or other optical component(s), and/or software necessary for creating a digital image file from a captured image. Alternatively, the camera module may comprise only the hardware for viewing an image, while a memory device of the electronic apparatus 10 stores instructions for execution by the processor 11 in the form of software for creating a digital image file from a captured image. In at least one example embodiment, the camera module may further comprise a processing element such as a co-processor that assists the processor 11 in processing image data and an encoder and/or decoder for compressing and/or decompressing image data. The encoder and/or decoder may encode and/or decode according to a standard format, for example, a Joint Photographic Experts Group (JPEG) standard format.
In the example of
It should be understood that, even though the example of
In some circumstances, a user may desire to have collaboration between apparatuses, such as between an apparatus and a separate apparatus, based on their proximity with each other. For example, it may be intuitive for a user to manage collaboration between apparatuses that are local to each other. A plurality of apparatuses may be proximate to each other based on location, availability of local communication among the apparatuses, and/or the like. For example, if the apparatuses collaborate by way of low power radio frequency communication, a radio frequency communication, near field communication, inductive communication, electric field communication, Bluetooth communication, infrared communication, local area network communication, wireless local area network communication, and/or the like, the apparatuses may be considered to be proximate with each other based, at least in part, on availability of such proximity-based communication with each other. In at least one example embodiment, an apparatus may be a phone, a tablet, a computer, a wearable apparatus, a head worn apparatus, a hand worn apparatus, an electronic apparatus, a peripheral apparatus, a host apparatus, and/or the like. In at least one example embodiment, apparatuses communicate with each other. For example, an apparatus may be an apparatus that automatically communicates with another apparatus for purposes such as identifying the apparatus, synchronizing data, exchanging status information, and/or the like. In at least one example embodiment, an apparatus retains information associated with communication with a separate apparatus. For example, the apparatus may comprise information associated with identifying, communicating with, authenticating, performing authentication with, and/or the like, the separate apparatus. In this manner, the apparatus may be privileged to perform operations in conjunction with the separate apparatus that a different apparatus may lack the privilege to perform.
In at least one example embodiment, communication based, at least in part, on short range communication is referred to as proximity-based communication. In at least one example embodiment, proximity-based communication relates to wireless communication that is associated with a short range, such as low power radio frequency communication, radio frequency communication, near field communication, inductive communication, electric field communication, Bluetooth communication, infrared communication, local area network communication, wireless local area network communication, and/or the like. In such an example, the exchange of information may be by way of the short range wireless communication between the apparatus and a separate apparatus, host apparatus, and/or the like.
In at least one example embodiment, a proximity-based communication channel is a low power radio frequency communication channel, a radio frequency communication channel, a near field communication channel, a wireless communication channel, a wireless local area network communication channel, a Bluetooth communication channel, an electric field communication channel, an inductive communication channel, an infrared communication channel, and/or the like. For example, as depicted in
In at least one example embodiment, an apparatus and a separate apparatus communicate by way of non-proximity-based communication channels. For example, as depicted in
In many circumstances, a user of an electronic apparatus interacts with the electronic apparatus. For example, the user may listen to music on the electronic apparatus, watch a movie by way of the electronic apparatus, play games on the electronic apparatus, share one or more files with another electronic apparatus by way of the electronic apparatus, and/or the like. In some circumstances, a user interacts with an electronic apparatus by way of a touch sensitive display. In such circumstances, the touch sensitive display is often constrained in size, constrained in granularity of input control, and/or the like. As such, it may be desirable to configure an electronic apparatus such that a user of the electronic apparatus may interact with the electronic apparatus in a manner that allows for an increased granularity of input control, that avoids obscuration of a display of the electronic apparatus by a hand or a finger of the user, and/or the like.
In many circumstances, a user may commonly utilize more than one electronic apparatus simultaneously. For example, the user may utilize an electronic apparatus, for example a phone, a computer, a tablet, etc., and a different electronic apparatus, for example a head mounted apparatus, another phone, a tablet, etc., simultaneously, in conjunction with one another, and/or the like. In such an example, the electronic apparatus and the different electronic apparatus may be operatively connected, the electronic apparatus may be in communication with the different electronic apparatus by way of at least one communication channel, and/or the like. For example, a user of the electronic apparatus may render music by way of the different electronic apparatus. As such, in such circumstances, it may be desirable to configure an electronic apparatus that is associated with another electronic apparatus such that a user of the electronic apparatus may interact with the electronic apparatus in a manner that allows for an increased granularity of input control, that avoids obscuration of a display of the electronic apparatus by a hand or a finger of the user, and/or the like. In at least one example embodiment, an apparatus determines input axis that extends between an apparatus and a separate apparatus. The input axis may be a cross apparatus delineation that extends between a point on the apparatus and a point on the separate apparatus. The point on the apparatus may be a sensor, a camera module, and/or the like. The point on the separate apparatus may be a marker, a tag, and/or the like. As such, the sensor, the camera module, and/or the like of the apparatus may detect, track, etc. the marker, the tag, and/or the like on the separate apparatus.
In order to facilitate a user interacting with an electronic apparatus by way of an input axis associated with the electronic apparatus, it may be desirable to detect an input object that is associated with the input axis. In at least one example embodiment, an apparatus receives information indicative of an input object. The input object may be an object that has a predetermined characteristic that indicates that the object is intended to interact with the apparatus by way of the input axis. For example, the input object may be intended to invoke a determination of an operational directive associated with the apparatus, to manipulate one or more parameters associated with the apparatus by way of the operational directive, and/or the like. The operational directive may be similar as described regarding
In many circumstances, it may be desirable to infer a user intent to interact with an electronic apparatus by way of an input axis based, at least in part, on an input object being proximate to the input axis. In at least one example embodiment, an apparatus receives information indicative of an input object, the input object being proximate to the input axis. The input object being proximate to the input axis may correspond with the input object being within a threshold distance from the input axis. In at least one example embodiment, the threshold distance is a distance from the input axis within which the input object is determined to be proximate to the input axis. In at least one example embodiment, the threshold distance is based, at least in part, on a detection region of a sensor, a capture region of a camera module, and/or the like. In at least one example embodiment, an apparatus determines that the input object is proximate to the input axis. For example, the apparatus may determine that the input object intersects the input axis, the input object is within a threshold distance of the input axis, the input object is within a detection region of a sensor, the input object is within a capture region of a camera module, and/or the like.
For example, as depicted in the example of
In many circumstances, a user may desire to interact with an electronic apparatus by way of an input axis that extends from the electronic apparatus to another electronic apparatus, may desire to interact with the other electronic apparatus by way of the input axis, and/or the like. As such, it may be desirable to allow for quantification of user input that is associated with the input axis. In at least one example embodiment, an apparatus determines an input axis position on the input axis. The determination of the input axis position may be based, at least in part, on the input object, a position of the input object in relation to the input axis, and/or the like. The input axis position may be indicative of a position relative to a distance between the apparatus and the separate apparatus, a percentage of the distance between the apparatus and the separate apparatus, a percentage of the distance from the apparatus to separate apparatus, a percentage of the distance from the separate apparatus to apparatus, and/or the like.
In order to facilitate determination of the input axis position, it may be desirable to determine one or more distances associated with an apparatus, a separate apparatus, an input object, and/or the like. In at least one example embodiment, an apparatus determines a distance between an apparatus and a separate apparatus. In at least one example embodiment, the apparatus determines a distance between an input object and the apparatus. For example, the apparatus may determine a distance from an apparatus to the input object, a distance from the input object to the separate apparatus, and/or the like. As such, the apparatus may determine the input axis position based, at least in part, on the distance between the apparatus and the separate apparatus, the distance between the apparatus and the input object, the distance between the separate apparatus and the input object, and/or the like. For example, as illustrated in the example of
In many circumstances, it may be desirable to determine an input axis position when the input object is proximate to the input axis, and to preclude determination of an input axis position when the input object fails to be proximate to the input axis. For example, a user of the apparatus may desire to avoid unintentional interaction with the apparatus, to preclude inadvertent input associated with the input axis, and/or the like. In at least one example embodiment, an apparatus determines an input axis position of an input object on the input axis. The determination of the input axis position may be based, at least in part, on the determination that the input object is proximate to the input axis. In some circumstances, the input object may intersect the input axis. As such, the determination of the input axis position on the input axis may comprise determination of the input axis position to be a position on the input axis where the input object intersects the input axis. For example, as illustrated in the example of
In some circumstances, the input object may be proximate to the input axis, but may fail to intersect the input axis. For example, the input object may be within a threshold distance from the input axis, the input object may be within a detection region of a sensor, the input object may be within a capture region of a camera module, and/or the like. As such, the determination of the input axis position on the input axis may comprise determination of an input object correspondence delineation perpendicular to the input axis. The input object correspondence delineation may intersect at least a part of the input object, and may intersect the input axis at an input axis position. As such, the apparatus may determine the input axis position to be a position on the input axis where the input object correspondence delineation intersects the input axis.
In the example of
As discussed previously, in many circumstances, it may be desirable to preclude determination of an input axis position in circumstances where the input object fails to be proximate to the input axis. For example, a user of the apparatus may desire to avoid unintentional interaction with the apparatus, to preclude inadvertent input associated with the input axis, and/or the like. In at least one example embodiment, an apparatus determines that an input object fails to be proximate to an input axis. For example, the input object may be outside of a threshold distance from the input axis, may be outside of a detection region of a sensor, may be outside of a capture region of a camera module, and/or the like. In such an example embodiment, the apparatus may preclude determination of a position on the input axis based, at least in part, on the determination that the input object fails to be proximate to the input axis.
As discussed previously, in many circumstances, a user of an electronic apparatus may desire to interact with the electronic apparatus by way of an input axis associated with the electronic apparatus, another electronic apparatus, and/or the like. In order to facilitate such interaction, it may be desirable to configure the electronic apparatus such that the electronic apparatus may perform one or more operations based, at least in part, on input received from a user by way of the input axis. For example, the user may desire to scroll through content, may desire to manipulate one or more parameters associated with the apparatus, and/or the like. In at least one example embodiment, an apparatus determines an operational directive based, at least in part, on the input axis position.
It should be understood that determination of an operational directive may be performed in various manners, such as a lookup table, a comparison, and/or the like. For example, the apparatus may determine the operational directive by way of correlation of the input axis position with the operational directive. In such an example, a particular operational directive may be invoked when the apparatus determines an input axis position that corresponds with a specific input axis position, a range of input axis positions, and/or the like. In at least one example embodiment, the apparatus determines the operational directive based, at least in part, on correlation of the input axis position with a predetermined input axis position that is associated with the operational directive. In at least one example embodiment, the apparatus determines the operational directive based, at least in part, on correlation of the input axis position with a predetermined input axis position range that is associated with the operational directive. For example, the input axis position may be within an input axis position threshold from a predetermined input axis position that is associated with an operation directive, may be within a range of input axis positions that is associated with an operational directive, and/or the like.
In many circumstances, the user may desire the electronic apparatus to perform one or more operations in response to the user's input. For example, the user may desire to cause the apparatus to scroll through a list of content, to adjust a parameter associated with the apparatus, and/or the like. As such, it may be desirable to configure an electronic apparatus such that the electronic apparatus performs one or more operations in conformance with the operational directive. In at least one example embodiment, an operational directive is a directive that instructs the apparatus to perform a particular set of operations that are associated with the directive. For example, the operational directive may identify at least one operation of the set of operations, and the apparatus may perform the operation in conformance with the directive. In at least one example embodiment, an apparatus causes performance of one or more operations in conformance with the operational directive. The causation of performance of the operation in conformance with the operational directive may comprise performance of the operation in conformance with the operational directive, sending of the operational directive to a separate apparatus such that the separate apparatus is caused to perform the operation in conformance with the operational directive, and/or the like.
In some circumstances, a user may desire to interact with an electronic apparatus in different manners. For example, the user may desire to utilize the user's hand as an input object, to utilize a stylus as the input object, to utilize a single finger as the input object, to utilize three fingers as the input object, and/or the like. In such circumstances, it may be desirable to determine an operational directive based, at least in part, on the type of the input object, a characteristic of the input object, and/or the like. For example, the user may desire to interact with the apparatus in a particular manner based, at least in part, on the input object utilized by the user. In at least one example embodiment, an apparatus determines an input object classification of an input object. The determination of the input object classification may be based, at least in part, on the received information indicative of the input object. In such an example embodiment, the determination of the operational directive may be based, at least in part, on the input object classification. The input object classification may identify a characteristic of the input object. Such a characteristic of the input object may be a characteristic that is identifiable by way of a sensor, a camera module, and/or the like. The characteristic of the input object may be particularly suited for identification by a specific type of sensor, for example an infrared sensor or a radar sensor, or by a particular type of camera module. The input object classification may be indicative of a finger, a single finger, two fingers, a hand, an arm, a stylus, a wearable apparatus, and/or the like.
In many circumstances, a user may desire to interact with an electronic apparatus in various context associated with utilization of the electronic apparatus. For example, the user may be utilizing the electronic apparatus to play music, to watch a video, to view textual information, and/or the like. In at least one example embodiment, an operational directive corresponds with an operational directive type. In at least one example embodiment, the operational directive type corresponds with a parameter adjustment operational directive that causes a parameter to be set to a value. In such an example embodiment, the value may be based, at least in part, on the input axis position. For example, the input axis may represent a range of values for a specific parameter, and an input axis position associated with the input axis may indicate a particular value within the range of values.
In some circumstances, a user may desire to interact with and/or manipulate playback of audio content by way of an input axis. In at least one example embodiment, a parameter is an audio playback position parameter, and the value indicated by the input axis position is an audio playback position. For example, the user may desire to fast forward through a song, may desire to rewind a song, may desire to select a specific playback position within the song, and/or the like. In such an example, the input axis may correlate with the duration of the song that is being played. As such, a specific input axis position may indicate with a specific audio playback position within the song. For example, an input axis position that is halfway along the input axis may indicate an audio playback position associated with the middle of the song.
In order to facilitate interaction with an electronic apparatus by way of an input axis, a user of the electronic apparatus may desire to perceive information indicative of the input received by the electronic apparatus. For example, the user may desire to perceive confirmation regarding the user's interactions, may desire to perceive manipulation of content displayed on a display comprised by the electronic apparatus, and/or the like. In at least one example embodiment, an apparatus causes display of a slider interface element. For example, the apparatus may display the slider interface element, may cause a separate apparatus to display the slider interface element, and/or the like. In such an example embodiment, an operational directive may cause setting of a slider interface element selection point to correspond with the input axis position. For example, the apparatus may cause display of a slider interface element selection point in relation to the slider interface element. In such an example, the slider interface element may be associated with a parameter, and the slider interface element selection point may be associated with a value of the parameter. For example, the slider interface element selection point may be indicative of the value of the parameter, may be displayed at a position along the slider interface element based, at least in part, on the value of the parameter, may be displayed at a position along the slider interface element that is indicative of the value of the parameter, and/or the like.
In some circumstances, a user may desire to select specific audio content by way of an input axis. In at least one example embodiment, a parameter is an audio item parameter, and the value indicated by the input axis position is indicative of an audio item. For example, the user may interact with a playlist of songs by way of the input axis. As such, the user may scroll through the playlist by way of the input axis, may select a specific song by way of the input axis, and/or the like. As such, a specific input axis position may indicate a specific song that the user desires to cause rendering of by way of the apparatus, a separate apparatus, and/or the like.
In some circumstances, a user may desire to interact with one or more interface element associated with an electronic apparatus. For example, the user may desire to play a game by way of the electronic apparatus, to interact with an application that is associated with the electronic apparatus, and/or the like. In such an example, the user may interact with and/or manipulate one or more interface elements associated with a game, an application, and or the like. In at least one example embodiment, a parameter is a game input position parameter, and the value indicated by the input axis position is a game input position. For example, the game may be associated with a sling-shot mechanism, and the input axis position may indicate a position of the sling. For example, an input axis position that is closer to a separate apparatus may indicate a lesser straining of the sling, an input axis position that is further from the separate apparatus may indicate a greater straining of the sling, and/or the like. In at least one example embodiment, a parameter is an instrument playback parameter, and the value indicated by the input axis position is a pitch associated with playing of the instrument. For example, the instrument may be a harp, and the input axis position may indicate playing of a particular string associated with the harp. For example, an input axis position that is closer to a separate apparatus may indicate a pitch associated with a shorter string of the harp, an input axis position that is further from the separate apparatus may indicate a pitch associated with a longer string of the harp, and/or the like.
In some circumstances, a user may desire to control one or more additional parameters associated with an electronic apparatus. For example, the user may desire to increase an output volume associated with playback of auditory information, may desire to decrease an input volume associated with auditory information received from a microphone, and/or the like. In at least one example embodiment, a parameter is an output volume parameter, and the value indicated by the input axis position is an output volume. For example, an input axis position that is closer to a separate apparatus may indicate a higher output volume, an input axis position that is further from the separate apparatus may indicate a lower output volume, and/or the like. In at least one example, a parameter is an input volume parameter, and the value indicated by the input axis position is an input volume. The input volume may be associated with amplification of auditory information received from at least one microphone, gain associated with the auditory information received from the microphone, and/or the like. For example, an input axis position that is closer to a separate apparatus may indicate a higher input volume, an input axis position that is further from the separate apparatus may indicate a lower input volume, and/or the like. In at least one example embodiment, a parameter is a tag parameter, and the value indicated by the input axis position is indicative of a tag. A tag may be an identifying indicator, characteristic, descriptor, and/or the like. For example, a user may desire to tag a specific song as a favorite, tag a different song as a song by a particular artist, and/or the like. As such, various positions along the input axis may correspond with various tags. In this manner, receipt of an input at an input axis position may indicate a user's desire to select the tag that corresponds with the input axis position.
In some circumstances, a user may desire to interact with and/or manipulate playback of video content by way of an input axis. In at least one example embodiment, a parameter is a video playback position parameter, and the value indicated by the input axis position is a video playback position. For example, the user may desire to fast forward through a video, may desire to rewind a video, may desire to select a specific playback position within the video, and/or the like. In such an example, the input axis may correlate with the duration of the video that is being played. As such, a specific input axis position may indicate a specific video playback position within the video. For example, an input axis position that is halfway along the input axis may indicate a video playback position associated with the middle of the video.
In some circumstances, a user may desire to interact with and/or manipulate playback of video content by way of an input axis. In at least one example embodiment, a parameter is a video scene parameter, and the value indicated by the input axis position is indicative of a video scene. For example, the user may desire to cause playback of a particular scene associated with the video, may desire to replay a particular scene associate with the video, may desire to skip ahead within the video and being watching a scene further into the video, and/or the like. In such an example, the input axis position may correspond with an input axis segment. As such, the determination of the operational directive may be based, at least in part, on the input axis segment. In at least one example embodiment, an input axis segment identifies a range of input axis positions that correspond with a particular operational directive. As such, an input axis position may correspond with an input axis segment that is associated with a scene of a video, and a different input axis position may correspond with a different input axis segment that is associated with a different scene of the video. In this manner, the input received at the input axis position may cause rendering of the scene of the video, and the different input received at the different input axis position may cause rendering of the different scene of the video. For example, the apparatus may render the scene of the video, the apparatus may cause a separate apparatus to cause rendering of the scene of the video, and/or the like.
For example, an input axis position may correspond with an input axis segment and a different input axis position may correspond with a different input axis segment. In such an example, the determination of the operational directive may be based, at least in part, on the input axis segment. Similarly, the determination of the different operational directive may be based, at least in part, on the different input axis segment. In another example, an input axis position may correspond with an input axis segment and a different input axis position may correspond with the input axis segment. In such an example, the determination of the operational directive may be based, at least in part, on the input axis segment. Similarly, the determination of the different operational directive may be based, at least in part, on the input axis segment.
In order to permit a user to distinguish between input axis segments, it may be desirable to provide the user with feedback that indicates that a specific input axis position is associated with an input axis segment, the input object has transitioned from an input axis segment to a different input axis segment, and/or the like. In at least one example embodiment, an apparatus provides feedback when transitioning from an input axis segment to a different input axis segment. For example, the apparatus may provide auditory feedback, tactile feedback, visual feedback, and/or the like. For example, the apparatus may cause playing of auditory information when an input object transitions from an input axis segment to a different input axis segment, cause physical rendering of tactile information when an input object transitions from an input axis segment to a different input axis segment, and/or the like.
In the example of
In some circumstances, it may be desirable to allow for more than one type of interaction to be performed by way of an input axis. For example, it may be desirable to configure an electronic apparatus such that the electronic apparatus may distinguish between portions of an input axis, may set a specific parameter based, at least in part, on an input axis position corresponding with a particular portion of the input axis, and/or the like. In such an example, the input axis position may correspond with an input axis partition. As such, the determination of the operational directive may be based, at least in part, on the input axis partition. In at least one example embodiment, an input axis partition identifies a range of input axis positions that correspond with a particular operational directive, with a particular type of operational directive, with a particular parameter, and/or the like. As such, an input axis position may correspond with an input axis partition that is associated with an audio playback position parameter, and a different input axis position may correspond with a different input axis partition that is associated with an output volume parameter. In this manner, the input received at the input axis position may cause setting of the audio playback position parameter to a value indicated by the input axis position, and the different input received at the different input axis position may cause setting of the output volume parameter to a value indicated by the different input axis position.
For example, an input axis position may correspond with an input axis partition and a different input axis position may correspond with a different input axis partition. In such an example, the determination of the operational directive may be based, at least in part, on the input axis partition. Similarly, the determination of the different operational directive may be based, at least in part, on the different input axis partition. In another example, an input axis position may correspond with an input axis partition and a different input axis position may correspond with the input axis partition. In such an example, the determination of the operational directive may be based, at least in part, on the input axis partition. Similarly, the determination of the different operational directive may be based, at least in part, on the input axis partition.
In order to permit a user to distinguish between input axis partitions, it may be desirable to provide the user with feedback that indicates that a specific input axis position is associated with an input axis partition, the input object has transitioned from an input axis partition to a different input axis partition, and/or the like. In at least one example embodiment, an apparatus provides feedback when transitioning from an input axis partition to a different input axis partition. For example, the apparatus may provide auditory feedback, tactile feedback, visual feedback, and/or the like. For example, the apparatus may cause playing of auditory information when an input object transitions from an input axis partition to a different input axis partition, cause physical rendering of tactile information when an input object transitions from an input axis partition to a different input axis partition, and/or the like.
In the example of
In the example of
In many circumstances, users of electronic apparatuses have grown accustomed to certain types of input gestures. For example, many users may be familiar with a swipe gesture, a dragging gesture, a scrolling gesture, and/or the like. As such, it may be desirable to configure an electronic apparatus such that the electronic apparatus may receive information indicative of movement of an input object, may determine that an input object has moved to a different position that corresponds with a different input axis position, and/or the like. In at least one example embodiment, an apparatus receives information indicative of movement of an input object. In such an example embodiment, the apparatus may determine that the input object remains proximate to the input axis while the input object is in motion, after the input object movement has ceased, and/or the like. For example, the apparatus may determine that the input object remains within the threshold distance from the input axis, may determine that the input object movement ceases at a position that is within the threshold distance from the input axis, and/or the like. In such an example embodiment, the apparatus may determine a different input axis position on the input axis based, at least in part, on the input object. As such, the apparatus may determine a different operational directive that is based, at least in part, on the different input axis position.
For example, as depicted in the example of
As discussed previously, many users of electronic apparatuses have grown accustomed to certain types of input gestures. For example, many users may be familiar with a swipe gesture, a dragging gesture, a scrolling gesture, and/or the like. In some circumstances, it may be desirable to configure an electronic apparatus such that the electronic apparatus may receive information indicative of an input object performing a gesture. For example, the input object may perform a specific gesture while remaining proximate to the input axis, and the apparatus may determine the gesture based, at least in part, on an input axis position, a different input axis position, information indicative of movement of the input object, and/or the like. In at least one example embodiment, a gesture corresponds with a predetermined movement pattern of an input object. For example, a gesture may correspond with movement of an input object such that the input object traces a predetermined path, such that the input object is moved in conformance with a predetermined gesture, and/or the like. A gesture may be associated with movement of the input object parallel to an input axis, perpendicular to the input axis, and/or the like. In at least one example embodiment, an apparatus receives information indicative of a gesture performed by way of an input object. In such an example embodiment, the apparatus may determine an operational directive based, at least in part, on the gesture.
In many circumstances, a user may desire to cause transference of files, information, etc. from one apparatus to a separate apparatus. In such circumstances, it may be desirable to configure an apparatus such that the user of the apparatus may indicate such a desire in a manner that is easy and intuitive. In at least one example embodiment, an input axis is associated with one or more files that a user may desire to cause transference of to a separate apparatus. In such an example embodiment, a specific position along the input axis may indicate selection of a particular file to transfer, may indicate selection of a range of files to transfer, and/or the like. For example, an input object may begin at a position associated with an input axis position and subsequently move to a different position associated with a different input axis position. In such an example embodiment, the movement of the input object from the input axis position to the different input axis position may indicate a selection of a portion of the files associated with input axis positions between the input axis position and the different input axis position, a selection of all of the files associated with input axis positions between the input axis position and the different input axis position, and/or the like. In such circumstances, an apparatus may determine that an operational directive is a file transfer operational directive. As such, the apparatus may cause transfer of one or more files from an apparatus to a separate apparatus, from the apparatus to a different separate apparatus, from the separate apparatus to the different separate apparatus, and/or the like, in conformance with the operational directive.
At block 702, the apparatus determines an input axis that extends between an apparatus and a separate apparatus. The determination, the input axis, the apparatus, and the separate apparatus may be similar as described regarding
At block 704, the apparatus receives information indicative of an input object, the input object being proximate to the input axis. The receipt, the input object, and the input object being proximate to the input axis may be similar as described regarding
At block 706, the apparatus determines an input axis position on the input axis based, at least in part, on the input object. The determination and the input axis position may be similar as described regarding
At block 708, the apparatus determines an operational directive based, at least in part, on the input axis position. The determination and the operational directive may be similar as described regarding
As discussed previously, in many circumstances, it may be desirable to determine an input axis position when the input object is proximate to the input axis, and to preclude determination of an input axis position when the input object fails to be proximate to the input axis. For example, a user of the apparatus may desire to avoid unintentional interaction with the apparatus, to preclude inadvertent input associated with the input axis, and/or the like.
At block 802, the apparatus determines an input axis that extends between an apparatus and a separate apparatus. The determination, the input axis, the apparatus, and the separate apparatus may be similar as described regarding
At block 804, the apparatus receives information indicative of an input object. The receipt and the input object may be similar as described regarding
At block 806, the apparatus determines whether the input object is proximate to the input axis. If the apparatus determines that the input object is proximate to the input axis, flow proceeds to block 808. If the apparatus determines that the input object fails to be proximate to the input axis, flow proceeds to block 812. The determination and the input object being proximate to the input axis may be similar as described regarding
At block 808, the apparatus determines an input axis position on the input axis based, at least in part, on the input object and the determination that the input object is proximate to the input axis. The determination and the input axis position may be similar as described regarding
At block 810, the apparatus determines an operational directive based, at least in part, on the input axis position. The determination and the operational directive may be similar as described regarding
At block 812, the apparatus precludes determination of an input axis position on the input axis based, at least in part, on the determination that the input object fails to be proximate to the input axis. The preclusion, the determination, and the input axis position may be similar as described regarding
As discussed previously, in many circumstances, it may be desirable to configure an apparatus such that a user of the apparatus may determine an operational directive based, at least in part, on movement of an input object from an input axis position to a different input axis position.
At block 902, the apparatus determines an input axis that extends between an apparatus and a separate apparatus. The determination, the input axis, the apparatus, and the separate apparatus may be similar as described regarding
At block 904, the apparatus receives information indicative of an input object, the input object being proximate to the input axis. The receipt, the input object, and the input object being proximate to the input axis may be similar as described regarding
At block 906, the apparatus determines an input axis position on the input axis based, at least in part, on the input object. The determination and the input axis position may be similar as described regarding
At block 908, the apparatus determines an operational directive based, at least in part, on the input axis position. The determination and the operational directive may be similar as described regarding
At block 910, the apparatus receives information indicative of movement of the input object, the input object remaining proximate to the input axis. The receipt, the movement of the input object, and the input object remaining proximate to the input axis may be similar as described regarding
At block 912, the apparatus determines a different input axis position on the input axis based, at least in part, on the input object. The determination and the different input axis position may be similar as described regarding
At block 914, the apparatus determines a different operational directive based, at least in part, on the different input axis position. The determination and the different operational directive may be similar as described regarding
In many circumstances, a user of an electronic apparatus and a separate electronic apparatus may desire to interact with one or more of the electronic apparatuses by way of an input axis that extends from the electronic apparatus to the separate electronic apparatus. In such circumstances, the user may desire to ensure that an association the electronic apparatus and the separate electronic apparatus has been established, that communications between the electronic apparatus and the separate electronic apparatus are secure, and/or the like. As such, it may be desirable to configure an electronic apparatus such that a user of the electronic apparatus may indicate a desire for the electronic apparatus to establish a connection with a particular separate electronic apparatus in a manner that is quick, easy, and intuitive. For example, a user may desire to trace a virtual line between an electronic apparatus and a separate electronic apparatus such that the virtual line connects the electronic apparatus and the separate electronic apparatus. In this manner, the user may indicate to the electronic apparatus that the user desires to interact with the electronic apparatus by way of an input axis that extends from the electronic apparatus to the separate electronic apparatus, indicate to the separate electronic apparatus that the user desires to interact with the separate electronic apparatus by way of an input axis that extends from the separate electronic apparatus to the electronic apparatus, indicate to the electronic apparatus that the user desires to interact with the separate electronic apparatus by way of an input axis that extends from the electronic apparatus to the separate electronic apparatus, and/or the like.
In at least one example embodiment, an apparatus receives information indicative of movement of an input object. In such an example, the apparatus may determine a different input axis position on the input axis based, at least in part, on the input object. In at least one example embodiment, the apparatus determines that the different input axis position is further from the apparatus than the input axis position. For example, the user may trace a virtual line from the apparatus and to the separate apparatus. As such, the input object may initially be located at a position that corresponds to an input axis position that is proximate to the apparatus, and may move to a different position that corresponds with a different input axis position that is proximate to the separate apparatus. In such an example embodiment, the apparatus may determine the operational directive based, at least in part, on the determination that the different input axis position is further from the apparatus than the input axis position.
In such circumstances, it may be desirable to perform one or more operations associated with an operational directive such that an association between the electronic apparatus and the separate electronic apparatus is established. In at least one example embodiment, an operational directive is a separate apparatus authentication directive. The operational directive type may correspond with a separate apparatus authentication operational directive that causes establishment of a pairing association between the apparatus and the separate apparatus. As such, in at least one example embodiment, the operational directive causes the apparatus to determine an apparatus authentication parameter based, at least in part, on the input axis position. In such an example embodiment, the apparatus may send an authentication request to the separate apparatus, and subsequently receive an authentication response from the separate apparatus that is indicative of a separate apparatus authentication parameter. In such an example embodiment, the apparatus may determine that the separate apparatus authentication parameter corresponds with the apparatus authentication parameter, and causation of establishment of a pairing association between the apparatus and the separate apparatus based, at least in part, on the correspondence of the apparatus authentication parameter and the separate apparatus authentication parameter.
Similarly, the apparatus may receive an authentication request from a separate apparatus. In such an example, the apparatus may determine an operational directive that causes the apparatus to perform determination of an apparatus authentication parameter based, at least in part, on the input axis position and to send an authentication response that is indicative of the apparatus authentication parameter to the separate apparatus.
In some circumstances, an electronic apparatus may be associated with more than one separate electronic apparatus. For example, a user of the electronic apparatus may commonly use the electronic apparatus in conjunction with another electronic apparatus and a different electronic apparatus. As such, a pairing association between the electronic apparatus and the other electronic apparatus, and between the electronic apparatus and the different electronic apparatus, may have been previously established. In such circumstances, the user may desire at times to utilize the electronic apparatus in conjunction with the other electronic apparatus, and may desire at other times to utilize the electronic apparatus in conjunction with the different electronic apparatus. As such, it may be desirable to configure an electronic apparatus such that the user of the electronic apparatus may indicate the user's desire to utilize the electronic apparatus in conjunction with the other electronic apparatus, to utilize the electronic apparatus in conjunction with the different electronic apparatus, and/or the like, in a manner that is easy and intuitive. In at least one example embodiment, an apparatus identifies the separate apparatus by way of an input associated with an input axis that extends between the apparatus and the separate apparatus. For example, if the user desires to utilize the user's electronic apparatus in conjunction with the user's other electronic apparatus, the user may indicate such a desire by way of the input axis that extends between the electronic apparatus and the other electronic apparatus. Alternatively, for example, if the user desires to utilize the user's electronic apparatus in conjunction with the user's different electronic apparatus, the user may indicate such a desire by way of the input axis that extends between the electronic apparatus and the different electronic apparatus. In this manner, the user may indicate the user's desire to interact with a specific separate apparatus, and to cause establishment of an operative connection between the apparatus and the specific separate apparatus, by way of an input associated with an input axis that extends between the apparatus and the specific separate apparatus.
In the example of
At block 1006, apparatus 1002 determines an input axis that extends between apparatus 1002 and separate apparatus 1004. The determination and the input axis may be similar as described regarding
At block 1008, separate apparatus 1004 determines an input axis that extends between separate apparatus 1004 and apparatus 1002. The determination and the input axis may be similar as described regarding
At block 1010, apparatus 1002 receives information indicative of an input object, the input object being proximate to the input axis. The receipt, the input object, and the input object being proximate to the input axis may be similar as described regarding
At block 1012, separate apparatus 1004 receives information indicative of an input object, the input object being proximate to the input axis. The receipt, the input object, and the input object being proximate to the input axis may be similar as described regarding
At block 1014, apparatus 1002 determines an input axis position on the input axis based, at least in part, on the input object. The determination and the input axis position may be similar as described regarding
At block 1016, separate apparatus 1004 determines an input axis position on the input axis based, at least in part, on the input object. The determination and the input axis position may be similar as described regarding
At block 1018, apparatus 1002 determines an operational directive based, at least in part, on the input axis position. The determination and the operational directive may be similar as described regarding
At block 1020, separate apparatus 1004 determines an operational directive based, at least in part, on the input axis position. The determination and the operational directive may be similar as described regarding
At block 1022, apparatus 1002 determines an apparatus authentication parameter based, at least in part, on the input axis position.
At interaction 1024, apparatus 1002 sends an authentication request to separate apparatus 1004. In this manner, separate apparatus 1004 receives the authentication request from apparatus 1002.
At block 1026, separate apparatus 1004 determines a separate apparatus authentication parameter based, at least in part, on the input axis position.
At interaction 1028, separate apparatus 1004 sends an authentication response to apparatus 1002 that is indicative of the separate apparatus authentication parameter. In this manner, apparatus 1002 receives the authentication response from separate apparatus 1004 that is indicative of the separate apparatus authentication parameter.
At block 1030, apparatus 1002 determines that the separate apparatus authentication parameter corresponds with the apparatus authentication parameter.
At block 1032, apparatus 1002 causes establishment of a pairing association between apparatus 1002 and separate apparatus 1004.
Embodiments of the invention may be implemented in software, hardware, application logic or a combination of software, hardware, and application logic. The software, application logic and/or hardware may reside on the apparatus, a separate device, or a plurality of separate devices. If desired, part of the software, application logic and/or hardware may reside on the apparatus, part of the software, application logic and/or hardware may reside on a separate device, and part of the software, application logic and/or hardware may reside on a plurality of separate devices. In an example embodiment, the application logic, software or an instruction set is maintained on any one of various conventional computer-readable media.
If desired, the different functions discussed herein may be performed in a different order and/or concurrently with each other. For example, block 908 of
Although various aspects of the invention are set out in the independent claims, other aspects of the invention comprise other combinations of features from the described embodiments and/or the dependent claims with the features of the independent claims, and not solely the combinations explicitly set out in the claims.
It is also noted herein that while the above describes example embodiments of the invention, these descriptions should not be viewed in a limiting sense. Rather, there are variations and modifications which may be made without departing from the scope of the present invention as defined in the appended claims.
Eronen, Antti, Lehtiniemi, Arto, Eskolin, Peter, Arrasvuori, Juha, Coyle-Gilchrist, Matthew
Patent | Priority | Assignee | Title |
Patent | Priority | Assignee | Title |
7978091, | Aug 24 2006 | NAVISENSE, LLC | Method and device for a touchless interface |
9389594, | Sep 27 2012 | Nokia Technologies Oy | Method and apparatus for providing an interactive cable-based interface to applications and services |
20090265470, | |||
20100231509, | |||
20110080490, | |||
20110249024, | |||
20110316790, | |||
20120131518, | |||
20120326966, | |||
20130050069, | |||
20130229508, | |||
20140145942, | |||
20140157210, | |||
20140320383, | |||
20170347265, | |||
EP2302883, | |||
EP2348383, | |||
EP2455841, | |||
EP2610708, | |||
JP2012032945, | |||
WO2007096893, | |||
WO2009074887, | |||
WO2009144685, | |||
WO2012141350, | |||
WO2012153227, | |||
WO2013094786, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Mar 05 2014 | COYLE-GILCHRIST, MATTHEW | Nokia Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 040186 | /0304 | |
Mar 09 2014 | ARRASVUORI, JUHA | Nokia Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 040186 | /0304 | |
Mar 12 2014 | LEHTINIEMI, ARTO | Nokia Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 040186 | /0304 | |
Mar 13 2014 | ERONEN, ANTTI | Nokia Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 040186 | /0304 | |
Mar 14 2014 | ESKOLIN, PETER | Nokia Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 040186 | /0304 | |
Jan 16 2015 | Nokia Corporation | Nokia Technologies Oy | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 040186 | /0359 | |
Feb 27 2015 | Nokia Technologies Oy | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Jan 17 2024 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Date | Maintenance Schedule |
Aug 04 2023 | 4 years fee payment window open |
Feb 04 2024 | 6 months grace period start (w surcharge) |
Aug 04 2024 | patent expiry (for year 4) |
Aug 04 2026 | 2 years to revive unintentionally abandoned end. (for year 4) |
Aug 04 2027 | 8 years fee payment window open |
Feb 04 2028 | 6 months grace period start (w surcharge) |
Aug 04 2028 | patent expiry (for year 8) |
Aug 04 2030 | 2 years to revive unintentionally abandoned end. (for year 8) |
Aug 04 2031 | 12 years fee payment window open |
Feb 04 2032 | 6 months grace period start (w surcharge) |
Aug 04 2032 | patent expiry (for year 12) |
Aug 04 2034 | 2 years to revive unintentionally abandoned end. (for year 12) |