Described herein are systems and methods for managing video data. Embodiments are described by reference to a digital video Management (DVM) system, for example methods for providing a user with access to view live feed from cameras in a DVM system. In overview, a DVM system is configured to provide a user interface for allowing an operator to view a plurality of video display objects, wherein each display object is configured to display live video data captured at a camera of the DVM system. The user interface is further configured to provide access to one or more camera groups, wherein the one or more camera groups are respectively defined by dynamic parameters. The DVM system executes a process for updating the constituent cameras of each camera group responsive to changes in the dynamic parameters. For example, in one embodiment there is a camera group defined by cameras recently viewed by a particular client. That group is updated thereby to provide an operator with quick and convenient access to recently viewed cameras.

Patent
   10863143
Priority
Aug 03 2012
Filed
Jun 11 2019
Issued
Dec 08 2020
Expiry
Aug 03 2032

TERM.DISCL.
Assg.orig
Entity
Large
0
320
currently ok
11. A display of a digital video management (DVM) system, the display configured to display:
a camera display history object including a plurality of video display objects each simultaneously displaying live video data capture at a camera of the DVM system; and
a set of camera objects, at least one of the camera objects including one or more camera groups, each camera group including one or more constituent cameras that are all providing live video streaming; and
wherein:
the one or more camera groups are defined based on dynamic parameters which change automatically over time such that the constituent cameras that are in a given group change over time; and
the constituent cameras that make up each camera group in response to changes in the dynamic parameters.
1. A user interface for a digital video management (DVM) system, the user interface comprising:
a camera display history object including a plurality of video display objects each simultaneously displaying live video data capture at a camera of the DVM system; and
a set of camera objects, at least one of the camera objects including one or more camera groups, each camera group including one or more constituent cameras that are all providing live video streaming; and
wherein:
the one or more camera groups are defined based on dynamic parameters which change automatically over time such that the constituent cameras that are in a given group change over time; and
the constituent cameras that make up each camera group in response to changes in the dynamic parameters.
17. A non-transitory computer-readable medium storing instructions that when executed by a processing unit are configured to display:
a camera display history object including a plurality of video display objects each simultaneously displaying live video data capture at a camera of the DVM system; and
a set of camera objects, at least one of the camera objects including one or more camera groups, each camera group including one or more constituent cameras that are all providing live video streaming; and
wherein:
the one or more camera groups are defined based on dynamic parameters which change automatically over time such that the constituent cameras that are in a given group change over time; and
the constituent cameras that make up each camera group in response to changes in the dynamic parameters.
2. The user interface of claim 1, wherein the one or more camera groups includes a recently viewed camera group defined by the most recent n cameras viewed via the user interface, wherein 0<n<a predefined maximum.
3. The user interface of claim 2, wherein the user interface is configured to display a screen object containing video display objects indicative of the cameras in the recently viewed camera group.
4. The user interface of claim 3, wherein the video display objects provide live video streaming.
5. The user interface of claim 4, wherein the live video streaming is at a rate less than the rate of streaming when the cameras where being viewed.
6. The user interface of claim 1, wherein the live video streaming rate varies between the cameras in the viewed camera groups in response to specified criteria.
7. The user interface of claim 1, wherein the user interface is configured to enable user creation of camera groups based on specified dynamic parameters.
8. The user interface of claim 1, wherein the one or more groups are defined by dynamic parameters thereby to configure groups for one or more of the following:
cameras recently viewed at a given client terminal;
cameras recently viewed across the DVM system;
cameras having a high view rate over a predefined period;
cameras having greater than a threshold incident rate over a predefined period;
cameras having greater than a threshold motion rate over a predefined period;
cameras having greater than a threshold number of movement observations, based on analytics software, over a predefined period;
cameras in respect of which alarms have been raised over a predefined period;
cameras which have had greater than a threshold number of fail or error conditions over a predetermined period; and
cameras having a fail or error condition.
9. The user interface of claim 8, wherein each group has a maximum size, and is populated by a set of cameras of up to and including the maximum size that best satisfy the dynamic parameters.
10. The user interface of claim 1, wherein a given camera group is displayable in a user interface object including a screenshot or streaming feed for each of the cameras.
12. The display of claim 11, wherein a camera server is configured to utilize video data from an assigned one or more video streaming units and a streaming unit is configured to stream, onto a network, video data for a respective camera.
13. The display of claim 11, wherein the one or more camera groups includes a recently viewed camera group defined by the most recent n cameras viewed via the user interface, wherein 0<n<a predefined maximum.
14. The display of claim 13 further configured to display a screen object containing video display objects indicative of the cameras in the recently viewed camera group.
15. The display of claim 11, wherein the live video streaming rate varies between the cameras in the viewed camera groups in response to specified criteria.
16. The display of claim 11, wherein the user interface is configured to enable user creation of camera groups based on specified dynamic parameters.
18. The non-transitory computer-readable medium according to claim 17 wherein the one or more camera groups includes a recently viewed camera group defined by the most recent n cameras viewed via the user interface, wherein 0<n<a predefined maximum.
19. The non-transitory computer-readable medium according to claim 17 wherein the live video streaming rate varies between the cameras in the viewed camera groups in response to specified criteria.
20. The non-transitory computer-readable medium according to claim 17 wherein the user interface is configured to enable user creation of camera groups based on specified dynamic parameters.

This application is a continuation of U.S. patent application Ser. No. 13/566,777, filed Aug. 3, 2012. U.S. patent application Ser. No. 13/566,777, filed on Aug. 3, 2012, is hereby incorporated by reference.

This application claims priority under 35 U.S.C. § 119 to Australian (AU) Patent Application No. 2011903152, filed on Aug. 5, 2011. The Australian (AU) Patent Application No. 2011903152, filed on Aug. 5, 2011, is hereby incorporated by reference.

The present invention relates to systems and methods for managing video data. Embodiments of the invention have been particularly developed for managing user access to cameras in a Digital Video Management (DVM) system. While some embodiments will be described herein with particular reference to that application, it will be appreciated that the invention is not limited to such a field of use, and is applicable in broader contexts.

Any discussion of the background art throughout the specification should in no way be considered as an admission that such art is widely known or forms part of common general knowledge in the field.

Digital Video Management (DVM) systems, such as those based on the Honeywell DVM model, are widely used. In overview, a plurality of cameras are assigned to a plurality camera servers, with each camera server being configured to make available (for live viewing or recording purposes) video data from an assigned one or more cameras. The camera servers are all centrally managed by a DVM database server.

In sites with a large number of cameras (for example more than 1,000 cameras), there are significant challenges in providing quick and convenient access to a desired one or more cameras. Such quick and convenient access is in some cases crucial, for example during an incident or emergency.

There is a need in the art for improved systems and methods for managing video data.

It is an object of the present invention to overcome or ameliorate at least one of the disadvantages of the prior art, or to provide a useful alternative.

One embodiment provides a method for managing video data in a DVM system, the method including:

providing a user interface for allowing an operator to view a plurality of video display objects, wherein each display object is configured to display live video data captured at a camera of the DVM system;

via the user interface, providing access to one or more camera groups, wherein the one or more camera groups are respectively defined by dynamic parameters; and

updating the constituent cameras of each camera group responsive to changes in the dynamic parameters.

One embodiment provides a DVM system configured to perform a method as described herein.

One embodiment provides a tangible non-transitive carrier medium carrying computer executable code that, when executed via one or more processes, allows the performance of a method as described herein.

Reference throughout this specification to “one embodiment”, “some embodiments” or “an embodiment” means that a particular feature, structure or characteristic described in connection with the embodiment is included in at least one embodiment of the present invention. Thus, appearances of the phrases “in one embodiment”, “in some embodiments” or “in an embodiment” in various places throughout this specification are not necessarily all referring to the same embodiment, but may. Furthermore, the particular features, structures or characteristics may be combined in any suitable manner, as would be apparent to one of ordinary skill in the art from this disclosure, in one or more embodiments.

As used herein, unless otherwise specified the use of the ordinal adjectives “first”, “second”, “third”, etc., to describe a common object, merely indicate that different instances of like objects are being referred to, and are not intended to imply that the objects so described must be in a given sequence, either temporally, spatially, in ranking, or in any other manner.

In the claims below and the description herein, any one of the terms comprising, comprised of or which comprises is an open term that means including at least the elements/features that follow, but not excluding others. Thus, the term comprising, when used in the claims, should not be interpreted as being limitative to the means or elements or steps listed thereafter. For example, the scope of the expression a device comprising A and B should not be limited to devices consisting only of elements A and B. Any one of the terms including or which includes or that includes as used herein is also an open term that also means including at least the elements/features that follow the term, but not excluding others. Thus, including is synonymous with and means comprising.

Embodiments of the invention will now be described, by way of example only, with reference to the accompanying drawings in which:

FIG. 1 schematically illustrates a DVM system according to one embodiment.

FIG. 2A schematically illustrates a DVM system according to one embodiment.

FIG. 2B schematically illustrates a DVM system client terminal and user interface according to one embodiment.

FIG. 2C schematically illustrates a DVM system client terminal and user interface according to one embodiment.

FIG. 3 illustrates a method according to one embodiment.

Described herein are systems and methods for managing video data. Embodiments are described by reference to a Digital Video Management (DVM) system, for example methods for providing a user with access to view live feed from cameras in a DVM system. In overview, a DVM system is configured to provide a user interface for allowing an operator to view a plurality of video display objects, wherein each display object is configured to display live video data captured at a camera of the DVM system. The user interface is further configured to provide access to one or more camera groups, wherein the one or more camera groups are respectively defined by dynamic parameters. The DVM system executes a process for updating the constituent cameras of each camera group responsive to changes in the dynamic parameters. For example, in one embodiment there is a camera group defined by cameras recently viewed by a particular client. That group is updated thereby to provide an operator with quick and convenient access to recently viewed cameras.

FIG. 1 illustrates a general Digital Video Management (DVM) system 101. System 101 is described to provide general context to various embodiments discussed below. Although embodiments are described by reference to DVM systems based on system 101, the present invention is not limited as such. That is, system 101 is provided as a general example to highlight various features of an exemplary DVM system. In practice, many systems omit one or more of these features, and/or include additional features.

System 101 includes a plurality of video streaming units 102. Units 102 include conventional cameras 104 (including analogue video cameras) coupled to discrete video streaming units, and IP streaming cameras 105. Video streaming units 102 stream video data, presently in the form of surveillance footage, on a TCP/IP network 106. This is readily achieved using IP streaming cameras 105, which are inherently adapted for such a task. However, in the case of other cameras 104 (such as conventional analogue cameras), a discrete video streaming unit 107 is required to convert a captured video signal into a format suitable for IP streaming.

For the purposes of the present disclosure, the term “video streaming unit” should be read to include IP streaming cameras 105 and video streaming units 107. That is, the term “video streaming unit” describes any hardware component configured to stream video data onto a network, independent of the source of the originating analogue video data.

For the present purposes, the terms “video streaming unit” and “camera” are generally used interchangeably, on the assumption that each video streaming unit corresponds to a unique set of optical components used to capture video. That is, there is a one-to-one relationship between streaming units 107 and cameras 104. However, in other embodiments there is a one-to-many relationship between streaming units 107 and cameras 104 (i.e. a streaming unit is configured for connection to multiple cameras).

One or more camera servers 109 are also connected to network 106 (these may be either physical servers or virtual servers). Each camera server is enabled to have assigned to it one or more of video streaming units 102. In some embodiments the assignment is on a stream-by-stream basis rather than a camera-by-camera basis. This assignment is carried out using a software-based configuration tool, and it follows that camera assignment is virtual rather than physical. That is, the relationships are set by software configuration rather than hardware manipulation. In practice, each camera has a unique identifier. Data indicative of this identifier is included with surveillance footage being streamed by that camera such that components on the network are able to ascertain from which camera a given stream originates.

In the present embodiment, camera servers are responsible for making available both live and stored video data. In relation to the former, each camera server provides a live stream interface, which consists of socket connections between the camera manager and clients. Clients request live video through the camera server's COM interfaces and the camera server then pipes video and audio straight from the relevant streaming unit to the client through TCP sockets. In relation to the latter, each camera server has access to a data store for recording video data. Although FIG. 1 suggests a one-to-one relationship between camera servers and data stores, this is by no means necessary. Each camera server also provides a playback stream interface, which consists of socket connections between the camera manager and clients. Clients create and control the playback of video stored that the camera server's data store through the camera manager's COM interfaces and the stream is sent to clients via TCP sockets.

Although, in the context of the present disclosure, there is discussion of one or more cameras or streaming units being assigned to a common camera server, this is a conceptual notion, and is essentially no different from a camera server being assigned to one or more cameras or streaming units.

Clients 110 execute on a plurality of client terminals, which in some embodiments include all computational platform on network 106 that are provided with appropriate permissions. Clients 110 provide a user interface (UI) that allows surveillance footage to be viewed in real time by an end-user. For example, one UI component is a render window, in which streamed video data is rendered for display to a user. In some cases this user interface is provided through an existing application (such as Microsoft Internet Explorer), whilst in other cases it is a standalone application. The user interface optionally provides the end-user with access to other system and camera functionalities, including mechanical, digital and optical camera controls, control over video storage, and other configuration and administrative functionalities (such as the assignment and reassignment of cameras to camera servers). Typically clients 110 are relatively “thin”, and commands provided via the relevant user interfaces are implemented at a remote server, typically a camera server. In some embodiments different clients have different levels of access rights. For example, in some embodiments there is a desire to limit the number of users with access to change configuration settings or mechanically control cameras.

System 101 also includes a DVM database server 115. Database server 115 is responsible for maintaining various information relating to configurations and operational characteristics of system 101, and for managing events within the system. In terms of events, the general notion is that an action in the system (such as the modification of data in the database, or the reservation of a camera, as discusses below) causes an event to be “fired” (i.e. published), this having follow-on effects depending on the nature of the event.

In the present example, the system makes use of a preferred and redundant database server (115 and 116 respectively), the redundant server essentially operating as a backup for the preferred server. The relationship between these database servers is generally beyond the concern of the present disclosure.

Some embodiments of the present invention are directed to distributed DVM systems, also referred to as “distributed system architecture” (DSA). In general terms, a distributed DVM system includes a plurality of (i.e. two or more) discrete DVM systems, such as system 101. These systems are discrete in the sense that they are in essence standalone systems, able to function autonomously without the other by way of their own DVM servers. They may be distributed geographically (for example in different buildings, cities or countries), or notionally (in a common geographic location, but split due to individual system constraints, for example camera server numbers, or simply to take advantage of benefits of a distributed architecture). In the context of FIG. 1, a remote system 150, communicates with the local system via a DSA link 151. For the present purposes, it is assumed that remote system 150 is in a general sense similar to the local system. Various components (hardware and software) are configured to allow communications between the systems, for example via a network connection (including, but not limited to, an Intranet or Internet connection), or other communications interface. For the sake of the present embodiments, it is assumed that the inter-system communications occur by way of TCP/IP connections, and in this manner any communications channel supporting TCP/IP may be used.

FIG. 2 illustrates components of an exemplary DVM system (such as the system of FIG. 1), but simplified to illustrate components relevant to camera group management.

A camera 201 is associated with a camera server 202. Camera server 202 is configured to access video data made available by camera 201, either for live viewing or for recording to a storage device 203. Camera server 202 is configured/controlled by a DVM server 204. There may be a large number of cameras and camera servers configured/controlled by DVM server 204.

DVM server 204 executes DVM administration modules 205. The functional block for modules 205 is used to simplistically represent a wide range of software components implemented within a DVM system. Only a selection of these are shown, being user interface modules 230 and a content sharing module 250.

DVM server 204 communicates with a user interface 210 which executes on a client terminal 211. In the present embodiment, this user interface is provided via module 230 via a web-server type arrangement (i.e. user interface 210 is provided via a web-browser at terminal 211 which renders data transmitted by server 211). In this manner, module 230 is configured to allow a plurality of terminals 211 to independently provide respective instantiations user interface 210 for a respective plurality of operators.

User interface 210 is configured to display live and recorded video data to a user via a video display objects (and, in some embodiments, other DVM content such as screenshots, maps, and the like). In the example of FIG. 2, a plurality of display objects are shown as being rendered on-screen simultaneously, including a major display object 212 and minor display objects 213A-E. These are configured to each display live video data from respective cameras (such as camera 201).

User interface 210 provides various controls 217 (which simplistically represents a variety of GUI controls available to an operator of terminal 211, such as record control, camera position control, camera view selection, and so on).

DVM server 204 additionally provides a camera group management module 250. However, it will be appreciated that in other embodiments some or all of the functionality of module 250 is handled at client terminal 211 within user interface 210. In overview, module 250 is responsible for managing camera groups that are respectively defined by dynamic parameters, such as camera parameters, user parameters, user interface parameters, or camera server parameters. The general crux of the parameters being “dynamic” is that the values for the parameters change over time, such that the constituent cameras that make up a given group vary over time. Module 250 is configured to keep the groups up to date, achieved by either assessing parameters periodically, updating the group in response to monitored events, or updating the group in response to a request that requires the group to be updated (for example a “refresh” command).

In some embodiments user interface 210 is configured to enable user creation of camera groups based on specified dynamic parameters, for example by way of a camera group controls object 216, which provides access an interface by which a user is permitted to create a new group or modify an existing group. In either case, the creation/modification process includes selection of parameters (such as dynamic parameters) that are to define the group, with the selection being made from a list of available parameters. By way of example, parameters are optionally selected thereby to configure groups for one or more of the following:

Cameras recently viewed at a given client terminal. For example, the dynamic parameter may be a camera “last accessed time” for cameras viewed at the relevant client terminal.

Cameras recently viewed across the DVM system. For example, the dynamic parameter may be a camera “last accessed time”.

Cameras having a high view rate over a predefined period. For example, the dynamic parameter may be a “times viewed” and/or a parameter related to the number of viewing clients and time viewed by each client.

Cameras having greater than a threshold incident rate over a predefined period. For example, the dynamic parameter may be related to an “incidents recorded” value and/or data supplied by video analytics software.

Cameras having greater than a threshold motion rate over a predefined period. For example, the dynamic parameter may be a measure of camera movement, or number of times the camera was moved (e.g. in terms of pan, tilt and zoom).

Cameras having greater than a threshold number of movement observations, based on analytics software, over a predefined period. For example, the dynamic parameter may be a value supplied by video analytics software that is configured to identify movement in a camera's field of view.

Cameras in respect of which alarms have been raised over a predefined period. For example, the dynamic parameter may be a camera alarm count value.

Cameras displaying a fail condition. For example, the dynamic parameter may be an active camera fail or error value.

Cameras which have had greater than a threshold number of fail or error conditions over a predetermined period. This is useful in allowing a user to observe the feed from these cameras, which may intermittently fail given the error rate.

Cameras that are proximally close (in location or field of view) to currently or recently viewed cameras.

It will be appreciated that there is a wide range of groups able to be created, given the flexible approach of allowing a user to select defining parameters.

In some cases, wherein each group has a maximum size, and is populated by a set of cameras of up to and including the maximum size that best satisfy the dynamic parameters. For example, the maximum size may be 10, and the 10 cameras selected to populate the group are either those that best match the dynamic parameter, or those which have most recently matched the dynamic parameter. In the case of a “recently used cameras” group, this would display the 10 most recently sued cameras, with the 11th most recent dropping out of the group each time another camera ceases to be used.

In some cases a camera group is displayable in a list or tree structure, and cameras in the list are able to be dragged and dropped into user interface video display objects. In some cases, as an alternative or in additional to “drag and drop”, a click-based selection is made, and the relevant camera view is automatically moved to a major display object, or moved to a major/minor display object selected based on a prioritization algorithm.

In some cases a camera group is able to be launched into a multi-object view (having an object arrangement similar to that of FIG. 2A, or alternately a different geometric layout). The example of FIG. 2A shows a user interface object, in the form of a camera group preview object 260, which includes a screenshot or streaming live feed for each of the cameras in the group. The screenshots, or streaming live feeds, are displayed in video display objects 261A-E. In some cases, where live streaming is provided via objects 261A-E, the live video streaming rate varies between the cameras in response to specified criteria. For example, in some cases there are additional objects displayed off-screen (requiring a user to scroll down to make them visible); these are streamed at a lower rate until displayed on screen. In the cases of a “recently viewed” camera group, the streaming rate decreases based on the time elapsed since a camera was viewed and/or as the number of cameras more recently viewed than that camera increases.

FIG. 2B illustrates an embodiment specifically directed towards maintaining convenient access to recently used cameras. To this end, user interface 210 includes a camera display history object 270, which is configured to contain a plurality of video display objects (with objects 271A-E being shown, and potentially additional accessible by way of a scrolling operation). During use of user interface 210, which cameras are displayed in objects 212 and 213A-E varies over time (for example as a result of user activity, or as a result of automated events such as alarms or the like). Each time a new camera is displayed by one of the display objects, another camera which was previously displayed ceases to be displayed (given the limited number of display objects). That is, one camera moves from a state of being presently used, to being the most recently used. The recently viewed camera group of object 270 preferably ahs a maximum size of n. That is, the group is defined by the most recent n cameras viewed via the user interface, wherein 0<n<a predefined maximum.

Objects 271A-E are respectively indicative of the most recently used cameras. These objects in some embodiments provide live video streaming of their respective cameras (although preferably at a lower rate/quality thereby to conserve CPU and network resources). In other embodiments they each provide a still image taken from the respective cameras (which may be predefined or taken at a designated time, such as the time at which the camera was last viewed, and is optionally periodically updated). In some embodiments, objects 271A-E are camera names or descriptors.

Preferably, a selected one of objects 271A-E is able to be dragged-and-dropped onto a selected major or minor video display object, thereby to instruct the user interface to display the video stream from the camera associated with the selected object 217A-E via the selected major/minor display object.

FIG. 3 illustrates an exemplary method 300 for managing a recently sued camera group. In the context of method 300, 301 includes identifying a change in cameras displayed in major/minor video display objects. 302 includes determining a camera or set of cameras that are no longer displayed in major/minor display objects based on the identified change. Those cameras are added to the “recently viewed” camera group at 303, and the display of that group (for example object 270) us updated accordingly at 304.

FIG. 2C illustrates a further exemplary user interface instantiation. This is similar to FIG. 2B, but with two additional objects on the left hand side. An attribute based camera tree object 290 sets out cameras based on attributes (such as location or the like). This may be in the form of a hierarchical tree, or a list. A group based camera tree 291 sets out cameras based on camera groups. That is, a user is able to navigate camera groups using object 291 and then, for example, drag and drop a desired camera into one of the major or minor video display objects thereby to view live video data from that camera. In some cases a user ca select a camera group, and launch a multi-display-object view for simultaneously viewing live video data for a selection of or all of the cameras in that group.

It will be appreciated that the disclosure above provides various significant systems and methods for managing video data. For example, the present embodiments allows for cameras to be more readily accessed by a user, either by way of flexibly defined groups based on dynamic parameters, or by configuring an object to keep track of recently used cameras.

Unless specifically stated otherwise, as apparent from the following discussions, it is appreciated that throughout the specification discussions utilizing terms such as “processing,” “computing,” “calculating,” “determining”, analyzing” or the like, refer to the action and/or processes of a computer or computing system, or similar electronic computing device, that manipulate and/or transform data represented as physical, such as electronic, quantities into other data similarly represented as physical quantities.

In a similar manner, the term “processor” may refer to any device or portion of a device that processes electronic data, e.g., from registers and/or memory to transform that electronic data into other electronic data that, e.g., may be stored in registers and/or memory. A “computer” or a “computing machine” or a “computing platform” may include one or more processors.

The methodologies described herein are, in one embodiment, performable by one or more processors that accept computer-readable (also called machine-readable) code containing a set of instructions that when executed by one or more of the processors carry out at least one of the methods described herein. Any processor capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken are included. Thus, one example is a typical processing system that includes one or more processors. Each processor may include one or more of a CPU, a graphics processing unit, and a programmable DSP unit. The processing system further may include a memory subsystem including main RAM and/or a static RAM, and/or ROM. A bus subsystem may be included for communicating between the components. The processing system further may be a distributed processing system with processors coupled by a network. If the processing system requires a display, such a display may be included, e.g., a liquid crystal display (LCD) or a cathode ray tube (CRT) display. If manual data entry is required, the processing system also includes an input device such as one or more of an alphanumeric input unit such as a keyboard, a pointing control device such as a mouse, and so forth. The term memory unit as used herein, if clear from the context and unless explicitly stated otherwise, also encompasses a storage system such as a disk drive unit. The processing system in some configurations may include a sound output device, and a network interface device. The memory subsystem thus includes a computer-readable carrier medium that carries computer-readable code (e.g., software) including a set of instructions to cause performing, when executed by one or more processors, one of more of the methods described herein. Note that when the method includes several elements, e.g., several steps, no ordering of such elements is implied, unless specifically stated. The software may reside in the hard disk, or may also reside, completely or at least partially, within the RAM and/or within the processor during execution thereof by the computer system. Thus, the memory and the processor also constitute computer-readable carrier medium carrying computer-readable code.

Furthermore, a computer-readable carrier medium may form, or be included in a computer program product.

In alternative embodiments, the one or more processors operate as a standalone device or may be connected, e.g., networked to other processor(s), in a networked deployment, the one or more processors may operate in the capacity of a server or a user machine in server-user network environment, or as a peer machine in a peer-to-peer or distributed network environment. The one or more processors may form a personal computer (PC), a tablet PC, a set-top box (STB), a Personal Digital Assistant (PDA), a cellular telephone, a web appliance, a network router, switch or bridge, or any machine capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that machine.

Note that while some diagrams only show a single processor and a single memory that carries the computer-readable code, those in the art will understand that many of the components described above are included, but not explicitly shown or described in order not to obscure the inventive aspect. For example, while only a single machine is illustrated, the term “machine” shall also be taken to include any collection of machines that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methodologies discussed herein.

Thus, one embodiment of each of the methods described herein is in the form of a computer-readable carrier medium carrying a set of instructions, e.g., a computer program that is for execution on one or more processors, e.g., one or more processors that are part of web server arrangement. Thus, as will be appreciated by those skilled in the art, embodiments of the present invention may be embodied as a method, an apparatus such as a special purpose apparatus, an apparatus such as a data processing system, or a computer-readable carrier medium, e.g., a computer program product. The computer-readable carrier medium carries computer readable code including a set of instructions that when executed on one or more processors cause the processor or processors to implement a method. Accordingly, aspects of the present invention may take the form of a method, an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present invention may take the form of carrier medium (e.g., a computer program product on a computer-readable storage medium) carrying computer-readable program code embodied in the medium.

The software may further be transmitted or received over a network via a network interface device. While the carrier medium is shown in an exemplary embodiment to be a single medium, the term “carrier medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions. The term “carrier medium” shall also be taken to include any medium that is capable of storing, encoding or carrying a set of instructions for execution by one or more of the processors and that cause the one or more processors to perform any one or more of the methodologies of the present invention. A carrier medium may take many forms, including but not limited to, non-volatile media, volatile media, and transmission media. Non-volatile media includes, for example, optical, magnetic disks, and magneto-optical disks. Volatile media includes dynamic memory, such as main memory. Transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise a bus subsystem. Transmission media also may also take the form of acoustic or light waves, such as those generated during radio wave and infrared data communications. For example, the term “carrier medium” shall accordingly be taken to included, but not be limited to, solid-state memories, a computer product embodied in optical and magnetic media; a medium bearing a propagated signal detectable by at least one processor of one or more processors and representing a set of instructions that, when executed, implement a method; a carrier wave bearing a propagated signal detectable by at least one processor of the one or more processors and representing the set of instructions a propagated signal and representing the set of instructions; and a transmission medium in a network bearing a propagated signal detectable by at least one processor of the one or more processors and representing the set of instructions.

It will be understood that the steps of methods discussed are performed in one embodiment by an appropriate processor (or processors) of a processing (i.e., computer) system executing instructions (computer-readable code) stored in storage. It will also be understood that the invention is not limited to any particular implementation or programming technique and that the invention may be implemented using any appropriate techniques for implementing the functionality described herein. The invention is not limited to any particular programming language or operating system.

Reference throughout this specification to “one embodiment” or “an embodiment” means that a particular feature, structure or characteristic described in connection with the embodiment is included in at least one embodiment of the present invention. Thus, appearances of the phrases “in one embodiment” or “in an embodiment” in various places throughout this specification are not necessarily all referring to the same embodiment, but may. Furthermore, the particular features, structures or characteristics may be combined in any suitable manner, as would be apparent to one of ordinary skill in the art from this disclosure, in one or more embodiments.

Similarly it should be appreciated that in the above description of exemplary embodiments of the invention, various features of the invention are sometimes grouped together in a single embodiment, FIG., or description thereof for the purpose of streamlining the disclosure and aiding in the understanding of one or more of the various inventive aspects. This method of disclosure, however, is not to be interpreted as reflecting an intention that the claimed invention requires more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive aspects lie in less than all features of a single foregoing disclosed embodiment. Thus, the claims following the Detailed Description are hereby expressly incorporated into this Detailed Description, with each claim standing on its own as a separate embodiment of this invention.

Furthermore, while some embodiments described herein include some but not other features included in other embodiments, combinations of features of different embodiments are meant to be within the scope of the invention, and form different embodiments, as would be understood by those skilled in the art. For example, in the following claims, any of the claimed embodiments can be used in any combination.

Furthermore, some of the embodiments are described herein as a method or combination of elements of a method that can be implemented by a processor of a computer system or by other means of carrying out the function. Thus, a processor with the necessary instructions for carrying out such a method or element of a method forms a means for carrying out the method or element of a method. Furthermore, an element described herein of an apparatus embodiment is an example of a means for carrying out the function performed by the element for the purpose of carrying out the invention.

In the description provided herein, numerous specific details are set forth. However, it is understood that embodiments of the invention may be practiced without these specific details. In other instances, well-known methods, structures and techniques have not been shown in detail in order not to obscure an understanding of this description.

Similarly, it is to be noticed that the term coupled, when used in the claims, should not be interpreted as being limited to direct connections only. The terms “coupled” and “connected,” along with their derivatives, may be used. It should be understood that these terms are not intended as synonyms for each other. Thus, the scope of the expression a device A coupled to a device B should not be limited to devices or systems wherein an output of device A is directly connected to an input of device B. It means that there exists a path between an output of A and an input of B which may be a path including other devices or means. “Coupled” may mean that two or more elements are either in direct physical or electrical contact, or that two or more elements are not in direct contact with each other but yet still co-operate or interact with each other.

Thus, while there has been described what are believed to be the preferred embodiments of the invention, those skilled in the art will recognize that other and further modifications may be made thereto without departing from the spirit of the invention, and it is intended to claim all such changes and modifications as falling within the scope of the invention. For example, any formulas given above are merely representative of procedures that may be used. Functionality may be added or deleted from the block diagrams and operations may be interchanged among functional blocks. Steps may be added or deleted to methods described within the scope of the present invention.

Worrill, Joshua

Patent Priority Assignee Title
Patent Priority Assignee Title
3753232,
3806911,
3857018,
3860911,
3866173,
3906447,
4095739, Aug 26 1977 CASI-RUSCO INC System for limiting access to security system program
4146085, Oct 03 1977 YORK INTERNATIONAL CORPORATION, 631 SOUTH RICHLAND AVENUE, YORK, PA 17403, A CORP OF DE Diagnostic system for heat pump
4148012, Sep 26 1975 CARDKEY SYSTEMS, INC , A CORP OF OREGON Access control system
4161778, Jul 19 1977 Honeywell Information Systems, Inc. Synchronization control system for firmware access of high data rate transfer bus
4213118, Nov 08 1976 UNIQEY LOCK COMPANY Combination changing system and method
4283710, Nov 08 1976 UNIQEY LOCK COMPANY Security system
4298946, Dec 18 1978 Honeywell INC Electronically controlled programmable digital thermostat
4332852, Mar 29 1978 CARDBORUNDUM COMPANY, THE; Unifrax Corporation Conditioned colloidal silica post impregnant to prevent binder migration in the production of insulation articles comprising randomly oriented refractory fibers
4336902, Oct 28 1977 Thermostat alterable by remote control
4337893, Dec 17 1979 Energy Savings Parhelion; ENERGY SAVINGS PARHELION A PARTNERSHIP OF Multi-phase modular comfort controlled heating system
4353064, Jan 14 1981 Honeywell Inc. Battery operated access control card
4373664, Oct 11 1977 Robertshaw Controls Company Wall thermostat and the like
4379483, Aug 17 1981 EVCON INDUSTRIES, INC , A CORP OF DE Method of controlling heating and cooling sources
4462028, Feb 19 1981 Honeywell Information Systems Inc. Access control logic for video terminal display memory
4525777, Aug 03 1981 Honeywell Information Systems Inc. Split-cycle cache system with SCU controlled cache clearing during cache store access period
4538056, Aug 27 1982 CASI-RUSCO INC , A CORP OF FLORIDA Card reader for time and attendance
4556169, Jun 07 1984 Honeywell Inc. On-off thermostat based modulating air flow controller
4628201, Mar 05 1983 Dr. Johannes Heidenhain GmbH Scanning signal balancing circuit
4646964, Mar 26 1982 Carrier Corporation Temperature control system
4685615, Dec 17 1984 Diagnostic thermostat
4821177, Sep 02 1986 Honeywell Bull Inc. Apparatus for controlling system accesses having multiple command level conditional rotational multiple port servicing priority hierarchy
4847839, Aug 26 1987 HONEYWELL INC , HONEYWELL PLAZA, MINNEAPOLIS, MINNESOTA 55408, A CORP OF DE Digital registers with serial accessed mode control bit
5070468, Jul 20 1988 Mitsubishi Jukogyo Kabushiki Kaisha; Idemitsu Kosan Company Limited Plant fault diagnosis system
5071065, Jan 13 1989 Halton Oy Procedure for controlling and maintaining air currents or equivalent in an air-conditioning installation, and an air-conditioning system according to said procedure
5099420, Jan 10 1989 Bull HN Information Systems Inc.; HONEYWELL BULL INC , A DE CORP Method and apparatus for limiting the utilization of an asynchronous bus with distributed controlled access
5172565, May 21 1990 Honeywell Inc. Air handling system utilizing direct expansion cooling
5204663, May 21 1990 Applied Systems Institute, Inc. Smart card access control system
5227122, Nov 02 1989 WESTINGHOUSE ELECTRIC CO LLC Display device for indicating the value of a parameter in a process plant
5259553, Apr 05 1991 Norm Pacific Automation Corp. Interior atmosphere control system
5271453, Jun 06 1991 Hitachi, Ltd. System and method for controlling air conditioner
5361982, Jul 12 1993 Johnson Controls Technology Company Temperature control system having central control for thermostats
5404934, May 19 1992 Currise & Carlson, Inc. Retrofit air conditioning system
5420927, Feb 01 1994 ASSA ABLOY AB Method for certifying public keys in a digital signature scheme
5449112, Mar 15 1994 Method and apparatus for monitoring and controlling air handling systems
5465082, Jul 27 1990 Hill-Rom Services, Inc Apparatus for automating routine communication in a facility
5479154, Apr 06 1990 Siemens Aktiengesellschaft Process for operating a remote-controllable central locking installation of a vehicle
5481481, Nov 23 1992 Architectural Energy Corporation Automated diagnostic system having temporally coordinated wireless sensors
5526871, Feb 08 1994 Quick connect diagnostic apparatus and method for a vehicle cooling system
5541585, Oct 11 1994 PREMDOR INTERNATIONAL INC ; Masonite International Corporation Security system for controlling building access
5591950, Nov 04 1992 ONITY, INC Programmable electronic lock
5594429, Oct 27 1993 ALPS ELECTRIC CO , LTD Transmission and reception system and signal generation method for same
5604804, Apr 23 1996 ASSA ABLOY AB Method for certifying public keys in a digital signature scheme
5610982, May 15 1996 ASSA ABLOY AB Compact certification with threshold signatures
5631825, Sep 29 1993 DOW BENELUX N V Operator station for manufacturing process control system
5640151, Jun 15 1990 OL SECURITY LIMITED LIABILITY COMPANY Communication system for communicating with tags
5644302, Dec 27 1994 Najib, Hana Device for remotely changing the set temperature of a thermostat
5663957, Jul 12 1995 Ericsson Inc. Dual mode satellite/cellular terminal
5666416, Nov 16 1995 ASSA ABLOY AB Certificate revocation system
5717757, Aug 29 1996 ASSA ABLOY AB Certificate issue lists
5717758, Nov 02 1995 ASSA ABLOY AB Witness-based certificate revocation system
5717759, Apr 23 1996 ASSA ABLOY AB Method for certifying public keys in a digital signature scheme
5732691, Oct 30 1996 Rheem Manufacturing Company Modulating furnace with two-speed draft inducer
5774058, Jul 20 1995 Vindicator Corporation Remote access system for a programmable electronic lock
5778256, Mar 24 1993 NISEL INVESTMENTS LIMITED LIABILITY COMPANY PDA having a separate infrared generating device connected to its printer port for controlling home appliances
5793868, Aug 29 1996 ASSA ABLOY AB Certificate revocation system
5914875, Jan 11 1996 Kabushiki Kaisha Toshiba Method and apparatus for diagnosing plant anomaly
5915473, Jan 29 1997 Trane International Inc Integrated humidity and temperature controller
5923328, Aug 07 1996 Microsoft Technology Licensing, LLC Method and system for displaying a hierarchical sub-tree by selection of a user interface element in a sub-tree bar control
5923817, Feb 23 1996 Mitsubishi Denki Kabushiki Kaisha Video data system with plural video data recording servers storing each camera output
5927398, Jun 22 1996 Carrier Corporation Device identification system for HVAC communication network
5930773, Dec 17 1997 ENGIE INSIGHT SERVICES INC Computerized resource accounting methods and systems, computerized utility management methods and systems, multi-user utility management methods and systems, and energy-consumption-based tracking methods and systems
5960083, Oct 24 1995 ASSA ABLOY AB Certificate revocation system
5973613, Jun 15 1990 OL SECURITY LIMITED LIABILITY COMPANY Personal messaging system and method
5992194, Dec 20 1995 VDO Adolf Schindling AG Device for unlocking doors
6072402, Jan 09 1992 GE SECURITY, INC Secure entry system with radio communications
6097811, Nov 02 1995 ASSA ABLOY AB Tree-based certificate revocation system
6104963, Apr 03 1998 Johnson Controls Technology Company Communication system for distributed-object building automation system
6119125, Apr 03 1998 Johnson Controls Technology Company Software components for a building automation system based on a standard object superclass
6141595, Apr 03 1998 Johnson Controls Technology Company Common object architecture supporting application-centric building automation systems
6149065, Oct 28 1998 BURNER SYSTEMS INTERNATIONAL, INC Modulating thermostat for gas oven burner
6154681, Apr 03 1998 Johnson Controls Technology Company Asynchronous distributed-object building automation system with support for synchronous object execution
6167316, Apr 03 1998 Johnson Controls Technology Company Distributed object-oriented building automation system with reliable asynchronous communication
6233954, Apr 28 1999 Ingersoll-Rand Company Method for controlling the operation of a compression system having a plurality of compressors
6241156, Feb 01 2000 Acutherm L.P. Process and apparatus for individual adjustment of an operating parameter of a plurality of environmental control devices through a global computer network
6249755, May 25 1994 VMWARE, INC Apparatus and method for event correlation and problem reporting
6260765, Feb 25 2000 SECUREALERT, INC Remotely controllable thermostat
6268797, Mar 15 2000 Detection Systems, Inc. Integrated portable tracking signal and access authorization signal generator
6292893, Oct 24 1995 ASSA ABLOY AB Certificate revocation system
6301659, Nov 02 1995 ASSA ABLOY AB Tree-based certificate revocation system
6318137, Apr 08 1998 David, Chaum Electronic lock that can learn to recognize any ordinary key
6324854, Nov 22 2000 Copeland Corporation Air-conditioning servicing system and method
6334121, May 04 1998 Virginia Commonwealth University Usage pattern based user authenticator
6347374, Jun 05 1998 INTRUSION INC Event detection
6366558, May 02 1997 Cisco Technology, Inc Method and apparatus for maintaining connection state between a connection manager and a failover device
6369719, Oct 28 1996 FIRST-CLASS MONITORING, LLC Apparatus and method for collecting and transmitting utility meter data and other information via a wireless network
6374356, Jun 17 1998 Integral Technologies, Inc Shared intelligence automated access control system
6393848, Feb 01 2000 LG Electronics Inc. Internet refrigerator and operating method thereof
6394359, Jul 12 2000 Remote control thermostat
6424068, Jun 27 1997 Asahi Kogaku Kogyo Kabushiki Kaisha Galvano mirror unit
6453426, Mar 26 1999 Microsoft Technology Licensing, LLC Separately storing core boot data and cluster configuration data in a server cluster
6453687, Jan 07 2000 Robertshaw Controls Company Refrigeration monitor unit
6483697, May 29 2001 Qualcomm Incorporated Nested flip cover lid for a hand-held computing system
6487658, Oct 02 1995 ASSA ABLOY AB Efficient certificate revocation
6490610, May 30 1997 Oracle International Corporation Automatic failover for clients accessing a resource through a server
6496575, Jun 08 1998 GATESPACE NETWORKS, INC Application and communication platform for connectivity based services
6516357, Feb 08 1998 International Business Machines Corporation System for accessing virtual smart cards for smart card application and data carrier
6518953, Oct 31 1997 ANASCAPE, LTD Analog controls housed with electronic displays for remote controllers having feedback display screens
6546419, May 07 1998 SAMSUNG ELECTRONICS CO , LTD , A KOREAN CORP Method and apparatus for user and device command and control in a network
6556899, Aug 17 2000 New Flyer Industries Canada ULC Bus diagnostic and control system and method
6574537, Feb 05 2001 Boeing Company, the Diagnostic system and method
6583712, Jan 06 1999 MAS-HAMILTON GROUP, INC , A KENTUCKY CORPORATION Supervisor and subordinate lock system
6604023, Apr 28 2000 International Business Machines Corporation Managing an environment utilizing a portable data processing system
6615594, Mar 27 2001 Copeland Corporation Compressor diagnostic system
6628997, Apr 28 2000 Carrier Corporation Method for programming a thermostat
6647317, Sep 06 2000 HITACHI GLOBAL LIFE SOLUTIONS, INC Air conditioner management system
6647400, Aug 30 1999 Symantec Corporation System and method for analyzing filesystems to detect intrusions
6658373, May 11 2001 MCLOUD TECHNOLOGIES USA INC Apparatus and method for detecting faults and providing diagnostics in vapor compression cycle equipment
6663010, Jan 22 2001 ArvinMeritor Technology, LLC Individualized vehicle settings
6665669, Jan 03 2000 Simon Fraser University Methods and system for mining frequent patterns
6667690, Jan 22 2002 Carrier Corporation System and method for configuration of HVAC network
6741915, Aug 22 2001 MMI CONTROLS, LTD Usage monitoring HVAC control system
6758051, Mar 27 2001 Copeland Corporation Method and system for diagnosing a cooling system
6766450, Oct 24 1995 ASSA ABLOY AB Certificate revocation system
6789739, Feb 13 2002 ROSEN TECHNOLOGIES LLC Thermostat system with location data
6796494, Jun 18 1999 REMBRANDT TRADING, LP Method and system for configuring a publicly accessible computer system
6801849, Jul 13 2001 BRP US INC Engine diagnostic via PDA
6801907, Apr 10 2000 Security Identification Systems Corporation System for verification and association of documents and digital images
6826454, Sep 19 2001 Air conditioning diagnostic analyzer
6829332, Mar 28 1997 GOOGLE LLC Personal dial tone service with personalized call waiting
6851621, Aug 18 2003 Honeywell International Inc PDA diagnosis of thermostats
6871193, Nov 29 2000 GOOGLE LLC Method and system for partitioned service-enablement gateway with utility and consumer services
6886742, Aug 09 1999 First Data Corporation; The Western Union Company Systems and methods for deploying a point-of sale device
6895215, Dec 28 2000 International Business Machines Corporation Method and apparatus for transferring correspondence information
6910135, Jul 07 1999 Raytheon BBN Technologies Corp Method and apparatus for an intruder detection reporting and response system
6967612, Oct 22 2004 RAPISCAN LABORATORIES, INC System and method for standoff detection of human carried explosives
6969542, Dec 20 2000 Merck Patent GmbH Liquid-crystal medium, and electro-optical display containing same
6970070, May 08 2003 EMC IP HOLDING COMPANY LLC Method and apparatus for selective blocking of radio frequency identification devices
6973410, May 15 2001 Chillergy Systems, LLC Method and system for evaluating the efficiency of an air conditioning apparatus
6983889, Mar 21 2003 EMME E2MS, LLC Forced-air zone climate control system for existing residential houses
6989742, Dec 25 2001 NAMIC VA, INC Device and system for detecting abnormality
7004401, Aug 10 2001 TARGETED GRAIN MANAGEMENT, INC System and method for regulating agriculture storage facilities in order to promote uniformity among separate storage facilities
7019614, Feb 07 1995 Schlage Lock Company LLC; Harrow Products LLC Door security system audit trail
7032114, Aug 30 2000 Symantec Corporation System and method for using signatures to detect computer intrusions
7055759, Aug 18 2003 Honeywell International Inc PDA configuration of thermostats
7076083, Dec 12 2002 Eastman Kodak Company Personnel access control system
7117356, May 21 2002 BIO-key International, Inc.; BIO-KEY INTERNATIONAL, INC Systems and methods for secure biometric authentication
7124943, Sep 24 2004 ASSA ABLOY AB RFID system having a field reprogrammable RFID reader
7130719, Mar 28 2002 Invensys Systems, Inc System and method of controlling an HVAC system
7183894, Jul 31 2002 Sony Corporation Communication system for accessing shared entrance of multiple dwelling house
7203962, Aug 30 1999 Symantec Corporation System and method for using timestamps to detect attacks
7205882, Nov 10 2004 ASSA ABLOY AB Actuating a security system using a wireless device
7216007, Jul 06 2005 Honeywell International Inc. System and method for providing direct web access to controllers in a process control environment
7216015, Aug 22 2001 MMI Controls, LTD. HVAC control system
7218243, Jul 23 1998 Universal Electronics Inc. System and method for automatically setting up a universal remote control
7222800, Aug 18 2003 Honeywell International Inc. Controller customization management system
7233243, Jan 09 2004 CTRL Systems, Inc. Method of defense-in-depth ultrasound intrusion detection
7243001, Jun 15 2004 Amazon Technologies, Inc Time-based warehouse movement maps
7245223, Nov 20 2002 Anti terrorist and homeland security public safety warning system
7250853, Dec 10 2004 Honeywell International Inc Surveillance system
7274676, Jul 14 2003 Honeywell International Inc Burst-mode weighted sender scheduling for ad-hoc wireless medium access control protocols
7283489, Mar 31 2003 Lucent Technologies Inc Multimedia half-duplex sessions with individual floor controls
7313819, Jul 20 2001 Intel Corporation Automated establishment of addressability of a network device for a target network environment
7321784, Oct 24 2001 Texas Instruments Incorporated Method for physically updating configuration information for devices in a wireless network
7337315, Oct 02 1995 ASSA ABLOY AB Efficient certificate revocation
7343265, Nov 23 2005 Lockheed Martin Corporation System to monitor the health of a structure, sensor nodes, program product, and related methods
7353396, Oct 02 1995 ASSA ABLOY AB Physical access control
7362210, Sep 05 2003 Honeywell International Inc. System and method for gate access control
7376839, May 04 2001 Cubic Corporation Smart card access control system
7379997, Jul 28 2003 Invensys Systems, Inc System and method of controlling delivery and/or usage of a commodity
7380125, May 22 2003 KYNDRYL, INC Smart card data transaction system and methods for providing high levels of storage and transmission security
7383158, Apr 16 2002 Trane International Inc HVAC service tool with internet capability
7397371, Jan 31 2005 Honeywell International Inc. Security system access control and method
7408925, Mar 31 2004 ARLINGTON TECHNOLOGIES, LLC Originator based directing and origination call processing features for external devices
7468662, Jun 16 2006 GENERAC HOLDINGS INC ; GENERAC POWER SYSTEMS, INC Method for spatio-temporal event detection using composite definitions for camera systems
7487538, Nov 19 2001 Security system
7505914, Aug 06 2001 Ecolab USA Inc Method and system for providing advisory information to a field service provider
7542867, Aug 14 2001 National Instruments Corporation Measurement system with modular measurement modules that convey interface information
7543327, Nov 21 2003 ARECONT VISION COSTAR, LLC Video surveillance system based on high resolution network cameras capable of concurrent transmission of multiple image formats at video rates
7574734, Aug 15 2002 ACTIVIDENTITY, INC System and method for sequentially processing a biometric sample
7576770, Feb 11 2003 TELESIS GROUP, INC , THE; E-WATCH, INC System for a plurality of video cameras disposed on a common network
7583401, Jun 27 2002 Snap-On Incorporated Portal for distributing business and product information
7586398, Jul 23 1998 Universal Electronics, Inc. System and method for setting up a universal remote control
7600679, Jul 11 2007 Honeywell International Inc. Automatic guidance of visitor in new facility through access control system integration with LCD display
7620644, Oct 19 2004 Microsoft Technology Licensing, LLC Reentrant database object wizard
7634662, Nov 21 2002 TELESIS GROUP, INC , THE; E-WATCH, INC Method for incorporating facial recognition technology in a multimedia surveillance system
7644367, May 16 2003 Microsoft Technology Licensing, LLC User interface automation framework classes and interfaces
7661603, Dec 10 2002 LG Electronics Inc. Central control system and method for controlling air conditioners
7680814, Oct 16 2002 Microsoft Technology Licensing, LLC Navigating media content by groups
7683940, Sep 12 2003 Canon Kabushiki Kaisha Streaming non-continuous video data
7735132, Jul 29 2005 Malikie Innovations Limited System and method for encrypted smart card PIN entry
7735145, Feb 18 2005 Microsoft Technology Licensing, LLC Portable secure media with timed erasure
7794536, Oct 22 2007 Canon Kabushiki Kaisha Method for manufacturing pigment dispersion
7797640, Sep 29 2005 Robert Bosch GmbH Method for generating a flexible display field for a video surveillance system
7801870, Oct 26 2006 Samsung Electronics Co., Ltd. Method of synchronizing information shared between a plurality of universal plug and play devices and apparatus therefor
7818026, Nov 24 2003 Nokia Technologies Oy Configuration of a terminal
7839926, Nov 17 2000 E-WATCH, INC Bandwidth management and control
7853987, Oct 10 2006 Honeywell International Inc. Policy language and state machine model for dynamic authorization in physical access control
7861314, Aug 10 2000 Shield Security Systems, LLC Interactive key control system and method of managing access to secured locations
7873441, Sep 25 2006 System for execution of a load operating plan for load control
7907753, Mar 08 2002 HONEYWELL SILENT WITNESS INC Access control system with symbol recognition
7937669, Jun 12 2007 Honeywell International Inc.; International Business Machines Corporation; Honeywell International, Inc Access control system with rules engine architecture
7983892, May 20 2008 Honeywell International Inc. System and method for accessing and presenting health information for field devices in a process control system
7995526, Apr 23 2008 Honeywell International Inc. Apparatus and method for medium access control in wireless communication networks
7999847, May 08 2007 TIERRA VISTA GROUP, LLC; SECURENET SOLUTIONS GROUP, LLC Audio-video tip analysis, storage, and alerting system for safety, security, and business productivity
8045960, May 31 2007 Honeywell International Inc. Integrated access control system and a method of controlling the same
8069144, Nov 15 2001 Malikie Innovations Limited System and methods for asynchronous synchronization
8089341, Nov 02 2004 DAI NIPPON PRINTING CO , LTD Management system
8095889, May 12 2008 Honeywell International Inc. Heuristic and intuitive user interface for access control systems
8102240, Dec 27 2007 Honeywell International, Inc Controller providing shared device access for access control systems
8108200, May 20 2008 Honeywell International Inc. System and method for accessing and configuring field devices in a process control system using distributed control components
8166532, Oct 10 2006 Honeywell International Inc. Decentralized access control framework
8174572, Mar 25 2005 Tyco Fire & Security GmbH Intelligent camera selection and object tracking
8179227, Nov 08 2007 Honeywell International Inc. Employing external storage devices as media for access control panel control information
8199196, Sep 27 2007 RPX Corporation Method and apparatus for controlling video streams
8222990, Dec 12 2008 Honeywell International Inc. Hybrid access control system and method for controlling the same
8232860, Oct 21 2005 Honeywell International Inc. RFID reader for facility access control and authorization
8272053, Dec 18 2003 HONEYWELL INTERNATIONAL IN C Physical security management system
8316407, Apr 04 2006 Honeywell International Inc Video system interface kernel
8341695, May 01 2008 Honeywell International Inc.; Honeywell International Inc Method of access control implemented in an Ethernet switch
8350666, Oct 15 2008 Honeywell International Inc.; Honeywell International Inc Apparatus and method for location-based access control in wireless networks
8351350, May 28 2007 Honeywell International Inc Systems and methods for configuring access control devices
8479029, Sep 11 2007 Intel Corporation Methods and apparatuses for reducing step loads of processors
8509987, Nov 11 2009 BRIDGESTONE MOBILITY SOLUTIONS B V Methods and apparatus for automatic internet logging and social comparison of vehicular driving behavior
8560970, Dec 20 2007 Canon Kabushiki Kaisha Hierarchical tag based browsing of media collections
8605151, Sep 21 2007 UTC Fire & Security Americas Corporation, Inc Methods and systems for operating a video surveillance system
8922658, Nov 05 2010 RAZBERI TECHNOLOGIES, INC Network video recorder system
9036828, Dec 19 2009 Bayerische Motoren Werke Aktiengesellschaft Method for outputting music information in a vehicle
20020011923,
20020016971,
20020022991,
20020046337,
20020118096,
20020121961,
20020165824,
20020170064,
20030033230,
20030071714,
20030174049,
20030208689,
20030233432,
20040061714,
20040062421,
20040064453,
20040068583,
20040087362,
20040125124,
20040205350,
20040267694,
20050138380,
20050200714,
20060017939,
20060077262,
20070109098,
20070132550,
20070171862,
20070268145,
20070272744,
20080037432,
20080173709,
20090018900,
20090063517,
20090080443,
20090086692,
20090097815,
20090168695,
20090258643,
20090266885,
20090292995,
20090328203,
20100026811,
20100036511,
20100220715,
20100269173,
20110043631,
20110071929,
20110115602,
20110133884,
20110153791,
20110167488,
20110181414,
20120096131,
20120106915,
20120121229,
20120133482,
20120326868,
20130010111,
20130036356,
20150363478,
CA2240881,
CN1265762,
DE19945861,
EP43270,
EP122244,
EP152678,
EP629940,
EP858702,
EP1339028,
EP1630639,
GB2251266,
GB2390705,
JP2003074942,
JP2003240318,
JP6019911,
WO11592,
WO76220,
WO142598,
WO157489,
WO160024,
WO2091311,
WO232045,
WO3090000,
WO2004092514,
WO2005038727,
WO2006021047,
WO2006049181,
WO2006126974,
WO2007043798,
WO2008045918,
WO2008144803,
WO2010039598,
WO2010106474,
WO8402786,
WO9419912,
WO9627858,
/
Executed onAssignorAssigneeConveyanceFrameReelDoc
Jun 11 2019Honeywell International Inc.(assignment on the face of the patent)
Date Maintenance Fee Events
Jun 11 2019BIG: Entity status set to Undiscounted (note the period is included in the code).
May 28 2024M1551: Payment of Maintenance Fee, 4th Year, Large Entity.


Date Maintenance Schedule
Dec 08 20234 years fee payment window open
Jun 08 20246 months grace period start (w surcharge)
Dec 08 2024patent expiry (for year 4)
Dec 08 20262 years to revive unintentionally abandoned end. (for year 4)
Dec 08 20278 years fee payment window open
Jun 08 20286 months grace period start (w surcharge)
Dec 08 2028patent expiry (for year 8)
Dec 08 20302 years to revive unintentionally abandoned end. (for year 8)
Dec 08 203112 years fee payment window open
Jun 08 20326 months grace period start (w surcharge)
Dec 08 2032patent expiry (for year 12)
Dec 08 20342 years to revive unintentionally abandoned end. (for year 12)