An application and network analytics platform can capture telemetry from servers and network devices operating within a network. The application and network analytics platform can determine an application dependency map (ADM) for an application executing in the network. Using the ADM, the application and network analytics platform can resolve flows into flowlets of various granularities, and determine baseline metrics for the flowlets. The baseline metrics can include transmission times, processing times, and/or data sizes for the flowlets. The application and network analytics platform can compare new flowlets against the baselines to assess availability, load, latency, and other performance metrics for the application. In some implementations, the application and network analytics platform can automate remediation of unavailability, load, latency, and other application performance issues.
|
1. A method comprising:
processing telemetry data for a plurality of flows associated with a set of service instances in a network, the set of service instances residing in a data center, the telemetry data received from one or more sensors installed in the data center;
generating, based on the processed telemetry data, an application dependency map for an application executing in the network, the application dependency map indicating dependencies between service instances of the set of service instances in the network, each service instance implementing one or more processes associated with the application;
determining one or more metrics associated with requests and responses transmitted between at least a first service instance and a second service instance of the application dependency map;
comparing the determined one or more metrics to respective ranges; and
responsive to detecting a deviation of at least one of the one or more metrics from a corresponding respective range, initiating one or more remediation actions, at least one of the one or more remediation actions comprising instantiating one or more new service instances associated with the application in a public cloud remote from the data center,
wherein at least a second one of the one or more remediation actions comprises instantiating one or more new service instances associated with the application in the data center.
13. A non-transitory computer-readable medium having instructions that, upon being executed by one or more processors, cause the one or more processors to:
process telemetry data for a plurality of flows associated with a set of service instances in a network, the set of service instances residing in a data center, the telemetry data received from one or more sensors installed in the data center;
generate, based on the processed telemetry data, an application dependency map for an application executing in the network, the application dependency map indicating dependencies between service instances of the set of service instances in the network, each service instance implementing one or more processes associated with the application;
determine one or more metrics associated with requests and responses transmitted between at least a first service instance and a second service instance of the application dependency map;
compare the determined one or more metrics to respective ranges; and
responsive to detecting a deviation of at least one of the one or more metrics from a corresponding respective range, initiate one or more remediation actions, at least one of the one or more remediation actions comprising instantiating one or more new service instances associated with the application in a public cloud remote from the data center,
wherein at least a second one of the one or more remediation actions comprises instantiating one or more new service instances associated with the application in the data center.
7. A system comprising:
one or more processors; and
memory including instructions that, upon being executed by the one or more processors, cause the system to:
process telemetry data for a plurality of flows associated with a set of service instances in a network, the set of service instances residing in a data center, the telemetry data received from one or more sensors installed in the data center;
generate, based on the processed telemetry data, an application dependency map for an application executing in the network, the application dependency map indicating dependencies between service instances of the set of service instances in the network, each service instance implementing one or more processes associated with the application;
determine one or more metrics associated with requests and responses transmitted between at least a first service instance and a second service instance of the application dependency map;
compare the determined one or more metrics to respective ranges; and
responsive to detecting a deviation of at least one of the one or more metrics from a corresponding respective range, initiate one or more remediation actions, at least one of the one or more remediation actions comprising instantiating one or more new service instances associated with the application in a public cloud remote from the data center,
wherein at least a second one of the one or more remediation actions comprises instantiating one or more new service instances associated with the application in the data center.
2. The method of
3. The method of
4. The method of
5. The method of
6. The method of
8. The system of
9. The system of
10. The system of
11. The system of
12. The system of
14. The non-transitory computer-readable medium of
15. The non-transitory computer-readable medium of
16. The non-transitory computer-readable medium of
17. The non-transitory computer-readable medium of
18. The non-transitory computer-readable medium of
|
This application is a continuation of U.S. Non-Provisional patent application Ser. No. 17/094,815, filed on Nov. 11, 2020, which is a continuation of U.S. Non-Provisional patent application Ser. No. 15/471,183, filed on Mar. 28, 2017, now U.S. Pat. No. 10,873,794, the full disclosures of each are hereby expressly incorporated by reference in their entireties.
The subject matter of this disclosure relates in general to the field of computer networks, and more specifically for application performance monitoring and management.
An application performance management or application performance monitoring (APM) system is a tool that an enterprise can use to assess the states of applications operating in the enterprise's data center, such as availability, load, latency, and other performance issues that may arise for the applications. Some conventional networks rely on an application's built-in monitoring utilities but not every application includes these utilities. Moreover, taking such an ad-hoc approach for application monitoring may be untenable to manage for a typical enterprise network with hundreds or thousands of applications. Certain conventional APM platforms may require injecting code into an application to integrate the application with these platforms. This is a non-trivial task that can require careful coordination between the development team and the team tasked with managing the application thereafter. This can also affect the performance of the application because logging and other monitoring features occur directly in the execution path of the application. This approach is also inadequate for third party applications and/or applications deployed off-premises (e.g., via a software as a service (SaaS) or public cloud provider) that the enterprise cannot modify. Some conventional APM systems use measurement hardware and/or software on servers hosting applications while others may rely on similar tools on network devices. Both approaches can negatively affect performance of the hosting hardware and/or software. In addition, these techniques are incomplete as they neglect network performance over application performance or vice versa.
Overview
An application and network analytics platform can capture telemetry (e.g., flow data, server data, process data, user data, policy data, etc.) from servers (e.g., physical and/or virtual servers) (sometimes also referred to as hosts, endpoints, computing devices, machines, and the like) and network devices (e.g., switches, routers, hubs, etc.) operating within a network. The application and network analytics platform can determine an application dependency map (ADM) for one or more applications executing in the network. Using the ADM, the application and network analytics platform can resolve flows into flowlets of varying granularities, and determine baseline metrics for the flowlets. The baseline metrics can include transmission times, processing times, and/or data sizes (in number of packets and/or bytes) for the flowlets. The application and network analytics platform can compare new flowlets against the baselines to assess availability, load, latency, and other performance metrics for one or more applications. In some embodiments, the application and network analytics platform can automate or substantially automate remediation of unavailability, load, latency, and other application performance issues.
Description
The data collection layer 110 may include software sensors 112, hardware sensors 114, and customer/third party data sources 116. The software sensors 112 can run within servers of a network, such as physical or bare-metal servers; hypervisors, virtual machine monitors, container orchestrators, or other virtual entity managers; virtual machines, containers, or other virtual entities. The hardware sensors 114 can reside on the application-specific integrated circuits (ASICs) of switches, routers, or other network devices (e.g., packet capture (pcap) appliances such as a standalone packet monitor, a device connected to a network device's monitoring port, a device connected in series along a main trunk of a data center, or similar device). The software sensors 112 can capture telemetry (e.g., flow data, server data, process data, user data, policy data, etc.) from the servers and the hardware sensors 114 can capture network telemetry (e.g., flow data) from network devices, and send the telemetry to the analytics engine 120 for further processing. For example, the software sensors 112 can sniff packets sent over their hosts' physical or virtual network interface cards (NICs), or individual processes on each server can report the telemetry to the software sensors 112. The hardware sensors 114 can capture network telemetry at line rate from all ports of the network devices hosting the hardware sensors.
As discussed, the input forwarding controller 214 may perform several operations on an incoming packet, including parsing the packet header, performing an L2 lookup, performing an L3 lookup, processing an ingress access control list (ACL), classifying ingress traffic, and aggregating forwarding results. Although describing the tasks performed by the input forwarding controller 214 in this sequence, one of ordinary skill will understand that, for any process discussed herein, there can be additional, fewer, or alternative steps performed in similar or alternative orders, or in parallel, within the scope of the various embodiments unless otherwise stated.
In some embodiments, when a unicast packet enters through a front-panel port (e.g., a port of ingress MAC 212), the input forwarding controller 214 may first perform packet header parsing. For example, the input forwarding controller 214 may parse the first 128 bytes of the packet to extract and save information such as the L2 header, EtherType, L3 header, and TCP IP protocols.
As the packet goes through the ingress forwarding pipeline 210, the packet may be subject to L2 switching and L3 routing lookups. The input forwarding controller 214 may first examine the destination MAC address of the packet to determine whether to switch the packet (i.e., L2 lookup) or route the packet (i.e., L3 lookup). For example, if the destination MAC address matches the network device's own MAC address, the input forwarding controller 214 can perform an L3 routing lookup. If the destination MAC address does not match the network device's MAC address, the input forwarding controller 214 may perform an L2 switching lookup based on the destination MAC address to determine a virtual LAN (VLAN) identifier. If the input forwarding controller 214 finds a match in the MAC address table, the input forwarding controller 214 can send the packet to the egress port. If there is no match for the destination MAC address and VLAN identifier, the input forwarding controller 214 can forward the packet to all ports in the same VLAN.
During L3 routing lookup, the input forwarding controller 214 can use the destination IP address for searches in an L3 host table. This table can store forwarding entries for directly attached hosts and learned/32 host routes. If the destination IP address matches an entry in the host table, the entry will provide the destination port, next-hop MAC address, and egress VLAN. If the input forwarding controller 214 finds no match for the destination IP address in the host table, the input forwarding controller 214 can perform a longest-prefix match (LPM) lookup in an LPM routing table.
In addition to forwarding lookup, the input forwarding controller 214 may also perform ingress ACL processing on the packet. For example, the input forwarding controller 214 may check ACL ternary content-addressable memory (TCAM) for ingress ACL matches. In some embodiments, each ASIC may have an ingress ACL TCAM table of 4000 entries per slice to support system internal ACLs and user-defined ingress ACLs. These ACLs can include port ACLs, routed ACLs, and VLAN ACLs, among others. In some embodiments, the input forwarding controller 214 may localize the ACL entries per slice and program them only where needed.
In some embodiments, the input forwarding controller 214 may also support ingress traffic classification. For example, from an ingress interface, the input forwarding controller 214 may classify traffic based on the address field, IEEE 802.1q class of service (CoS), and IP precedence or differentiated services code point in the packet header. In some embodiments, the input forwarding controller 214 can assign traffic to one of eight quality-of-service (QoS) groups. The QoS groups may internally identify the traffic classes used for subsequent QoS processes as packets traverse the system.
In some embodiments, the input forwarding controller 214 may collect the forwarding metadata generated earlier in the pipeline (e.g., during packet header parsing, L2 lookup, L3 lookup, ingress ACL processing, ingress traffic classification, forwarding results generation, etc.) and pass it downstream through the input data path controller 216. For example, the input forwarding controller 214 can store a 64-byte internal header along with the packet in the packet buffer. This internal header can include 16 bytes of iETH (internal communication protocol) header information, which the input forwarding controller 214 can prepend to the packet when transferring the packet to the output data path controller 222 through the broadcast network 230. The network device can strip the 16-byte iETH header when the packet exits the front-panel port of the egress MAC 226. The network device may use the remaining internal header space (e.g., 48 bytes) to pass metadata from the input forwarding queue to the output forwarding queue for consumption by the output forwarding engine.
In some embodiments, the input data path controller 216 can perform ingress accounting functions, admission functions, and flow control for a no-drop class of service. The ingress admission control mechanism can determine whether to admit the packet into memory based on the amount of buffer memory available and the amount of buffer space already used by the ingress port and traffic class. The input data path controller 216 can forward the packet to the output data path controller 222 through the broadcast network 230.
As discussed, in some embodiments, the broadcast network 230 can comprise a set of point-to-multipoint wires that provide connectivity between all slices of the ASIC. The input data path controller 216 may have a point-to-multipoint connection to the output data path controller 222 on all slices of the network device, including its own slice.
In some embodiments, the output data path controller 222 can perform egress buffer accounting, packet queuing, scheduling, and multicast replication. In some embodiments, all ports can dynamically share the egress buffer resource. In some embodiments, the output data path controller 222 can also perform packet shaping. In some embodiments, the network device can implement a simple egress queuing architecture. For example, in the event of egress port congestion, the output data path controller 222 can directly queue packets in the buffer of the egress slice. In some embodiments, there may be no virtual output queues (VoQs) on the ingress slice. This approach can simplify system buffer management and queuing.
As discussed, in some embodiments, one or more network devices can support up to 10 traffic classes on egress, 8 user-defined classes identified by QoS group identifiers, a CPU control traffic class, and a switched port analyzer (SPAN) traffic class. Each user-defined class can have a unicast queue and a multicast queue per egress port. This approach can help ensure that no single port will consume more than its fair share of the buffer memory and cause buffer starvation for other ports.
In some embodiments, multicast packets may go through similar ingress and egress forwarding pipelines as the unicast packets but instead use multicast tables for multicast forwarding. In addition, multicast packets may go through a multistage replication process for forwarding to multiple destination ports. In some embodiments, the ASIC can include multiple slices interconnected by a non-blocking internal broadcast network. When a multicast packet arrives at a front-panel port, the ASIC can perform a forwarding lookup. This lookup can resolve local receiving ports on the same slice as the ingress port and provide a list of intended receiving slices that have receiving ports in the destination multicast group. The forwarding engine may replicate the packet on the local ports, and send one copy of the packet to the internal broadcast network, with the bit vector in the internal header set to indicate the intended receiving slices. In this manner, only the intended receiving slices may accept the packet off of the wire of the broadcast network. The slices without receiving ports for this group can discard the packet. The receiving slice can then perform local L3 replication or L2 fan-out lookup and replication to forward a copy of the packet to each of its local receiving ports.
In
In addition to the traditional forwarding information, the flow cache 240 can also collect other metadata such as detailed IP and TCP flags and tunnel endpoint identifiers. In some embodiments, the flow cache 240 can also detect anomalies in the packet flow such as inconsistent TCP flags. The flow cache 240 may also track flow performance information such as the burst and latency of a flow. By providing this level of information, the flow cache 240 can produce a better view of the health of a flow. Moreover, because the flow cache 240 does not perform sampling, the flow cache 240 can provide complete visibility into the flow.
In some embodiments, the flow cache 240 can include an events mechanism to complement anomaly detection. This configurable mechanism can define a set of parameters that represent a packet of interest. When a packet matches these parameters, the events mechanism can trigger an event on the metadata that triggered the event (and not just the accumulated flow information). This capability can give the flow cache 240 insight into the accumulated flow information as well as visibility into particular events of interest. In this manner, networks, such as a network implementing the application and network analytics platform 100, can capture telemetry more comprehensively and not impact application and network performance.
Returning to
In some embodiments, the application and network analytics platform 100 can resolve flows into flowlets (i.e., sequences of requests and responses of a larger request and response) of various granularities. For example, a response to a request to an enterprise application may result in multiple sub-requests and sub-responses to various back-end services (e.g., authentication, static content, data, search, sync, etc.). The application and network analytics platform 100 can break a flow into its constituent components to provide greater insight into application and network performance. The application and network analytics platform 100 can perform this resolution in real time or substantially real time (e.g., no more than a few minutes after detecting the flow).
The application and network analytics platform 100 can associate a flow with a server sending or receiving the flow, an application or process triggering the flow, the owner of the application or process, and one or more policies applicable to the flow, among other telemetry. The telemetry captured by the software sensors 112 can thus include server data, process data, user data, policy data, and other data (e.g., virtualization information, tenant information, sensor information, etc.). The server telemetry can include the server name, network address, CPU usage, network usage, disk space, ports, logged users, scheduled jobs, open files, and similar information. In some embodiments, the server telemetry can also include information about the file system of the server, such as the lists of files (e.g., log files, configuration files, device special files, etc.) and/or directories stored within the file system as well as the metadata for the files and directories (e.g., presence, absence, or modifications of a file and/or directory). In some embodiments, the server telemetry can further include physical or virtual configuration information (e.g., processor type, amount of random access memory (RAM), amount of disk or storage, type of storage, system type (e.g., 32-bit or 64-bit), operating system, public cloud provider, virtualization platform, etc.).
The process telemetry can include the process name (e.g., bash, httpd, netstat, etc.), process identifier, parent process identifier, path to the process (e.g., /usr2/username/bin/, /usr/local/bin, /usr/bin, etc.), CPU utilization, memory utilization, memory address, scheduling information, nice value, flags, priority, status, start time, terminal type, CPU time taken by the process, and the command string that initiated the process (e.g., “/opt/tetration/collector/tet-collector--config_file/etc/tetration/collector/collector.config--timestamp_flow_info --logtostderr--utc_time_in_file_name_true--max_num_ssl_sw_sensors 63000-- enable_client_certificate true”). The user telemetry can include information regarding a process owner, such as the user name, user identifier, user's real name, e-mail address, user's groups, terminal information, login time, expiration date of login, idle time, and information regarding files and/or directories of the user.
The customer/third party data sources 116 can include out-of-band data such as power level, temperature, and physical location (e.g., room, row, rack, cage door position, etc.). The customer/third party data sources 116 can also include third party data regarding a server such as whether the server is on an IP watch list or security report (e.g., provided by Cisco®, Arbor Networks® of Burlington, Mass., Symantec® Corp. of Sunnyvale, Calif., Sophos® Group plc of Abingdon, England, Microsoft® Corp. of Seattle, Wash., Verizon® Communications, Inc. of New York, N.Y., among others), geolocation data, and Whois data, and other data from external sources.
In some embodiments, the customer/third party data sources 116 can include data from a configuration management database (CMDB) or configuration management system (CMS) as a service. The CMDB/CMS may transmit configuration data in a suitable format (e.g., JavaScript® object notation (JSON), extensible mark-up language (XML), yet another mark-up language (YAML), etc.)).
The processing pipeline 122 of the analytics engine 120 can collect and process the telemetry. In some embodiments, the processing pipeline 122 can retrieve telemetry from the software sensors 112 and the hardware sensors 114 every 100 ms or faster. Thus, the application and network analytics platform 100 may not miss or is much less likely than conventional systems (which typically collect telemetry every 60 seconds) to miss “mouse” flows. In addition, as the telemetry tables flush so often, the software sensors 112 and the hardware sensors 114 do not or are much less likely than conventional systems to drop telemetry because of overflow/lack of memory. An additional advantage of this approach is that the application and network analytics platform is responsible for flow-state tracking instead of network devices. Thus, the ASICs of the network devices of various embodiments can be simpler or can incorporate other features.
In some embodiments, the processing pipeline 122 can filter out extraneous or duplicative data or it can create summaries of the telemetry. In some embodiments, the processing pipeline 122 may process (and/or the software sensors 112 and hardware sensors 114 may capture) only certain types of telemetry and disregard the rest. For example, the processing pipeline 122 may process (and/or the sensors may monitor) only high-priority telemetry, telemetry associated with a particular subnet (e.g., finance department, human resources department, etc.), telemetry associated with a particular application (e.g., business-critical applications, compliance software, health care applications, etc.), telemetry from external-facing servers, etc. As another example, the processing pipeline 122 may process (and/or the sensors may capture) only a representative sample of telemetry (e.g., every 1,000th packet or other suitable sample rate).
Collecting and/or processing telemetry from multiple servers of the network (including within multiple partitions of virtualized hosts) and from multiple network devices operating between the servers can provide a comprehensive view of network behavior. The capture and/or processing of telemetry from multiple perspectives rather than just at a single device located in the data path (or in communication with a component in the data path) can allow the data to be correlated from the various data sources, which may be used as additional data points by the analytics engine 120.
In addition, collecting and/or processing telemetry from multiple points of view can enable capture of more accurate data. For example, a conventional network may consist of external-facing network devices (e.g., routers, switches, network appliances, etc.) such that the conventional network may not be capable of monitoring east-west traffic, including telemetry for VM-to-VM or container-to-container communications on a same host. As another example, the conventional network may drop some packets before those packets traverse a network device incorporating a sensor. The processing pipeline 122 can substantially mitigate or eliminate these issues altogether by capturing and processing telemetry from multiple points of potential failure. Moreover, the processing pipeline 122 can verify multiple instances of data for a flow (e.g., telemetry from a source (physical server, hypervisor, container orchestrator, other virtual entity manager, VM, container, and/or other virtual entity), one or more network devices, and a destination) against one another.
In some embodiments, the processing pipeline 122 can assess a degree of accuracy of telemetry for a single flow captured by multiple sensors and utilize the telemetry from a single sensor determined to be the most accurate and/or complete. The degree of accuracy can be based on factors such as network topology (e.g., a sensor closer to the source may be more likely to be more accurate than a sensor closer to the destination), a state of a sensor or a server hosting the sensor (e.g., a compromised sensor/server may have less accurate telemetry than an uncompromised sensor/server), or telemetry volume (e.g., a sensor capturing a greater amount of telemetry may be more accurate than a sensor capturing a smaller amount of telemetry).
In some embodiments, the processing pipeline 122 can assemble the most accurate telemetry from multiple sensors. For instance, a first sensor along a data path may capture data for a first packet of a flow but may be missing data for a second packet of the flow while the reverse situation may occur for a second sensor along the data path. The processing pipeline 122 can assemble data for the flow from the first packet captured by the first sensor and the second packet captured by the second sensor.
In some embodiments, the processing pipeline 122 can also disassemble or decompose a flow into sequences of request and response flowlets (e.g., sequences of requests and responses of a larger request or response) of various granularities. For example, a response to a request to an enterprise application may result in multiple sub-requests and sub-responses to various back-end services (e.g., authentication, static content, data, search, sync, etc.). The processing pipeline 122 can break a flow down into its constituent components to provide greater insight into application and network performance. The processing pipeline 122 can perform this resolution in real time or substantially real time (e.g., no more than a few minutes after detecting the flow).
The processing pipeline 122 can store the telemetry in a data lake (not shown), a large-scale storage repository characterized by massive storage for various types of data, enormous processing power, and the ability to handle nearly limitless concurrent tasks or jobs. In some embodiments, the analytics engine 120 may deploy at least a portion of the data lake using the Hadoop® Distributed File System (HDFS™) from Apache® Software Foundation of Forest Hill, Md. HDFS™ is a highly scalable and distributed file system that can scale to thousands of cluster nodes, millions of files, and petabytes of data. A feature of HDFS™ is its optimization for batch processing, such as by coordinating data computation to where data is located. Another feature of HDFS™ is its utilization of a single namespace for an entire cluster to allow for data coherency in a write-once, read-many access model. A typical HDFS™ implementation separates files into blocks, which are typically 64 MB in size and replicated in multiple data nodes. Clients can access data directly from the data nodes.
The processing pipeline 122 can propagate the processed data to one or more engines, monitors, and other components of the analytics engine 120 (and/or the components can retrieve the data from the data lake), such as an application dependency mapping (ADM) engine 124, an inventory monitor 126, a flow monitor 128, and an application performance monitoring (APM) engine 130.
The ADM engine 124 can determine dependencies of applications running in the network, i.e., how processes on different servers interact with one another to perform the functions of the application. Particular patterns of traffic may correlate with particular applications. The ADM engine 124 can evaluate telemetry processed by the processing pipeline 122 to determine the interconnectivity or dependencies of the application to generate a graph for the application (i.e., an application dependency mapping). For example, in a conventional three-tier architecture for a web application, first servers of the web tier, second servers of the application tier, and third servers of the data tier make up the web application. From flow data, the ADM engine 124 may determine that there is first traffic flowing between external servers on port 80 of the first servers corresponding to Hypertext Transfer Protocol (HTTP) requests and responses. The flow data may also indicate second traffic between first ports of the first servers and second ports of the second servers corresponding to application server requests and responses and third traffic flowing between third ports of the second servers and fourth ports of the third servers corresponding to database requests and responses. The ADM engine 124 may define an application dependency map or graph for this application as a three-tier application including a first endpoint group (EPG) (i.e., groupings of application tiers or clusters, applications, and/or application components for implementing forwarding and policy logic) comprising the first servers, a second EPG comprising the second servers, and a third EPG comprising the third servers.
The inventory monitor 126 can continuously track the network's assets (e.g., servers, network devices, applications, etc.) based on the telemetry processed by the processing pipeline 122. In some embodiments, the inventory monitor 126 can assess the state of the network at a specified interval (e.g., every 1 minute). That is, the inventory monitor 126 can periodically take snapshots of the states of applications, servers, network devices, and/or other elements of the network. In other embodiments, the inventory monitor 126 can capture the snapshots when events of interest occur, such as an application experiencing latency that exceeds an application latency threshold; the network experiencing latency that exceeds a network latency threshold; failure of a server, network device, or other network element; and similar circumstances. Snapshots can include a variety of telemetry associated with network elements. For example, a snapshot of a server can represent the processes executing on the server at a time of capture, the amount of CPU utilized by each process (e.g., as an amount of time and/or a relative percentage), the amount of memory utilized by each process (e.g., in bytes and/or as a relative percentage), the amount of disk utilized by each process (e.g., in bytes or as a relative percentage), and a distance (physical and/or logical, relative and/or absolute) from one or more other network elements.
In some embodiments, on a change to the network (e.g., a server updating its operating system or running a new process; a server communicating on a new port; a VM, container, or other virtualized entity migrating to a different host and/or subnet, VLAN, VxLAN, or other network segment; etc.), the inventory monitor 126 can alert the APM engine 130 to ensure that applications and the network remain performing as expected in view of the change(s) to the data center.
The flow monitor 128 can analyze flows to detect whether they are associated with anomalous or malicious traffic. In some embodiments, the flow monitor 128 may receive examples of past flows determined to perform at expectation (i.e., the length of time for the flow to reach a network device or a destination is within a threshold length of time) or perform below expectation (i.e., the length of time for the flow to reach the network device and/or destination exceeds the threshold length of time). The flow monitor 128 can utilize machine learning to analyze the telemetry processed by the processing pipeline 122 and classify each current flow based on similarity to past flows. On detection of an anomalous flow, such as a flow taking a shorter or longer duration from source to destination than a specified time range, a flow of a size less or more than a specified amount, or a flow previously classified as a network attack, the flow monitor 128 may transmit an alert to the APM engine 130 and/or to the presentation layer 140. In some embodiments, the network may operate within a trusted environment for a period of time so that the analytics engine 120 can establish a baseline of normal operation.
The APM engine 130 can be responsible for monitoring and managing applications running within the network. For example, the APM engine 130 may receive an alert from the inventory monitor 126 on a change to the network or an alert from the flow monitor upon the flow monitor 128 detecting an anomalous flow. The APM engine 130 can evaluate the telemetry to pinpoint the root cause of degraded performance. In some embodiments, the APM engine 130 can also perform remediation tasks, such as load balancing, migrating application components to be closer together (physically and/or logically), instantiating new server instances (in the private network and/or a public cloud), and disabling network connectivity for problematic servers, among other operations.
In some embodiments, the PFEs 310 can be responsible for storing platform-independent configuration information in memory, handling registration of the sensors 302, monitoring updates to the configuration information, distributing the updates to the sensors 302, and collecting telemetry captured by the sensors 302. In the example of
The coordinator cluster 320 can operate as the controller for the APM engine 300. In the example of
The telemetry store 330 can maintain sensor data captured by the sensors 302. In some embodiments, the APM engine can maintain recently captured and/or accessed telemetry in more readily-accessible data stores (e.g., solid state devices (SSD), optimized hard disk drives (HDD), etc.) and migrate older telemetry to less accessible data stores (e.g., commodity HDDs, tape, etc.). In some embodiments, the APM engine 300 may implement the telemetry store 330 using Druid® or other relational database platform. In other embodiments, the APM engine 300 may implement the telemetry store 330 using software provided by MongoDB®, Inc. of New York, N.Y. or other NoSQL database.
In some embodiments, the coordinator cluster 320 may also be responsible for load balancing the PFEs 310, ensuring high availability of the PFEs 310 to the sensors 302, and receiving and storing telemetry in the telemetry store 330. In other embodiments, the APM engine 300 can integrate the functionality of a PFE and a coordinator or further divide the functionality of the PFE and the coordinator into additional components.
Returning to
In some embodiments, the application and network analytics platform 100 can expose application programming interface (API) endpoints (e.g., such as those based on the simple object access protocol (SOAP), a service oriented architecture (SOA), a representational state transfer (REST) architecture, a resource oriented architecture (ROA), etc.) for monitor the performance of applications executing in a network and the network itself. In some embodiments, the application and network analytics platform 100 may implement the API endpoints 144 using Hadoop® Hive from Apache® for the back end, and Java® Database Connectivity (JDBC) from Oracle® Corporation of Redwood Shores, Calif., as an API layer. Hive is a data warehouse infrastructure that provides data summarization and ad hoc querying. Hive provides a mechanism to query data using a variation of structured query language (SQL) called HiveQL. JDBC is an application programming interface (API) for the programming language Java®, which defines how a client may access a database.
In some embodiments, the application and network analytics platform 100 may implement the event-based notification system using Hadoop® Kafka. Kafka is a distributed messaging system that supports partitioning and replication. Kafka uses the concept of topics. Topics are feeds of messages in specific categories. In some embodiments, Kafka can take raw packet captures and telemetry information as input, and output messages to a security information and event management (SIEM) platform that provides users with the capability to search, monitor, and analyze machine-generated data.
In some embodiments, each server in the network may include a software sensor and each network device may include a hardware sensor 114. In other embodiments, the software sensors 112 and hardware sensors 114 can reside on a portion of the servers and network devices of the network. In some embodiments, the software sensors 112 and/or hardware sensors 114 may operate in a full-visibility mode in which the sensors collect telemetry from every packet and every flow or a limited-visibility mode in which the sensors provide only the conversation view required for application insight and policy generation.
The network environment 400 can include a client computing device 402, a wide area network (WAN) 404, and a local area network (LAN) 410. Although not shown here for purposes of simplicity and conciseness, a typical data center may also include a firewall, a load balancer, and/or an additional edge router between an edge network device 406b and the web server 412. The client 402 can be any kind of computing device (i.e., of varying types, capabilities, operating systems, etc.) capable of communication over a network, such as a server (physical or virtual), a desktop computer, a laptop, a tablet, a smartphone, or a wearable device (e.g., a watch; eyeglasses, a visor, a head-mounted display or other device generally worn over a user's eyes; headphones, ear buds, or other device generally worn in or over a user's ears; etc.). The client 402 can also be an “infotainment system” (i.e., a computing device integrated with a means of transportation), a “smart” home device or Internet of Things (IoT) device (e.g., a television, a set-top box, a digital video recorder (DVR), a digital video disc (DVD) player or other media player, a video game console, etc.), or other electronic devices.
The WAN 404 can include one or more networks and/or network devices, such as the network devices 406a and 406b, for interconnecting the client 402 and the LAN 410. WANs can connect geographically dispersed nodes over long-distance communications links, such as common carrier telephone lines, optical light paths, synchronous optical networks (SONET), or synchronous digital hierarchy (SDH) links. LANs and WANs can include L2 and/or L3 networks and servers. The Internet is an example of a WAN that connects disparate networks throughout the world, providing global communication between nodes on various networks. The nodes typically communicate over the network by exchanging discrete frames or packets of data according to predefined protocols, such as the Transmission Control Protocol/Internet Protocol (TCP/IP). In this context, a protocol can refer to a set of rules defining how the nodes interact with each other. The WAN 404 can also be a private network, such as a global enterprise network, that operates using similar or the same technologies as the public Internet.
LANs can connect nodes over dedicated private communications links located in the same general physical location, such as a building or campus. In the example of
In the example of
In the example of
The web server 412 may confirm authentication of the user from the authentication response and begin retrieving content to provide a response to the originating request. The response may be a page of the website/web application that includes content from the content server 418 and personal information from the data server 420. The web server 412 may take up a time JK to prepare the request to the content server 418 and a time KS to prepare the request to the data server 420. The web server 412 may send the content request to the network device 414 over a time KL, upon which the network device 414 may spend a time LM to forward the request to the content server 418 for a duration MN. The content server 418 can receive the content request, take a time NO to process the request, and transmit a content response to the network device 414 over a time OP. The network device 414 can process the content response for a time PQ and forward the content response to the web server after which a time QR elapses.
In parallel or very near in time to the request/response to the content server 418, the web server 412 may send the data request to the network device 414 for a time ST. The network device can process the data request for a time TU and forward the data request over a period of time UV. The data server 420 may have an architecture such that it must retrieve requested data from the database 422, and therefore must perform some time VW processing the data request and preparing a database request (i.e., at point W). The database request may take a time WX to arrive at the database 422. Fetching the requested data may occur over a duration XY and transmitting the data (i.e., at point Y) back to the data server 420 may occur over a time YZ. The data server may process the database response within a time Zα before sending a data response back to the network device 414 over a time αβ. The network device may process the data response for a time βγ and forward the data response over a time γδ. The web server 412 may assemble the content retrieved from the content server 418 and the data retrieved from the data server 420 over a time δε before sending the response to the originating request to the client 402 over a time εζ.
In the example of
As shown in
A second data center that relies only on telemetry from servers may also suffer from various defects. Such a system may be able to detect anomalous (or no) response times in communications between the web server 412 and the data server 420 but may not be able to ascertain whether latency or failure is due to the web server 412, the network device 414, the data server 420, the database 422, or the connection between the data server and the database. For example, in the data request, the second data center may have a view of the data request and the data response but ST, TU, UV, VW, WX, XY, YZ, Zα, αβ, βγ, and γδ may be a black box. Various embodiments may overcome these and other deficiencies with prior art data center by capturing telemetry along multiple points of a data path, resolving flows into flowlets, establishing baselines for the flowlets, and evaluating new traffic against the baselines to more precisely diagnose the root cause of application and network performance degradation.
In the example of
After collection of the network telemetry and associated data, the method 600 may continue on to step 604, in which the application and network analytics platform can generate an application dependency map (ADM) for one or more applications executing in the network. In some embodiments, the nodes of the ADM can represent a pairing or concatenation of a server and a process (or application tier, application, application component, or other application granularity in other embodiments), and the edges of the ADM graph can represent the application and network analytics framework detecting flows between nodes. Clusters of nodes (i.e., host-process pairings within a similarity threshold of one another) can represent application components and each connected graph of the ADM can represent an application.
In some embodiments, generating the ADM can involve determining the nodes of the ADM by concatenating servers (e.g., the telemetry identifies a server as a source or destination in packet header metadata of a packet/flow) and applications/application components (e.g., the telemetry identifies a process generating a packet/flow). Generating the ADM can also include determining edges of the ADM by detecting one or more flows between nodes of the ADM. Generating the ADM can further include determining feature vectors for the nodes. The feature vector for each node may include a concatenation of server features, process features, and/or other features. The server features can include the server name, network address, operating system, CPU usage, network usage, disk space, ports, logged users, scheduled jobs, open files, and information regarding files and/or directories stored on the server. The process features can include the process name, process identifier, parent process identifier, process path, CPU utilization, memory utilization, memory address, scheduling information, nice value, flags, priority, status, start time, terminal type, CPU time taken by the process, the command string that started the process, and the process owner (e.g., user name, user identifier, user's real name, e-mail address, user's groups, terminal information, login time, expiration date of login, idle time, and information regarding files and/or directories of the user. In some embodiments, the feature vectors can also include features extracted from customer/third party data such as and customer/third party data (e.g., CMDB or CMS as a service, Whois, geocoordinates, etc.)
Telemetry used to detect flows between the nodes of the ADM may include packet header fields such as source address, source port, destination address, destination port, protocol type, class of service, etc. and/or aggregate packet data such as flow start time, flow end time, number of packets for a flow, number of bytes for a flow, the union of TCP flags for a flow, etc.
Generating the ADM can also involve determining similarity between the nodes of the ADM by comparing their feature vectors. Similarity can be a measure of how much alike two nodes are relative to other nodes, or a measure of two nodes being less distant to one another than other nodes. In some embodiments, the application and network analytics platform can use as similarity/distance measures one or more of Euclidean distance, Manhattan distance, Minkowski distance, cosine similarity, Jaccard similarity, and the like. In some embodiments, determining the ADM can further include setting the similarity threshold for clusters to specify a level of granularity with respect to a view of the applications executing in the network. For example, setting the similarity threshold to a very coarse degree of similarity can result in a single cluster representing the data center as a monolithic application. On the other hand, setting the similarity threshold to a very fine degree of similarity can result in singleton clusters for each host-process pairing in the network. Setting the similarity threshold can depend largely on the number and types of applications executing in the network and the level of granularity desired for the task at hand. In most situations, the similarity threshold may lie somewhere between very coarse and very fine. As discussed, clusters can represent application components, and a connected graph can represent an application.
After generating the ADM, the method 600 can proceed to step 606 in which the application and network analytics platform may devolve flows into one or more flowlets. That is, the application and network analytics platform can break a flow down into a series of sub-requests and sub-responses by tracing a flow from source to destination. A request flow can include hops over network devices from source to destination and processing of the flow by the network devices. A response flow can include hops over network devices, processing of the flow by the network devices, and sub-requests and sub-responses to intermediate endpoints (including hops over network devices and processing by these network devices) performed to generate a response to the originating request flow. For example, in
The method 600 may continue to step 608 in which the application and network analytics platform can determine baseline metrics for the flowlets such as an average or a range of the length of time for transmitting and/or processing each flowlet and/or an average amount of data or a range of the amount of data (e.g., in number of packets and/or bytes) transmitted and/or processed within each flowlet. For example, the application and network analytics platform can detect a request flow associated with an application executing on a server using network telemetry and associated data captured by a software sensor of the server, such as by taking a snapshot of the processes executing on the server around or at the time that the server generated the flow and correlating the flow with one of the processes. The application and network analytics platform can trace the request flow from the server to one or more network devices (i.e., the network devices may correspond to one or more hops of the flow) and determine the transmission time(s) (and sometimes the amounts of data transmitted) between the server and the network device(s) as well as the time spent by the network device(s) to process the flow based on network telemetry captured by the hardware sensor(s) of the network device(s). The application and network analytics platform can trace the response flow as the transmission and/or processing of sub-requests and sub-responses to other servers (and transmission and/or processing by network devices along the data paths of the sub-requests and responses) based on network telemetry and associated data captured by the sensors of these additional devices and determine the amount of times (and/or the amount of data in some embodiments) associated with transmitting and/or processing the sub-requests and sub-responses. The application and network analytics platform can determine a baseline metric for a flowlet by aggregating observations of the flowlet detected during a specified period of time.
At step 610, the application and network analytics platform can evaluate network telemetry and associated data over a second period of time by comparing new flowlets against their corresponding baselines. In some embodiments, the application and network analytics platform may use machine learning and pattern recognition algorithms to classify each new flowlet. Machine learning can involve receiving “trained” samples (i.e., flowlets associated with predetermined classifications), determining the features (sometimes weighted) that the samples of a particular classification may have in common with one another and may not have in common with other classifications, and using these (sometimes weighted) features to classify new flowlets. Machine-learning and pattern recognition algorithms can include decision trees, random forests, rules-based classifiers, support vector machine (SVM) classifiers, neural network classifiers, nearest neighbor classifiers, etc.
Decision trees use a hierarchical division of the underlying data and different features. The hierarchical division of the data creates class partitions skewed in terms of their class distribution. For a given set of differences between flowlets, the partition that a flowlet is most likely to belong to is that flowlet's classification.
Random forests rely on a combination of decision trees in which each tree depends on the values of a random vector sampled independently and with the same distribution for all trees in the forest. Implementers can train a random forest for a number of trees T by sampling n cases of the training data at random with replacement to create a subset of the training data. At each node, the classifier can select a number m of the features at random from the set of all features, and can ultimately select the feature that provides the best split to do a binary split on that node. At the next node, the classifier selects another number m of the features at random and repeats the process.
Rules-based classifiers determine patterns (e.g., regular expressions) that define the different classes. These types of classifiers define a set of rules in which the left-hand side corresponds to a pattern, and the right-hand side corresponds to a class label. The rules determine the classification.
SVM classifiers attempt to partition the underlying data using linear or non-linear delineations between the different classes. The premise of these classifiers is to determine the optimal boundaries between the different classes and use these boundaries for the purposes of classification.
Neural network classifiers utilize an interconnected group of functions or classifiers that process information using a connectionist approach. Neural networks change their structure during training, such as by merging overlapping detections within one network and training an arbitration network to combine the results from different networks.
Nearest neighbor classifiers determine the k-nearest neighbors in the training data to a specific set of differences between flowlets, and the majority class from the k neighbors determines the classification.
At decision point 612, if the new flowlets do not deviate from their baselines, such as by occurring within a specified time range and/or transmitting/processing an amount of data within a specified range of data sizes, then the application and network analytics platform may continue analyzing new flowlets.
However, if the application and network analytics platform determines that the new flowlets deviate from their baselines, at step 614, the application and network analytics platform can perform certain remediation tasks to address the deviations. In some embodiments, remediation can include generating an alert, such as via an event-based notification system (e.g., the event-based notification system 146 of
As discussed, in some embodiments, the application and network analytics platform can determine the physical and/or logical locations of servers within the network (e.g., CMDB/CMS, geocoordinates, IP address, etc.). The application and network analytics can use the location information to determine optimal placement and configuration of application components that are dependent on one another such that the components are closer together physically and/or logically. For example, the application and network analytics platform may determine that an anomalous flow is a bottleneck for a particular application and can locate a different location for the application component. As another example, the application and network analytics platform may have a comprehensive view of the application and network activity in the data center and can determine locations for application components that minimize bandwidth and latency. Thus, in some embodiments, remediation can include automatically migrating virtual entities having application dependencies closer together physically and/or logically to resolve application and network performance issues. In some embodiments, remediation can also include generating recommendations on physical placement or configuration of application components to be physically and/or logically closer together to reduce latency and other performance degradation.
To enable user interaction with the computing system 700, an input device 745 can represent any number of input mechanisms, such as a microphone for speech, a touch-protected screen for gesture or graphical input, keyboard, mouse, motion input, speech and so forth. An output device 735 can also be one or more of a number of output mechanisms known to those of skill in the art. In some instances, multimodal systems can enable a user to provide multiple types of input to communicate with the computing system 700. The communications interface 740 can govern and manage the user input and system output. There may be no restriction on operating on any particular hardware arrangement and various other embodiments may substitute the basic features here for improved hardware or firmware arrangements.
Storage device 730 can be a non-volatile memory and can be a hard disk or other types of computer readable media which can store data that are accessible by a computer, such as magnetic cassettes, flash memory cards, solid state memory devices, digital versatile disks, cartridges, random access memories (RAMs) 725, read only memory (ROM) 720, and hybrids thereof.
The storage device 730 can include software modules 732, 734, 736 for controlling the processor 710. Various embodiments may utilize other hardware or software modules. The storage device 730 can connect to the system bus 705. In one aspect, a hardware module that performs a particular function can include the software component stored in a computer-readable medium in connection with the necessary hardware components, such as the processor 710, bus 705, output device 735, and so forth, to carry out the function.
The chipset 760 can also interface with one or more communication interfaces 790 that can have different physical interfaces. The communication interfaces 790 can include interfaces for wired and wireless LANs, for broadband wireless networks, as well as personal area networks. Some applications of the methods for generating, displaying, and using the GUI disclosed herein can include receiving ordered datasets over the physical interface or be generated by the machine itself by processor 755 analyzing data stored in the storage device 770 or the RAM 775. Further, the computing system 700 can receive inputs from a user via the user interface components 785 and execute appropriate functions, such as browsing functions by interpreting these inputs using the processor 755.
It will be appreciated that computing systems 700 and 750 can have more than one processor 710 and 755, respectively, or be part of a group or cluster of computing devices networked together to provide greater processing capability.
For clarity of explanation, in some instances the various embodiments may be presented as including individual functional blocks including functional blocks comprising devices, device components, steps or routines in a method embodied in software, or combinations of hardware and software.
In some embodiments the computer-readable storage devices, mediums, and memories can include a cable or wireless signal containing a bit stream and the like. However, when mentioned, non-transitory computer-readable storage media expressly exclude media such as energy, carrier signals, electromagnetic waves, and signals per se.
Methods according to the above-described examples can be implemented using computer-executable instructions that are stored or otherwise available from computer readable media. Such instructions can comprise, for example, instructions and data which cause or otherwise configure a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions. Portions of computer resources used can be accessible over a network. The computer executable instructions may be, for example, binaries, intermediate format instructions such as assembly language, firmware, or source code. Examples of computer-readable media that may be used to store instructions, information used, and/or information created during methods according to described examples include magnetic or optical disks, flash memory, USB devices provided with non-volatile memory, networked storage devices, and so on.
Devices implementing methods according to these disclosures can comprise hardware, firmware, and/or software, and can take any of a variety of form factors. Typical examples of such form factors include laptops, smart phones, small form factor personal computers, personal digital assistants, rack mount devices, standalone devices, and so on. Other embodiments may implement the functionality described in this disclosure in peripherals or add-in cards. Various embodiments may also implement this functionality on a circuit board among different chips or different processes executing in a single device, by way of further example.
The instructions, media for conveying such instructions, computing resources for executing them, and other structures for supporting such computing resources are means for providing the functions described in these disclosures.
Although the disclosure uses a variety of examples and other information to explain aspects within the scope of the appended claims, a person having ordinary skill in art will understand not to limit the claims to particular features or arrangements in such examples, as one of ordinary skill can use these examples to derive a wide variety of implementations. Further, although the disclosure describes some subject matter in language specific to examples of structural features and/or method steps, one of ordinary skill will understand that the subject matter defined in the appended claims is not necessarily limited to these described features or acts. For example, such functionality can be distributed differently or performed in components other than those identified herein. Rather, the disclosure provides described features and steps as examples of components of systems and methods within the scope of the appended claims.
Chang, Kai, Parandehgheibi, Ali, Madani, Omid, Yadav, Navindra, Chang, Shih-Chun, Vu, Hai, Kulshreshtha, Ashutosh, Deen, Khawar, Jeyakumar, Vimal, Sloane, Andy
Patent | Priority | Assignee | Title |
Patent | Priority | Assignee | Title |
10009240, | Jun 05 2015 | Cisco Technology, Inc | System and method of recommending policies that result in particular reputation scores for hosts |
5086385, | Jan 31 1989 | Custom Command Systems | Expandable home automation system |
5319754, | Oct 03 1991 | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | Data transfer system between a computer and a host adapter using multiple arrays |
5400246, | May 09 1989 | I O PORT SYSTEMS PARTNERSHIP | Peripheral data acquisition, monitor, and adaptive control system via personal computer |
5436909, | Sep 17 1990 | CONCORD COMMUNICATIONS, INC ; Computer Associates Think, Inc | Network management system using status suppression to isolate network faults |
5555416, | Sep 22 1992 | Sun Microsystems, Inc. | Automated software installation and operating environment configuration for a computer system based on classification rules |
5726644, | Jun 30 1995 | Philips Electronics North America Corporation | Lighting control system with packet hopping communication |
5742829, | Mar 10 1995 | Microsoft Technology Licensing, LLC | Automatic software installation on heterogeneous networked client computer systems |
5822731, | Sep 15 1995 | BTG INTERNATIONAL INC | Adjusting a hidden Markov model tagger for sentence fragments |
5831848, | Nov 17 1995 | Phoenix Controls Corporation | Distributed environmental process control system |
5903545, | Apr 24 1995 | CDC PROPRIETE INTELLECTUELLE | Responsive routing control method and apparatus |
6012096, | Apr 23 1998 | Microsoft Technology Licensing, LLC | Method and system for peer-to-peer network latency measurement |
6085243, | Dec 13 1996 | Hewlett Packard Enterprise Development LP | Distributed remote management (dRMON) for networks |
6141595, | Apr 03 1998 | Johnson Controls Technology Company | Common object architecture supporting application-centric building automation systems |
6144962, | Oct 15 1996 | MICRO FOCUS LLC | Visualization of web sites and hierarchical data structures |
6239699, | Mar 03 1999 | Lucent Technologies Inc.; Lucent Technologies Inc | Intelligent alarm filtering in a telecommunications network |
6247058, | Mar 30 1998 | Hewlett Packard Enterprise Development LP | Method and apparatus for processing network packets using time stamps |
6249241, | Sep 21 1995 | The United States of America as represented by the Secretary of the Navy | Marine vessel traffic system |
6330562, | Jan 29 1999 | TREND MICRO INCORPORATED | System and method for managing security objects |
6353775, | Jul 28 1998 | Honeywell International Inc. | Multiple instance single value identifiers environmental control communication method and system |
6525658, | Jun 11 2001 | ENSCO, INC. | Method and device for event detection utilizing data from a multiplicity of sensor sources |
6546420, | Mar 31 1999 | Cisco Technology, Inc. | Aggregating information about network message flows |
6597663, | Mar 03 1997 | Cisco Technology, Inc. | Technique for handling forwarding transients with link state routing protocol |
6611896, | Aug 25 1999 | EMC IP HOLDING COMPANY LLC | Dynamic mirror service policy with seek adjustment in a non-physical mirrored storage environment |
6654750, | May 05 1998 | TAC, LLC | Network management system having a network database |
6728779, | Dec 01 1999 | THE CHASE MANHATTAN BANK, AS COLLATERAL AGENT | Method and apparatus for exchanging routing information in a packet-based data network |
6801878, | Apr 08 1999 | George Mason University | System and method for managing sensors of a system |
6816461, | Jun 16 2000 | Ciena Corporation | Method of controlling a network element to aggregate alarms and faults of a communications network |
6847993, | May 31 2000 | International Business Machines Corporation | Method, system and program products for managing cluster configurations |
6848106, | Oct 05 1999 | Veritas Technologies LLC | Snapshot restore of application chains and applications |
6925490, | May 31 2000 | International Business Machines Corporation | Method, system and program products for controlling system traffic of a clustered computing environment |
6958998, | Jul 09 2001 | International Business Machines Corporation | Traffic management in packet-based networks |
6983323, | Aug 12 2002 | TREND MICRO INCORPORATED | Multi-level packet screening with dynamically selected filtering criteria |
6996817, | Dec 12 2001 | Valve Corporation | Method and system for upgrading and rolling back versions |
6999452, | Dec 16 1999 | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | Ethernet units adapted for loop configuration and method of operating same |
7002464, | Mar 19 2003 | HOME DATA SOURCE, INC | Relative timing mechanism for event sequencing without clock synchronization |
7024468, | Apr 27 2000 | Hewlett Packard Enterprise Development LP | Internet usage data recording system and method with configurable data collector system |
7096368, | Aug 01 2001 | JPMORGAN CHASE BANK, N A ; MORGAN STANLEY SENIOR FUNDING, INC | Platform abstraction layer for a wireless malware scanning engine |
7111055, | Aug 30 2001 | Oracle America, Inc | Method and apparatus to facilitate automated software installation on remote computers over a network |
7120934, | Mar 30 2000 | IRDETO USA, INC | System, method and apparatus for detecting, identifying and responding to fraudulent requests on a network |
7133923, | Dec 11 2000 | ACME PACKET, INC | System and method for assisting in controlling real-time transport protocol flow through multiple networks via screening |
7162643, | Jun 15 2001 | Informatica LLC | Method and system for providing transfer of analytic application data over a network |
7181769, | Aug 25 2000 | TRIPWIRE, INC | Network security system having a device profiler communicatively coupled to a traffic monitor |
7185103, | Sep 10 2002 | Juniper Networks, Inc. | Rate-controlled transmission of traffic flow information |
7203740, | Dec 22 1999 | Intel Corporation | Method and apparatus for allowing proprietary forwarding elements to interoperate with standard control elements in an open architecture for network devices |
7302487, | Mar 22 2001 | Rambus Inc | Security system for a data communications network |
7337206, | Jul 15 2002 | RIVERBED TECHNOLOGY LLC | Method for detecting congestion in internet traffic |
7349761, | Feb 07 2002 | Cisco Technology, Inc | System and method for distributed facility management and operational control |
7353511, | Feb 01 1999 | Cisco Technology, Inc. | Method and system for dynamically distributing updates in a network |
7356679, | Apr 11 2003 | VMware, Inc. | Computer image capture, customization and deployment |
7360072, | Mar 28 2003 | Cisco Technology, Inc.; Cisco Technology, Inc | iSCSI system OS boot configuration modification |
7370092, | Sep 12 2002 | Computer Sciences Corporation | System and method for enhanced software updating and revision |
7395195, | Dec 27 2004 | SAP SE | Sensor network modeling and deployment |
7444404, | Feb 05 2001 | ARBOR NETWORKS, INC | Network traffic regulation including consistency based detection and filtering of packets with spoof source addresses |
7466681, | Mar 19 2004 | RPX CLEARINGHOUSE LLC | Method and apparatus for sensor network routing |
7467205, | May 12 2003 | Cisco Technology, Inc | Systems and methods for identifying the client applications of a network |
7496040, | Jul 22 2004 | LG Electronics Inc | Roundtrip delay time measurement apparatus and method for variable bit rate multimedia data |
7496575, | Nov 22 2004 | Digital Guardian LLC | Application instrumentation and monitoring |
7530105, | Mar 21 2006 | Northrop Grumman Systems Corporation | Tactical and strategic attack detection and prediction |
7539770, | Apr 25 2001 | Cisco Technology, Inc. | DHCP over mobile IP |
7568107, | Aug 20 2003 | Extreme Networks | Method and system for auto discovery of authenticator for network login |
7610330, | Mar 30 2006 | CA, INC | Multi-dimensional computation distribution in a packet processing device having multiple processing architecture |
7633942, | Oct 15 2001 | Gula Consulting Limited Liability Company | Network traffic generation and monitoring systems and methods for their use in testing frameworks for determining suitability of a network for target applications |
7644438, | Oct 27 2004 | MICRO FOCUS LLC | Security event aggregation at software agent |
7676570, | Jun 14 2002 | ZARBAÑA DIGITAL FUND LLC | Determining client latencies over a network |
7681131, | Nov 10 2004 | INTERNETPERILS, INC | Method and apparatus for aggregating, condensing, supersetting, and displaying network topology and performance data |
7693947, | Mar 08 2002 | JPMORGAN CHASE BANK, N A , AS ADMINISTRATIVE AGENT | Systems and methods for graphically displaying messaging traffic |
7743242, | Oct 16 2006 | Dell Products L P | Method and system for automatic generation of operating system boot images |
7752307, | Jul 28 2006 | TWITTER, INC | Technique of analyzing an information system state |
7774498, | Nov 06 2006 | Cisco Technology, Inc. | Methods and apparatus for trusted application centric QoS provisioning |
7783457, | Jun 15 2006 | Oracle America, Inc | Sensor localization using lateral inhibition |
7787480, | Mar 04 2009 | Juniper Networks, Inc. | Routing frames in a trill network using service VLAN identifiers |
7788477, | Jan 31 2007 | VALTRUS INNOVATIONS LIMITED | Methods, apparatus and articles of manufacture to control operating system images for diskless servers |
7808897, | Mar 01 2005 | LinkedIn Corporation | Fast network security utilizing intrusion prevention systems |
7813822, | Oct 05 2000 | Blanding Hovenweep, LLC; HOFFBERG FAMILY TRUST 1 | Intelligent electronic appliance system and method |
7844696, | Jun 27 2001 | Arbor Networks | Method and system for monitoring control signal traffic over a computer network |
7844744, | Apr 25 2008 | International Business Machines Corporation | Providing server security via a security sensor application shared by multiple operating system partitions |
7864707, | Feb 24 2009 | LinkedIn Corporation | Determination of network topology using flow-based traffic information |
7873025, | Feb 23 2006 | Cisco Technology, Inc. | Network device that determines application-level network latency by monitoring option values in a transport layer message |
7873074, | Jun 01 2006 | AVAYA LLC | Adaptive selection of bandwidth parameters to meet a service provider pricing model |
7874001, | Jul 15 2005 | Microsoft Technology Licensing, LLC | Detecting user-mode rootkits |
7885197, | Nov 17 2003 | GUANGDONG OPPO MOBILE TELECOMMUNICATIONS CORP , LTD | System and method for measuring per node packet loss in a wireless network |
7895649, | Apr 04 2003 | FORCEPOINT FEDERAL HOLDINGS LLC | Dynamic rule generation for an enterprise intrusion detection system |
7904420, | Aug 26 2008 | Raytheon Company | Identification and verification of common cluster files residing on nodes in a cluster |
7930752, | Nov 17 2006 | Nexthink S.A. | Method for the detection and visualization of anomalous behaviors in a computer network |
7934248, | Jun 27 2007 | EMC IP HOLDING COMPANY LLC | Network policy enforcement dashboard views |
7957934, | May 15 2007 | Dynatrace LLC | Method and system for processing application performance data ouside of monitored applications to limit overhead caused by monitoring |
7961637, | Jun 07 2004 | SPIRENT COMMUNICATIONS OF ROCKVILLE INC | Method and apparatus for monitoring latency, jitter, packet throughput and packet loss ratio between two points on a network |
7970946, | Nov 27 2007 | GOOGLE LLC | Recording and serializing events |
7975035, | Dec 01 2003 | International Business Machines Corporation | Method and apparatus to support application and network awareness of collaborative applications using multi-attribute clustering |
8001610, | Sep 28 2005 | Pulse Secure, LLC | Network defense system utilizing endpoint health indicators and user identity |
8005935, | Apr 05 2007 | International Business Machines Corporation | Methods and computer program products for managing application performance on a network |
8040232, | Sep 01 2006 | KAIFI, LLC | USN middleware apparatus and method for generating information based on data from heterogeneous sensor networks and information service providing system using the same |
8040822, | Jun 04 2009 | RPX Corporation | Configuring communication services using policy groups |
8056134, | Sep 10 2006 | Malware detection and identification via malware spoofing | |
8115617, | Nov 17 2008 | Cisco Technology, Inc. | Alarm reordering to handle alarm storms in large networks |
8135657, | Jul 01 2005 | CA, INC | Systems and methods for processing data flows |
8156430, | Dec 16 2002 | Palo Alto Research Center Incorporated | System and method for clustering nodes of a tree structure |
8160063, | Jun 09 2008 | Microsoft Technology Licensing, LLC | Data center interconnect and traffic engineering |
8179809, | Aug 23 1999 | Oracle America, Inc | Approach for allocating resources to an apparatus based on suspendable resource requirements |
8181248, | Nov 23 2006 | F1 SECURITY INC | System and method of detecting anomaly malicious code by using process behavior prediction technique |
8185824, | Dec 09 1998 | Cisco Technology, Inc. | Method and apparatus providing a graphical user interface for representing and navigating hierarchical networks |
8239365, | Aug 04 2006 | Flexible request and response communications interfaces | |
8239915, | Jun 30 2006 | CA, INC | Endpoint management using trust rating data |
8250657, | Dec 29 2006 | CA, INC ; NORTONLIFELOCK INC | Web site hygiene-based computer security |
8255972, | Jun 06 2008 | International Business Machines Corporation | Method to automatically map business function level policies to it management policies |
8266697, | Mar 04 2006 | Northrop Grumman Systems Corporation | Enabling network intrusion detection by representing network activity in graphical form utilizing distributed data sensors to detect and transmit activity data |
8272875, | Mar 09 1999 | REALITYWORKS, INC | Educational device for simulating addictive behavior and method of using |
8281397, | Apr 29 2010 | Telcordia Technologies, Inc | Method and apparatus for detecting spoofed network traffic |
8291495, | Aug 08 2007 | Juniper Networks, Inc. | Identifying applications for intrusion detection systems |
8296847, | Jul 25 2003 | Hewlett Packard Enterprise Development LP | Method of managing utilization of network intrusion detection systems in a dynamic data center |
8311973, | Sep 24 2011 | Z ADVANCED COMPUTING, INC | Methods and systems for applications for Z-numbers |
8365286, | Jun 30 2006 | DEUTSCHE BANK AG NEW YORK BRANCH, AS COLLATERAL AGENT | Method and system for classification of software using characteristics and combinations of such characteristics |
8370407, | Jun 28 2011 | Go Daddy Operating Company, LLC | Systems providing a network resource address reputation service |
8381289, | Mar 31 2009 | CA, INC | Communication-based host reputation system |
8391270, | Jun 14 2007 | KONINKLIJKE PHILIPS ELECTRONICS N V | Network synchronization with reduced energy consumption |
8407164, | Oct 02 2006 | TRUSTEES OF COLUMBIA UNIVERSITY IN THE CITY OF NEW YORK, THE | Data classification and hierarchical clustering |
8407798, | Oct 01 2002 | SKYBOX SECURITY INC | Method for simulation aided security event management |
8413235, | Sep 10 2010 | CA, INC | Malware detection using file heritage data |
8442073, | Oct 25 2007 | Siemens Healthcare GmbH | Method and an apparatus for analyzing a communication network |
8451731, | Jul 25 2007 | Xangati, Inc | Network monitoring using virtual packets |
8462212, | Dec 04 2008 | NCR Voyix Corporation | Correlating detected events with image data |
8489765, | May 28 2010 | Cisco Technology, Inc. | Dynamic directed acyclic graph (DAG) adjustment |
8499348, | Dec 28 2010 | Amazon Technologies, Inc.; Amazon Technologies, Inc | Detection of and responses to network attacks |
8516590, | Apr 25 2009 | DASIENT, INC | Malicious advertisement detection and remediation |
8527977, | Jun 07 1996 | McAfee, Inc. | Software uninstallation system, method and computer program product |
8549635, | Apr 01 2003 | JPMORGAN CHASE BANK, N A , AS ADMINISTRATIVE AGENT | Malware detection using external call characteristics |
8570861, | Jun 29 2010 | Amazon Technologies Inc | Reputation-based networking |
8572600, | Jul 26 2007 | Gangneung-Wonju National University Industrial Academy Cooperation Group | Method for updating firmware of sensor nodes on the wireless sensor network |
8572734, | Aug 12 2004 | PALO ALTO NETWORKS, INC | Geographical intrusion response prioritization mapping through authentication and flight data correlation |
8572735, | Mar 29 2007 | GEORGE MASON RESEARCH FOUNDATION, INC | Attack resistant continuous network service trustworthiness controller |
8572739, | Oct 27 2009 | TREND MICRO INCORPORATED | Detection of malicious modules injected on legitimate processes |
8588081, | Jul 14 2010 | Cisco Technology, Inc. | Monitoring a flow set to detect faults |
8600726, | Apr 16 2008 | Keysight Technologies, Inc | System and method for virtualization of networking system software via emulation |
8613084, | Sep 18 2007 | JPMORGAN CHASE BANK, N A , AS ADMINISTRATIVE AGENT | System, method, and computer program product for detecting at least potentially unwanted activity based on execution profile monitoring |
8615803, | Nov 29 2001 | International Business Machines Corporation | Method, computer program element and a system for processing alarms triggered by a monitoring system |
8630316, | Jun 19 2008 | Microsoft Technology Licensing, LLC | Clock synchronization using correlation events |
8631464, | Apr 20 2004 | ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL | Method of detecting anomalous behaviour in a computer network |
8640086, | Dec 29 2006 | SAP SE | Graphical user interface system and method for presenting objects |
8656493, | May 22 2006 | GOSECURE, INC | Decoy network technology with automatic signature generation for intrusion detection and intrusion prevention systems |
8661544, | Aug 31 2010 | Cisco Technology, Inc. | Detecting botnets |
8677487, | Oct 18 2011 | JPMORGAN CHASE BANK, N A , AS ADMINISTRATIVE AGENT | System and method for detecting a malicious command and control channel |
8683389, | Sep 08 2010 | THE NEW ENGLAND COMPLEX SYSTEMS INSTITUTE, INC | Method and apparatus for dynamic information visualization |
8706914, | Apr 23 2007 | Computing infrastructure | |
8713676, | May 13 2010 | Verisign, Inc | Systems and methods for identifying malicious domains using internet-wide DNS lookup patterns |
8719452, | Jul 29 2011 | GOOGLE LLC | Correction of client-assigned timestamps |
8719835, | Jun 07 2011 | TELEFONAKTIEBOLAGET L M ERICSSON PUBL | Ranking service units to provide and protect highly available services using the Nway redundancy model |
8750287, | May 06 2009 | RPX Corporation | Traffic-engineered connection establishment across resource domains for data transport |
8752042, | Aug 27 2008 | CardinalCommerce Corporation | Intelligent server routing |
8752179, | Mar 14 2012 | AO Kaspersky Lab | System and method for removal of malicious software from computer systems and management of treatment side-effects |
8755396, | Sep 11 2008 | Juniper Networks, Inc | Methods and apparatus related to flow control within a data center switch fabric |
8762951, | Mar 21 2007 | Oracle America, Inc. | Apparatus and method for profiling system events in a fine grain multi-threaded multi-core processor |
8769084, | Jun 07 2010 | JPMORGAN CHASE BANK, N A , AS SUCCESSOR AGENT | System and method for modeling interdependencies in a network datacenter |
8775577, | Dec 18 2007 | Amazon Technologies, Inc | System and method for configuration management service |
8776180, | May 01 2012 | TAASERA LICENSING LLC | Systems and methods for using reputation scores in network services and transactions to calculate security risks to computer systems and platforms |
8812448, | Nov 09 2011 | ACCESS SCIENCES CORPORATION | Computer implemented method for accelerating electronic file migration from multiple sources to multiple destinations |
8812725, | May 18 2012 | Cisco Technology Inc. | System and method for latency reduction in a network environment |
8813236, | Jan 07 2013 | Narus, Inc. | Detecting malicious endpoints using network connectivity and flow information |
8825848, | Mar 20 2012 | EMC IP HOLDING COMPANY LLC | Ordering of event records in an electronic system for forensic analysis |
8832013, | Jul 24 2009 | Decision Lens, Inc. | Method and system for analytic network process (ANP) total influence analysis |
8832461, | Jun 25 2010 | Microsoft Technology Licensing, LLC | Trusted sensors |
8849926, | Aug 06 2010 | Simon Fraser University | System and method for self-calibrating, self-organizing and localizing sensors in wireless sensor networks |
8881258, | Aug 24 2011 | JPMORGAN CHASE BANK, N A , AS ADMINISTRATIVE AGENT | System, method, and computer program for preventing infections from spreading in a network environment using dynamic application of a firewall policy |
8887238, | Dec 07 2011 | Time Warner Cable Enterprises LLC | Mechanism for establishing reputation in a network environment |
8904520, | Mar 19 2009 | CA, INC | Communication-based reputation system |
8908685, | Sep 30 2010 | Microsoft Technology Licensing, LLC | Routing using global address pairs |
8914497, | Dec 15 2011 | Amazon Technologies, Inc | System and method for throttling service requests having non-uniform workloads |
8931043, | Apr 10 2012 | JPMORGAN CHASE BANK, N A , AS ADMINISTRATIVE AGENT | System and method for determining and using local reputations of users and hosts to protect information in a network environment |
8954610, | Jul 21 2010 | Dell Products L.P. | System-wide time synchronization across power management interfaces and sensor data |
8955124, | Apr 28 2010 | Electronics and Telecommunications Research Institute | Apparatus, system and method for detecting malicious code |
8966021, | Dec 20 2011 | Amazon Technologies, Inc.; Amazon Technologies, Inc | Composable machine image |
8966625, | May 24 2011 | PALO ALTO NETWORKS, INC | Identification of malware sites using unknown URL sites and newly registered DNS addresses |
8973147, | Dec 29 2011 | JPMORGAN CHASE BANK, N A , AS ADMINISTRATIVE AGENT | Geo-mapping system security events |
8984331, | Sep 06 2012 | RPX Corporation | Systems and methods for automated memory and thread execution anomaly detection in a computer network |
8990386, | Aug 28 2009 | HCL Technologies Limited | Searching virtual resources |
8996695, | Nov 10 2011 | CopperEgg Corporation | System for monitoring elastic cloud-based computing systems as a service |
8997227, | Feb 27 2012 | Amazon Technologies, Inc | Attack traffic signature generation using statistical pattern recognition |
9014047, | Jul 10 2007 | Level 3 Communications, LLC | System and method for aggregating and reporting network traffic data |
9015716, | Apr 30 2013 | SPLUNK Inc.; SPLUNK INC | Proactive monitoring tree with node pinning for concurrent node comparisons |
9071575, | Apr 21 2011 | Method and system for abuse route aggregation and distribution | |
9088598, | Nov 14 2013 | Narus, Inc. | Systematic mining of associated server herds for uncovering malware and attack campaigns |
9110905, | Jun 11 2010 | Cisco Technology, Inc | System and method for assigning network blocks to sensors |
9117075, | Nov 22 2010 | TREND MICRO, INC | Early malware detection by cross-referencing host data |
9130836, | Feb 04 2013 | Cisco Technology, Inc.; Cisco Technology, Inc | Provisoning of a new node joining an existing cluster in a data center environment |
9152789, | May 28 2008 | ZSCALER, INC | Systems and methods for dynamic cloud-based malware behavior analysis |
9160764, | Jul 15 2011 | NORSE NETWORKS, INC | Systems and methods for dynamic protection from electronic attacks |
9170917, | Apr 01 2011 | PayPal, Inc | Flow tracing though diverse logical and physical application and infrastructure layers/dependencies |
9178906, | Oct 28 2010 | CA, INC | Detecting and remediating malware dropped by files |
9185127, | Jul 06 2011 | AKAMAI TECHNOLOGIES, INC | Network protection service |
9191400, | Jun 12 2013 | The United States of America, as represented by the Secretary of the Navy | Cyphertext (CT) analytic engine and method for network anomaly detection |
9191402, | Jan 25 2013 | Cisco Technology, Inc | Domain classification based on client request behavior |
9197654, | Jun 28 2013 | JPMORGAN CHASE BANK, N A , AS ADMINISTRATIVE AGENT | Rootkit detection by using HW resources to detect inconsistencies in network traffic |
9225793, | Jan 28 2011 | Cisco Technology, Inc.; Cisco Technology, Inc | Aggregating sensor data |
9237111, | Mar 14 2013 | KYNDRYL, INC | Credit-based flow control in lossless ethernet networks |
9246702, | Jan 31 2012 | Cisco Technology, Inc. | System and method for configuring service appliances as virtual line cards in a network environment |
9246773, | Jul 30 2013 | SYSDIG, INC | System, method, and graphical user interface for application topology mapping in hosted computing environments |
9253042, | Oct 05 2012 | NEC Corporation | Network management |
9253206, | Dec 18 2014 | DocuSign, Inc. | Systems and methods for protecting an online service attack against a network-based attack |
9258203, | Sep 29 2006 | Amazon Technologies, Inc.; Amazon Technologies, Inc | Monitoring computer performance metrics utilizing baseline performance metric filtering |
9258217, | Dec 16 2008 | AT&T Intellectual Property I, L.P. | Systems and methods for rule-based anomaly detection on IP network flow |
9281940, | Mar 02 2012 | Sony Corporation | Information processing apparatus, information processing method, and program |
9286047, | Feb 13 2013 | Cisco Technology, Inc.; Cisco Technology, Inc | Deployment and upgrade of network devices in a network environment |
9294486, | Mar 05 2014 | National Technology & Engineering Solutions of Sandia, LLC | Malware detection and analysis |
9317574, | Jun 11 2012 | QUEST SOFTWARE INC F K A DELL SOFTWARE INC ; Aventail LLC | System and method for managing and identifying subject matter experts |
9319384, | Apr 30 2014 | Fortinet, Inc.; Fortinet, INC | Filtering hidden data embedded in media files |
9369435, | Sep 30 2013 | Cisco Technology, Inc. | Method for providing authoritative application-based routing and an improved application firewall |
9369479, | Jun 25 2012 | EMC IP HOLDING COMPANY LLC | Detection of malware beaconing activities |
9378068, | Mar 13 2013 | KYNDRYL, INC | Load balancing for a virtual networking system |
9396327, | May 16 2011 | D2L Corporation | Systems and methods for security verification in electronic learning systems and other systems |
9405903, | Oct 31 2013 | PALO ALTO NETWORKS, INC | Sinkholing bad network domains by registering the bad network domains on the internet |
9417985, | Nov 14 2014 | Microsoft Technology Licensing, LLC | Distributed analysis and attribution of source code |
9418222, | Sep 27 2013 | CA, INC | Techniques for detecting advanced security threats |
9426068, | Feb 24 2012 | FUTUREWEI TECHNOLOGIES, INC | Balancing of forwarding and address resolution in overlay networks |
9454324, | Dec 18 2013 | EMC IP HOLDING COMPANY LLC | Methods and apparatus for data lifecycle analysis |
9462013, | Apr 29 2015 | KYNDRYL, INC | Managing security breaches in a networked computing environment |
9465696, | Jun 03 2011 | Apple Inc. | Methods and apparatus for multi-phase multi-source backup |
9501744, | Jun 11 2012 | QUEST SOFTWARE INC F K A DELL SOFTWARE INC ; Aventail LLC | System and method for classifying data |
9531589, | May 30 2014 | Cisco Technology, Inc. | Automating monitoring using configuration event triggers in a network environment |
9563517, | Dec 30 2013 | EMC IP HOLDING COMPANY LLC | Cloud snapshots |
9621413, | Mar 29 2012 | ARRIS ENTERPRISES LLC | Displaying dynamic host configuration protocol (DHCP) transaction states using a DHCP relay agent |
9634915, | Aug 14 2009 | Microsoft Technology Licensing, LLC | Methods and computer program products for generating a model of network application health |
9645892, | Jul 08 2014 | EMC IP HOLDING COMPANY LLC | Recording file events in change logs while incrementally backing up file systems |
9684453, | Mar 08 2011 | CITIBANK, N A , AS COLLATERAL AGENT | Cluster federation and trust in a cloud environment |
9697033, | Nov 15 2011 | Nicira, Inc. | Architecture of networks with middleboxes |
9733973, | Sep 16 2015 | Cisco Technology, Inc. | Automatically determining sensor location in a virtualized computing environment |
9749145, | Mar 08 2013 | KYNDRYL, INC | Interoperability for distributed overlay virtual environment |
9800608, | Sep 25 2000 | CA, INC | Processing data flows with a data flow processor |
9904584, | Nov 26 2014 | Microsoft Technology Licensing, LLC | Performance anomaly diagnosis |
9916538, | Sep 15 2012 | Z Advanced Computing, Inc. | Method and system for feature detection |
9935851, | Jun 05 2015 | Cisco Technology, Inc. | Technologies for determining sensor placement and topology |
20010028646, | |||
20020053033, | |||
20020097687, | |||
20020103793, | |||
20020107857, | |||
20020141343, | |||
20020184393, | |||
20030023601, | |||
20030065986, | |||
20030097439, | |||
20030126242, | |||
20030145232, | |||
20030151513, | |||
20030154399, | |||
20030177208, | |||
20040019676, | |||
20040030776, | |||
20040039728, | |||
20040213221, | |||
20040220984, | |||
20040243533, | |||
20040255050, | |||
20040268149, | |||
20050028154, | |||
20050039104, | |||
20050063377, | |||
20050083933, | |||
20050108331, | |||
20050122325, | |||
20050138157, | |||
20050166066, | |||
20050177829, | |||
20050182681, | |||
20050185621, | |||
20050198247, | |||
20050198371, | |||
20050198629, | |||
20050207376, | |||
20050240799, | |||
20050257244, | |||
20050289244, | |||
20060048218, | |||
20060077909, | |||
20060080733, | |||
20060089985, | |||
20060095968, | |||
20060143432, | |||
20060156408, | |||
20060159032, | |||
20060173912, | |||
20060195448, | |||
20060272018, | |||
20060274659, | |||
20060280179, | |||
20060294219, | |||
20070014275, | |||
20070025306, | |||
20070044147, | |||
20070097976, | |||
20070118654, | |||
20070127491, | |||
20070162420, | |||
20070169179, | |||
20070195729, | |||
20070195794, | |||
20070195797, | |||
20070201474, | |||
20070211637, | |||
20070214348, | |||
20070230415, | |||
20070232265, | |||
20070250930, | |||
20070300061, | |||
20080002697, | |||
20080022385, | |||
20080028389, | |||
20080046708, | |||
20080049633, | |||
20080056124, | |||
20080082662, | |||
20080101234, | |||
20080120350, | |||
20080126534, | |||
20080141246, | |||
20080155245, | |||
20080250122, | |||
20080270199, | |||
20080282347, | |||
20080295163, | |||
20080301765, | |||
20090059934, | |||
20090064332, | |||
20090109849, | |||
20090133126, | |||
20090138590, | |||
20090180393, | |||
20090241170, | |||
20090292795, | |||
20090296593, | |||
20090300180, | |||
20090307753, | |||
20090313373, | |||
20090313698, | |||
20090319912, | |||
20090323543, | |||
20090328219, | |||
20100005288, | |||
20100049839, | |||
20100054241, | |||
20100077445, | |||
20100095293, | |||
20100095367, | |||
20100095377, | |||
20100138526, | |||
20100138810, | |||
20100148940, | |||
20100153316, | |||
20100153696, | |||
20100180016, | |||
20100194741, | |||
20100220584, | |||
20100235514, | |||
20100235879, | |||
20100235915, | |||
20100287266, | |||
20100303240, | |||
20100306180, | |||
20100317420, | |||
20100319060, | |||
20110004935, | |||
20110010585, | |||
20110022641, | |||
20110055381, | |||
20110055388, | |||
20110066719, | |||
20110069685, | |||
20110072119, | |||
20110083125, | |||
20110085556, | |||
20110103259, | |||
20110107074, | |||
20110107331, | |||
20110126136, | |||
20110126275, | |||
20110145885, | |||
20110153039, | |||
20110153811, | |||
20110158088, | |||
20110170860, | |||
20110173490, | |||
20110185423, | |||
20110196957, | |||
20110202655, | |||
20110214174, | |||
20110225207, | |||
20110228696, | |||
20110238793, | |||
20110246663, | |||
20110277034, | |||
20110283277, | |||
20110302652, | |||
20110314148, | |||
20110317982, | |||
20120005542, | |||
20120079592, | |||
20120089664, | |||
20120102361, | |||
20120102543, | |||
20120110188, | |||
20120117226, | |||
20120117642, | |||
20120136996, | |||
20120137278, | |||
20120137361, | |||
20120140626, | |||
20120195198, | |||
20120197856, | |||
20120198541, | |||
20120210318, | |||
20120216271, | |||
20120218989, | |||
20120219004, | |||
20120233348, | |||
20120233473, | |||
20120240232, | |||
20120246303, | |||
20120254109, | |||
20120260227, | |||
20120278021, | |||
20120281700, | |||
20120300628, | |||
20130003538, | |||
20130003733, | |||
20130006935, | |||
20130007435, | |||
20130038358, | |||
20130041934, | |||
20130054682, | |||
20130085889, | |||
20130086272, | |||
20130103827, | |||
20130107709, | |||
20130124807, | |||
20130125107, | |||
20130145099, | |||
20130148663, | |||
20130159999, | |||
20130166730, | |||
20130173784, | |||
20130174256, | |||
20130179487, | |||
20130179879, | |||
20130198517, | |||
20130198839, | |||
20130201986, | |||
20130205293, | |||
20130219161, | |||
20130219500, | |||
20130232498, | |||
20130242999, | |||
20130246925, | |||
20130247201, | |||
20130254879, | |||
20130268994, | |||
20130275579, | |||
20130283374, | |||
20130290521, | |||
20130297771, | |||
20130301472, | |||
20130304900, | |||
20130305369, | |||
20130318357, | |||
20130326623, | |||
20130333029, | |||
20130336164, | |||
20130346736, | |||
20130347103, | |||
20140006610, | |||
20140006871, | |||
20140010069, | |||
20140012814, | |||
20140019972, | |||
20140031005, | |||
20140033193, | |||
20140036688, | |||
20140040343, | |||
20140047185, | |||
20140047372, | |||
20140056318, | |||
20140059200, | |||
20140074946, | |||
20140089494, | |||
20140092884, | |||
20140096058, | |||
20140105029, | |||
20140115219, | |||
20140129942, | |||
20140136680, | |||
20140137109, | |||
20140140244, | |||
20140143825, | |||
20140149490, | |||
20140156814, | |||
20140156861, | |||
20140164607, | |||
20140165054, | |||
20140165200, | |||
20140165207, | |||
20140173623, | |||
20140192639, | |||
20140201717, | |||
20140215573, | |||
20140215621, | |||
20140224784, | |||
20140225603, | |||
20140233387, | |||
20140269777, | |||
20140280499, | |||
20140281030, | |||
20140286354, | |||
20140289854, | |||
20140298461, | |||
20140301213, | |||
20140307686, | |||
20140317278, | |||
20140317737, | |||
20140330616, | |||
20140331048, | |||
20140331276, | |||
20140331280, | |||
20140331304, | |||
20140348182, | |||
20140351203, | |||
20140351415, | |||
20140359695, | |||
20150006689, | |||
20150006714, | |||
20150009840, | |||
20150026809, | |||
20150033305, | |||
20150036480, | |||
20150036533, | |||
20150039751, | |||
20150046882, | |||
20150052441, | |||
20150058976, | |||
20150067143, | |||
20150067786, | |||
20150082151, | |||
20150082430, | |||
20150085665, | |||
20150095332, | |||
20150112933, | |||
20150113133, | |||
20150124608, | |||
20150124652, | |||
20150128133, | |||
20150128205, | |||
20150138993, | |||
20150142962, | |||
20150195291, | |||
20150222939, | |||
20150249622, | |||
20150256555, | |||
20150261842, | |||
20150261886, | |||
20150271008, | |||
20150271255, | |||
20150295945, | |||
20150312233, | |||
20150347554, | |||
20150356297, | |||
20150358352, | |||
20160006753, | |||
20160019030, | |||
20160020959, | |||
20160021131, | |||
20160026552, | |||
20160036636, | |||
20160036837, | |||
20160050132, | |||
20160072815, | |||
20160080414, | |||
20160087861, | |||
20160094394, | |||
20160094529, | |||
20160103692, | |||
20160105350, | |||
20160112270, | |||
20160112284, | |||
20160119234, | |||
20160127395, | |||
20160147585, | |||
20160162308, | |||
20160162312, | |||
20160173446, | |||
20160173535, | |||
20160183093, | |||
20160191476, | |||
20160205002, | |||
20160216994, | |||
20160217022, | |||
20160255082, | |||
20160269424, | |||
20160269442, | |||
20160269482, | |||
20160294691, | |||
20160308908, | |||
20160337204, | |||
20160357424, | |||
20160357546, | |||
20160357587, | |||
20160357957, | |||
20160359592, | |||
20160359628, | |||
20160359658, | |||
20160359673, | |||
20160359677, | |||
20160359678, | |||
20160359679, | |||
20160359680, | |||
20160359686, | |||
20160359695, | |||
20160359696, | |||
20160359697, | |||
20160359698, | |||
20160359699, | |||
20160359700, | |||
20160359701, | |||
20160359703, | |||
20160359704, | |||
20160359705, | |||
20160359709, | |||
20160359711, | |||
20160359712, | |||
20160359740, | |||
20160359759, | |||
20160359872, | |||
20160359877, | |||
20160359878, | |||
20160359879, | |||
20160359880, | |||
20160359881, | |||
20160359888, | |||
20160359889, | |||
20160359890, | |||
20160359891, | |||
20160359897, | |||
20160359905, | |||
20160359912, | |||
20160359913, | |||
20160359914, | |||
20160359915, | |||
20160359917, | |||
20160373481, | |||
20160380865, | |||
20160380869, | |||
20170006141, | |||
20170024453, | |||
20170032310, | |||
20170034018, | |||
20170048121, | |||
20170070582, | |||
20170085483, | |||
20170104658, | |||
20170126532, | |||
20170208487, | |||
20170250880, | |||
20170250951, | |||
20170289067, | |||
20170295141, | |||
20170302691, | |||
20170318034, | |||
20170331747, | |||
20170346736, | |||
20170364380, | |||
20180006911, | |||
20180007115, | |||
20180013670, | |||
20180096081, | |||
20180145906, | |||
20200084121, | |||
CN101093452, | |||
CN101770551, | |||
CN102521537, | |||
CN103023970, | |||
CN103716137, | |||
CN104065518, | |||
CN107196807, | |||
EP811942, | |||
EP1076848, | |||
EP1383261, | |||
EP1450511, | |||
EP2043320, | |||
EP2045974, | |||
EP2860912, | |||
EP2887595, | |||
JP2009016906, | |||
KR1394338, | |||
WO2007014314, | |||
WO2007070711, | |||
WO2008069439, | |||
WO2013030830, | |||
WO2015042171, | |||
WO2015099778, | |||
WO2016004075, | |||
WO2016019523, | |||
WO2016196683, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Mar 10 2017 | CHANG, SHIH-CHUN | Cisco Technology, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 058152 | /0398 | |
Mar 11 2017 | VU, HAI | Cisco Technology, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 058152 | /0398 | |
Mar 16 2017 | JEYAKUMAR, VIMAL | Cisco Technology, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 058152 | /0398 | |
Mar 20 2017 | KULSHRESHTHA, ASHUTOSH | Cisco Technology, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 058152 | /0398 | |
Mar 20 2017 | DEEN, KHAWAR | Cisco Technology, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 058152 | /0398 | |
Mar 20 2017 | SLOANE, ANDY | Cisco Technology, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 058152 | /0398 | |
Mar 22 2017 | PARANDEHGHEIBI, ALI | Cisco Technology, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 058152 | /0398 | |
Mar 23 2017 | MADANI, OMID | Cisco Technology, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 058152 | /0398 | |
Mar 27 2017 | CHANG, KAI | Cisco Technology, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 058152 | /0398 | |
Mar 02 2021 | YADAV, NAVINDRA | Cisco Technology, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 058152 | /0398 | |
Nov 18 2021 | Cisco Technology, Inc. | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Nov 18 2021 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Date | Maintenance Schedule |
Jun 20 2026 | 4 years fee payment window open |
Dec 20 2026 | 6 months grace period start (w surcharge) |
Jun 20 2027 | patent expiry (for year 4) |
Jun 20 2029 | 2 years to revive unintentionally abandoned end. (for year 4) |
Jun 20 2030 | 8 years fee payment window open |
Dec 20 2030 | 6 months grace period start (w surcharge) |
Jun 20 2031 | patent expiry (for year 8) |
Jun 20 2033 | 2 years to revive unintentionally abandoned end. (for year 8) |
Jun 20 2034 | 12 years fee payment window open |
Dec 20 2034 | 6 months grace period start (w surcharge) |
Jun 20 2035 | patent expiry (for year 12) |
Jun 20 2037 | 2 years to revive unintentionally abandoned end. (for year 12) |