Remediation of network devices that are failing is accomplished using a shared failure domain (SFD) database that provides neighboring device/link information to remediation tools. SFD refers to a group of objects (links/devices) that share a same failure model. A state change of one or multiple of the objects results in a corresponding action on other devices linked together through the SFD. Moreover, the SFD data is available in a central repository and software tools consult the central repository for failure domain data before taking remedial actions. SFD data is generated using configuration generation and device state. Software tools lookup SFD data during operational events (device/link down) and take appropriate actions on the neighboring devices.
|
6. A method, comprising:
detecting a failing network device in a network;
finding an entry within a shared failure domain (SFD) database associated with the failing network device, wherein the entry identifies neighbor network devices associated with the failing network device by having a same address prefix;
determining at least two neighbor devices having the same address prefix using the entry in the SFD database; and
disabling operational capacity of the neighbor devices having the same address prefix as identified in the SFD and the failing network device.
17. A system, comprising:
a network switch having a plurality of operational ports and a failing port;
a plurality of neighbor switches, wherein a first of the plurality of neighbor switches is coupled to one of the plurality of operational ports, and a second of the plurality of neighbor switches is coupled to the failing port, wherein the first and the second of the plurality of neighbor switches are on a same tier; and
a controller coupled to the network switch and the plurality of neighbor switches, the controller configured to:
read a database associating ports in the network switch with ports in the plurality of neighbor switches, wherein the database is populated by using a forwarding table of the network switch and identifying the first and second neighbor switches in the forwarding table that have a same address prefix as the network switch; and
perform a given action on the first and the second of the plurality of neighbor switches having the same address prefix within the database as the network switch in response to detecting the failing port is not operational.
1. A method of performing remedial action on a network device, the method comprising:
detecting a failing port on a network device;
retrieving an entry for the network device from a database including shared failure domains, which associates the failing port to other ports on the network device or ports on neighboring devices, wherein the network device and the neighboring devices are in a multi-tiered network and wherein the entry from the database is generated by identifying neighbor network devices having a same address prefix as the network device found in a forwarding table of the network device;
using the entry in the database for the network device, identifying neighboring devices to the network device that are related to the network device based on the database;
identifying an operational port on the network device that is coupled to one of the neighboring devices and identifying the failing port as coupled to another of the neighboring devices; and
disabling ports including the failing port on the network device that are identified in the entry in the database as having the same address prefix as the network device.
2. The method of
disabling the operational ports on the neighboring devices.
4. The method of
5. The method of
7. The method of
8. The method of
9. The method of
10. The method of
11. The method of
12. The method of
13. The method of
14. The method of
16. The method of
18. The system of
19. The system of
20. The system of
|
Network switches play a critical role in large computer networks, such as those found in a data center. Server computers in the network can be divided into racks, and network switches can connect the server computers within a rack to routers in the data center. Data passed by switches is generally divided into a series of packets that can be transmitted between devices. Packets include control information and payload data. The control information includes information used to deliver the payload data. For example, control information can include source and destination network addresses, error detection codes, packet sequencing identification, and the like. Typically, control information is found in packet headers and trailers included within the packet and adjacent to the payload data.
Generally, network switches have two primary planes: a control plane and a data plane. The control plane is a management plane that configures the data plane. The data plane receives packets on input ports and transmits the received packets to output ports based on the configuration. A forwarding table within the data plane controls which next hops receive the packets. Typically, the forwarding table is programmed in accordance with routing strategies, such as equal-cost multi-path routing (ECMP), which is a strategy where next-hop packet forwarding to a destination can occur over multiple best paths. ECMP is per-hop decision limited to a single router, and can be inefficient if routers in the network are not functioning.
Failure modes pertaining to hardware (devices/links/transceivers) failures in three-tier Clos networks are difficult to calculate and remediate due to the complexity of Clos fabrics and combined failures. Combined failures are defined as two or more failures in a Clos network degrading the overall capacity of the fabric. Combined failures can happen over a period of time where the fabric's overall capacity is gradually reduced below an acceptable threshold, which increases the risk of traffic congestion. Clos fabrics are composed of a large number of devices and links that increases the possibility of combined failures in the network. Sometimes a device or link failure requires remediation action on a neighboring device. However, neighboring device information is not available to operational tools. The result can be congestion on one or more ports on the network devices.
Shared Failure Domains (SFD) provide neighboring device/link information to remediation tools. SFD refers to a group of objects (links/devices) that share a same failure model. More particularly, if a device or port of a device fails, then other devices or other ports are treated similarly in terms of actions taken in response to the failure. In one example, if one port fails, then both that port and any other port in the group of objects are deactivated. In other embodiments, if the entire device fails or is shifted away, then other related devices within the SFD can be shut down entirely or shifted away. A state change of one or multiple of the objects results in a corresponding action on other devices linked together through the SFD. Moreover, the SFD data is available in a central repository and software tools consult the central repository for failure domain data before taking remedial actions. SFD data is generated using configuration generation, network topology and device state. Software tools lookup SFD data during operational events (e.g., if a device or link is down) and take appropriate actions on the neighboring devices. Software tools can define workflows for each network design and failure event. SFD data can also have a field for specifying network design. Based on the network design, software tools can identify which particular workflow should be executed for a given device.
The SFD database 140 includes an object (e.g., a record) per device in the network 100. For example, an object for device 1 is shown at 150 and includes one or more fields 152 of related devices. The structure of the objects can differ based on the particular implementation. However, each device port can have its own entry with devices related to that port should it fail. The related devices include at least one other related port on the same device and neighbor devices and ports on those neighbor devices. The SFD database can be populated using a variety of techniques. One technique is to analyze a topology of the network, and any neighboring devices that are within a similar grouping on the network can be considered related. Thus, the device object can be populated based on a logical network analysis of the topology. Network management rules can also be used in conjunction or in place of the logical network analysis. For example, in some cases where network traffic is low, a reduction in capacity can be considered acceptable. In a very particular example, capacity can be reduced by 50%, but the utilization of the interfaces is only 5%. Thus, the network management rules can consider utilization and other factors as a supplement to the logical network analysis to determine an entry in the SFD database. In other embodiments, the SFD database can include some of the network management rules, and other network management rules are implemented directly in the controller 130. Turning briefly to
To overcome this problem with overloading an operational port, the controller 130 can shut down (i.e., disable) all of the related ports listed within the SFD database 140. Alternatively, the controller 130 can disable the entire device or shift away the device. In the particular example shown in
An advantage of shutting down operational ports is that for certain prefixes that use ports X and Y, other devices will not send packets to device 1 and will send such packets to other network devices. As a result, throughput actually increases and packet loss decreases, as opposed to port X being left operational and getting overloaded, such that packets are dropped. As a particular example, a device can have two 10 Gb ports (i.e., interface capacity) with a device utilization of 12 Gb, which is evenly distributed amongst the ports (6 Gb each). If one of the two ports is shut down, the entire 12 Gb is shifted to the other of the two ports. The result is dropped packets and congestion. In this particular example, rather than lose packets, it is more advantageous to shut down both ports.
As illustrated, each core switch 208 is able to communicate with each of a plurality of aggregation switches 210, 212, which in at least some embodiments are utilized in pairs. Utilizing aggregation switches in pairs provides a redundant capability in case one of the switches experiences a failure or is otherwise unavailable, such that the other devices can route traffic through the connected devices. Each pair of aggregation switches 210, 212 is linked to a plurality of physical racks 214, 215, each of which typically contains a top of rack (TOR) or “access” switch 216 and a plurality of physical host machines 218, such as data servers and other processing devices. As an additional benefit, the use of aggregation switch pairs enables the capability of a link to be exceeded during peak periods, for example, wherein both aggregation switches can concurrently handle and route traffic. Each pair of aggregation switches can service a dedicated number of racks based on factors such as capacity, a number of ports, etc. There can be any appropriate number of aggregation switches in a data center, such as six aggregation pairs. The traffic from the aggregation pairs can be aggregated by the core switches, which can pass the traffic “up and out” of the data center, such as back across the network 204. In some embodiments, the core switches are provided in pairs as well.
Similar to
Due to the relationship between switches 340, 342 and 330, a shared failure domain table 370 associates ports on these switches together. A controller 372 detects the failing port on switch 330 and checks the SFD table 370 to determine related ports. As indicated above, ports on switches 340, 342 are associated with the malfunctioning port on switch 330. Accordingly, the controller 372 can disable all ports on the three switches. In particular, the controller 372 can communicate with the switch's control plane and request that the ports be disabled. Other ports on the switches can continue to operate and pass packets through a data plane of the switches. Instead of disabling the switches, other remedial actions can be taken. However, the actions can be the same for all related ports within the SFD. In other embodiments, the controller 372 can detect if the entire network device is failing rather than just individual ports. In still other embodiments, the controller can detect if the network device is shifted away. In any of these cases, the controller 372 can search in the SFD to determine actions to take in response.
With reference to
A computing system may have additional features. For example, the computing environment 600 includes storage 640, one or more input devices 650, one or more output devices 660, and one or more communication connections 670. An interconnection mechanism (not shown) such as a bus, controller, or network interconnects the components of the computing environment 600. Typically, operating system software (not shown) provides an operating environment for other software executing in the computing environment 600, and coordinates activities of the components of the computing environment 600.
The tangible storage 640 may be removable or non-removable, and includes magnetic disks, magnetic tapes or cassettes, CD-ROMs, DVDs, or any other medium which can be used to store information in a non-transitory way and which can be accessed within the computing environment 600. The storage 640 stores instructions for the software 680 implementing one or more innovations described herein.
The input device(s) 650 may be a touch input device such as a keyboard, mouse, pen, or trackball, a voice input device, a scanning device, or another device that provides input to the computing environment 600. The output device(s) 660 may be a display, printer, speaker, CD-writer, or another device that provides output from the computing environment 600.
The communication connection(s) 670 enable communication over a communication medium to another computing entity. The communication medium conveys information such as computer-executable instructions, audio or video input or output, or other data in a modulated data signal. A modulated data signal is a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media can use an electrical, optical, RF, or other carrier.
Although the operations of some of the disclosed methods are described in a particular, sequential order for convenient presentation, it should be understood that this manner of description encompasses rearrangement, unless a particular ordering is required by specific language set forth below. For example, operations described sequentially may in some cases be rearranged or performed concurrently. Moreover, for the sake of simplicity, the attached figures may not show the various ways in which the disclosed methods can be used in conjunction with other methods.
Any of the disclosed methods can be implemented as computer-executable instructions stored on one or more computer-readable storage media (e.g., one or more optical media discs, volatile memory components (such as DRAM or SRAM), or non-volatile memory components (such as flash memory or hard drives)) and executed on a computer (e.g., any commercially available computer, including smart phones or other mobile devices that include computing hardware). The term computer-readable storage media does not include communication connections, such as signals and carrier waves. Any of the computer-executable instructions for implementing the disclosed techniques as well as any data created and used during implementation of the disclosed embodiments can be stored on one or more computer-readable storage media. The computer-executable instructions can be part of, for example, a dedicated software application or a software application that is accessed or downloaded via a web browser or other software application (such as a remote computing application). Such software can be executed, for example, on a single local computer (e.g., any suitable commercially available computer) or in a network environment (e.g., via the Internet, a wide-area network, a local-area network, a client-server network (such as a cloud computing network), or other such network) using one or more network computers.
For clarity, only certain selected aspects of the software-based implementations are described. Other details that are well known in the art are omitted. For example, it should be understood that the disclosed technology is not limited to any specific computer language or program. For instance, the disclosed technology can be implemented by software written in C++, Java, Perl, any other suitable programming language. Likewise, the disclosed technology is not limited to any particular computer or type of hardware. Certain details of suitable computers and hardware are well known and need not be set forth in detail in this disclosure.
It should also be well understood that any functionality described herein can be performed, at least in part, by one or more hardware logic components, instead of software. For example, and without limitation, illustrative types of hardware logic components that can be used include Field-programmable Gate Arrays (FPGAs), Program-specific Integrated Circuits (ASICs), Program-specific Standard Products (ASSPs), System-on-a-chip systems (SOCs), Complex Programmable Logic Devices (CPLDs), etc.
Furthermore, any of the software-based embodiments (comprising, for example, computer-executable instructions for causing a computer to perform any of the disclosed methods) can be uploaded, downloaded, or remotely accessed through a suitable communication means. Such suitable communication means include, for example, the Internet, the World Wide Web, an intranet, software applications, cable (including fiber optic cable), magnetic communications, electromagnetic communications (including RF, microwave, and infrared communications), electronic communications, or other such communication means.
The disclosed methods, apparatus, and systems should not be construed as limiting in any way. Instead, the present disclosure is directed toward all novel and nonobvious features and aspects of the various disclosed embodiments, alone and in various combinations and subcombinations with one another. The disclosed methods, apparatus, and systems are not limited to any specific aspect or feature or combination thereof, nor do the disclosed embodiments require that any one or more specific advantages be present or problems be solved.
In view of the many possible embodiments to which the principles of the disclosed invention may be applied, it should be recognized that the illustrated embodiments are only examples of the invention and should not be taken as limiting the scope of the invention. We therefore claim as our invention all that comes within the scope of these claims.
Zuber, Owais Bin, Callaghan, Stephen
Patent | Priority | Assignee | Title |
11165665, | Dec 28 2017 | Fujitsu Limited | Apparatus and method to improve precision of identifying a range of effects of a failure in a system providing a multilayer structure of services |
11431556, | Mar 06 2020 | Beijing University of Posts and Telecommunications; State Grid Liaoning Power Co., Ltd. Dalian Power Suoply Company | Cross-layer network fault recovery system and method based on configuration migration |
11909671, | Aug 18 2022 | Hewlett Packard Enterprise Development LP | Efficient port reconfiguration |
Patent | Priority | Assignee | Title |
8554952, | Jun 29 2010 | Amazon Technologies, Inc | Fused data center fabrics |
8995249, | Feb 13 2013 | Amazon Technologies, Inc | Predicting route utilization and non-redundant failures in network environments |
20110170405, | |||
20150172172, | |||
20160020939, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Jun 04 2018 | CALLAGHAN, STEPHEN | Amazon Technologies, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 046307 | /0593 | |
Jun 06 2018 | Amazon Technologies, Inc. | (assignment on the face of the patent) | / | |||
Jun 06 2018 | ZUBER, OWAIS BIN | Amazon Technologies, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 046307 | /0593 |
Date | Maintenance Fee Events |
Jun 06 2018 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Nov 04 2024 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Date | Maintenance Schedule |
May 04 2024 | 4 years fee payment window open |
Nov 04 2024 | 6 months grace period start (w surcharge) |
May 04 2025 | patent expiry (for year 4) |
May 04 2027 | 2 years to revive unintentionally abandoned end. (for year 4) |
May 04 2028 | 8 years fee payment window open |
Nov 04 2028 | 6 months grace period start (w surcharge) |
May 04 2029 | patent expiry (for year 8) |
May 04 2031 | 2 years to revive unintentionally abandoned end. (for year 8) |
May 04 2032 | 12 years fee payment window open |
Nov 04 2032 | 6 months grace period start (w surcharge) |
May 04 2033 | patent expiry (for year 12) |
May 04 2035 | 2 years to revive unintentionally abandoned end. (for year 12) |