A method and apparatus that promote expedited fast link failover decisions by and end station with redundant multilayer switching paths to a network are disclosed. In response to detecting a change in connectivity with an upstream link to a network, an upstream switch may send a notification to an intermediate switch with a downstream link to the end station. In response to the notification, the intermediate switch may disable (“bring down”) the downstream link to the end station. Bringing the downstream link down in this manner may cause the end station to switch (“failover”) to a different downstream link, for example, allowing the end station to establish connectivity to the network via a redundant path.
|
22. A method for managing network connectivity, comprising:
receiving, at a second network device, a notification of a loss of connectivity in an upstream communication link between a first network device and a network core, wherein the first network device and the second network device are connected by an aggregate communication link;
identifying a group of one or more end stations in communication with the second network device affected by the loss of connectivity in the upstream communication link between the first network device and the network core;
in response to receiving the notification, disabling a network connection with each end station in the identified group of end stations in a manner that causes each end station in the identified group to failover to a downstream link between that end station and a third network device in order to establish an alternate path to the network core that does not include the first network device; and
continuing to carry network traffic received on the aggregate communication link connecting the first network device and second network device originating from at least one end station not in the identified group of end stations after receiving the notification of the loss of connectivity.
1. A method for managing a network path between at least one end station and a network core, comprising:
detecting, by a first network device, a change in status of a first upstream communication link between the first network device and the network core, wherein the change in status of the first upstream communication link comprises a loss in connectivity between the first network device and the network core;
sending, by the first network device, via an aggregate link between the first network device and a second network device, notification of the change in status to the second network device, wherein the second network device is in communication with a plurality of end stations via respective first downstream communication links, and wherein the notification identifies a group of one or more of the end stations affected by the loss in connectivity;
disabling the first downstream communication links in a manner that causes each end station in the group to failover to a second downstream link between a third network device and the end station to establish an alternate path between the group of one or more end stations and the network core that does not include the first network device; and
continuing to carry network traffic received on the aggregate link originating from at least one end station after sending the notification of change in status to the second network device.
8. A network switching system, comprising:
at least a first network device configured to detect a change in status of a first upstream communication link between the first network device and a network core, wherein the change in status of the first upstream communication link comprises a loss in connectivity between the first network device and the network core;
at least a second network device configured to receive, from the first network device, notification of the change in status from the first network device, via an aggregate link between the first network device and the second network device, wherein the notification identifies a group of one or more of the end stations affected by the loss in connectivity
a first downstream communication link between the second network device and each end station in the group that is disabled in response to the notification of the change in status, wherein the aggregate link continues to carry traffic for management of the network switching system after disabling the first downstream communication link and further continues to carry network traffic originating from at least one end station after sending the notification of change in status to the second network device; and
a second downstream link between a third network device and each end station in the group that is established as an alternate path between the end station and the network core that does not include the first network device when the first downstream communication link is disabled.
15. An aggregation device comprising:
a processor; and
a non-transitory computer readable medium containing instructions which, when executed by the processor, cause the aggregation device to perform operations comprising:
detecting a change in status of a first upstream communication link between the aggregation device and a network core, wherein the change in status of the first upstream communication link is a loss in connectivity between the aggregation device and the network core,
sending notification of the change in status, via an aggregate link between the aggregation device and a first network device, to the first network device in communication with a plurality of end stations via respective first downstream communication links,
identifying a group of one or more of the end stations in communication with the first network device effected by the loss in connectivity,
disabling the first downstream communication link of each end station in the group of one or more end stations in a manner that causes the end stations in the group to each failover to a respective second downstream link to a second network device and the respective end station to establish an alternate path between the respective end station and the network core that does not include the aggregation device and further continues to carry network traffic originating from at least one of the plurality of end stations after sending the notification of change in status to the first network device, and
maintaining the aggregate link to continue to carry traffic for management of a network after sending the notification and to carry network traffic originating from at least a second end station after sending the notification of change in status to the first network device.
12. A system, comprising:
a plurality of end stations;
at least a first network device configured to detect a change in status of a first upstream communication link between the first network device and a network core, wherein the change in status of the first upstream communication link is a loss in connectivity between the first network device and the network core; and
at least a second network device configured to receive, from the first network device, via an aggregate link between the first network device and the second network device, notification of the change in status from the first network device, wherein the notification identifies a group of one or more of the end stations in communication with the second network device affected by the loss in connectivity;
downstream communication links between first network interface devices on the plurality of end stations including a respective first downstream communication link with each of the identified group of the end stations that is disabled in response to the notification of the change in status, wherein the aggregate link continues to carry traffic for management of the system after the first downstream communication link is disabled and further continues to carry network traffic originating from at least one of the end stations after sending the notification of change in status to the second network device;
additional downstream communications links between second network interface devices thereon and at least a third network device including a respective second downstream communications link that is established in response to the disabling of the first downstream communication link to provide an alternate path between of the identified group of end stations and the network core that does not include the first network device.
2. The method of
3. The method of
sending a message from the first network device to one or more network devices in communication with the first network device.
4. The method of
5. The method of
6. The method of
7. The method of
detecting, by the first network device, recovery of connectivity between the first network device and the network core; and
sending, by the first network device, notification of the recovery in status to the second network device; and
in response to receiving the notification of the recovery in status, enabling by the second network device, the first downstream communication link.
9. The system of
10. The system of
11. The system of
13. The system of
14. The system of
16. The aggregation device of
detecting recovery of connectivity between the aggregation device and the network core; and
sending notification of the recovery in status to the first network device.
17. The method of
18. The network switching system of
19. The system of
20. The aggregation device of
21. The aggregation device of
|
1. Field of the Invention
The present invention generally relates to management of switched computer networks and, more particularly, to propagating detected changes in upstream connectivity to downstream devices.
2. Description of the Related Art
A computer network generally includes a number of devices, including switches, routers and hubs, connected so as to allow communication among the devices. The devices within a network may be generally categorized into two classes: end stations connected along the edge of a network and network devices. End stations typically include devices that primarily supply or consume information, such as workstations, desktop PCs, printers, servers, hosts, fax machines, and other such devices. Network devices typically include devices that primarily transfer information between other devices (e.g., between end stations and other network devices), such as gateways, switches, routers, and other such devices. In other words, network devices generally provide connectivity between end stations and the network core.
In an effort to provide greater reliability, such networks are often designed to provide redundant Layer 2 connectivity between end stations 110 and the network core 120. To this end, at least some end stations 110 may each have multiple network interface cards NICs 112 (or other type network interface/adapter), allowing the end stations 110 to connect to the network through primary links/connections (Lp) and secondary links (Ls). For example, the end stations may be configured to connect to the network through a primary path (via SW3 and SW1) by default and switch (or “failover”) to a secondary path (via SW4 and SW2) if a connectivity problem is detected in the primary path, such as a loss of connectivity to a link LC between switch SW1 and the network core 120.
In conventional systems, end stations do not become aware of a change in upstream connectivity until they discover it on their own. For example, end stations may only discover a loss of upstream connectivity after experiencing consistent protocol packet timeouts. As a result, after a loss in network connectivity on the primary path, end stations may not immediately switch from their primary path to their alternative path despite a loss in network connectivity on the primary path, which may degrade network performance.
Therefore, there is a need for techniques that cause end stations to switch over to alternative paths more rapidly in response to remote upstream connectivity changes.
Embodiments of the present invention provide techniques for propagating changes in upstream connectivity downstream to connected end stations.
One embodiment provides a method for managing a network path between at least one end station and a network. The method generally includes detecting, by a first switching device, a change in status of a first upstream communication link between the first switching device and the network, sending, by the first switching device, notification of the change in status to a second switching device in communication with the end station via a first downstream communication link, and, in response to receiving the notification, disabling by the second switch, the first downstream communication link in a manner that causes the end station to failover to a second downstream link between a third switch and the end station.
Another embodiment provides a network switching system generally including at least a first switching device and at least a second switching device. The first switching device is generally configured to detect a change in status of a first upstream communication link between the first switching device and a network core. The second switching device is generally configured to receive, from the first switching device, notification of the change in status from the first switching device and, in response, disable a first downstream communication link between the second switching device and an end station. For some embodiments, their may be one or more intermediate switching devices, between the first and second switching devices, that receive the notification of the change in status from the first switching device and propagate the notification (possibly through other intermediate switches) to the second switching device that ultimately disables the first downstream communication link with the end station.
Another embodiment provides a system generally including a plurality of end stations, at least a first switching device, and at least a second switching device. The first switching device is generally configured to detect a change in status of a first upstream communication link between the first switching device and a network core. The second switching device is generally configured to receive, from the first switching device, notification of the change in status from the first switching device and, in response, disable first downstream communication links between first network interface devices on the plurality of end stations, thereby causing at least some of the plurality of end stations to switch to second downstream communications links between second network interface devices thereon and at least a third switching device.
Another embodiment provides an aggregation switching device generally including a processor and a computer readable medium containing instructions which, when executed by the processor, cause the switching device to perform operations. The operations generally include detecting a change in status of a first upstream communication link between the first switching device and a network and sending notification of the change in status to a second switching device in communication with an end station via a first downstream communication link.
Another embodiment provides an access switching device generally including a processor and a computer readable medium containing instructions which, when executed by the processor, cause the switching device to perform operations. The operations generally include in response to receiving, from an upstream switching device, notification of a detected change in status between the first switching device and a network, disabling at least one downstream communication link between the access switching device and at least one end station.
So that the manner in which the above recited features of the present invention can be understood in detail, a more particular description of the invention, briefly summarized above, may be had by reference to embodiments, some of which are illustrated in the appended drawings. It is to be noted, however, that the appended drawings illustrate only typical embodiments of this invention and are therefore not to be considered limiting of its scope, for the invention may admit to other equally effective embodiments.
Embodiments of the present invention generally provide methods and apparatus for propagating changes in upstream connectivity downstream. For some embodiments, an upstream switch connected to a network core may detect a change in connectivity and, in response, send a message indicating the change to a downstream switch connected to an end station. In response to the message, the downstream switch may disable a link with end station. Disabling the link may cause the end station to promptly switch (fail over) to a different link, in an effort to establish an alternate path to the network. Further, by maintaining the link between the intermediate or downstream switch and the upstream device, network traffic routed through that switch need not be re-routed and the link can continue to carry traffic destined for other networks.
Methods and apparatus for managing a multilayer switching path are described. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. It will be appreciated, however, to one skilled in the art that the present invention may be practiced without these specific details. In other instances, well-known structures and devices are shown in block diagram form (absent specific details) in order to avoid unnecessarily obscuring the present invention.
To facilitate understanding, the following terms, as used herein, have the following meaning. As used herein, the term end station generally refers to any device capable of a network connection via one or more network interfaces and need not necessarily reside at an edge of a network. Examples of end stations include a server, PC, fax machine, printer, or the like. The term switch (or switching device) generally refers to any type device that switches network traffic, such as a Layer 2 switch (per the OSI model) that switches Ethernet traffic. The term access device generally refers to a network element that is connected to one or more end stations. The term aggregation device generally refers to a network element that connects to one or more access devices. The term upstream generally refers to a path (or device along a path) in the direction from an end station towards the core of the network. The term downstream generally refers to a path (or device along a path) in the direction from the core of the network to the edge (e.g., towards an end station). The terms link or connection may be used herein interchangeably and generally refer to any connection between devices, such as a Layer 2 network connection between two devices. The term failure generally refers to a loss of connectivity to a link from a device, regardless of the particular cause, including if the link is intentionally disabled by another device.
Embodiments of the present invention allow changes in upstream connectivity to be communicated downstream, thereby allowing an end station to quickly make decisions regarding which paths to use to the network. The techniques described herein may be implemented, for example, as software methods (e.g., executable instructions residing in memory) running on network devices (e.g., aggregation devices and/or access devices) that are directly or indirectly connected to a network interface of an end station. As will be described herein, the indirectly connected switches may track the link status on upstream interfaces and communicate any change in connectivity (loss/restoration) to a set of downstream interfaces, for example, by sending a message. Such a message may be intercepted by downstream network devices, such as access switches directly connected to network interfaces of the end stations. In response, these switches may take appropriate action based on the message communicated, for example, disabling corresponding active links if the message indicates loss of connectivity. This allows conventional end stations with redundant network interfaces to detect that their primary connection is down and to automatically switch over to their secondary network interfaces.
To facilitate understanding, only a limited portion of the network shown in
The operations begin, at step 202, by detecting a change in upstream connectivity between a first switch and a network core. Referring to
At step 204, the change in connectivity is communicated to a second switch located downstream between the first switch and an end station. As further illustrated in
At step 206, a downstream connection between the second switch and the end station is modified in response to the change in connectivity. For example, as illustrated in
In any case, end stations equipped with redundant network interfaces may be configured to react to such a loss in connectivity to a primary link by automatically switching (or “failing over”) to a secondary interface in order to establish an alternate path to the network. Thus, in the illustrated example as shown in
For some embodiments, should the connectivity between SW1 and the network core be restored, switch SW1 may forward another message to SW3 (and/or other connected devices) indicating such. In response to the message, SW3 may bring the link with the connected end station back up, restoring it to an operational state. In response to detecting that the primary link has been brought back up, one or more of the end stations 110 may switch back to the primary connections immediately, or at any appropriate later time, for example, depending on the particular circumstance.
By propagating an upstream Layer 2 connectivity failure in this manner, switchover times for the end stations may be reduced. In other words, end stations may switchover much more rapidly than in conventional systems that rely on waiting for a higher layer protocol running on these end stations to detect that its link to the network is not working properly. These higher layer protocols typically rely upon timers to fail, and also have to account for network congestion to avoid erroneously detecting “false” failures.
In addition to improving switchover times, the techniques described herein also allow upstream switches (e.g., SW1 and SW2) to maintain their interfaces to downstream devices (e.g., SW3 and SW4). This is advantageous in that other network traffic may continue to pass through the interface from access devices to aggregation devices and allows multiple networks to be active that would otherwise be inactive if this link were brought down. These links (LX) may continue to carry traffic for management, as well as other local network traffic.
The particular manner in which changes in upstream connectivity are propagated down to an end station may vary, depending on a particular implementation. As an example, the particular format and contents of the message sent by an upstream device to communicate a change in upstream connectivity may vary. For some embodiments, the contents of the message may include at least an indication of the detected change (e.g., loss or restoration), as well as any other suitable information, such as identification of an intended recipient (destination). The type and/or number of devices to receive a message indicating a change in connectivity may also vary depending on a particular implementation and, for some embodiments, may be configurable by a user (e.g., a network administrator).
The contents of messages may vary with different embodiments and different types of messages may be sent (e.g., broadcast to many devices or sent directly to identified recipients). In any case, messages may include a common set of fields including, but not limited to one or more of the following: 1) a Switch identifier (source of the msg); 2) a Msg_type/Action_type (e.g., Link Lost, Link Restored, or link removed from being monitored, for example, if a user reconfigures a link from being the cause of downstream interfaces being brought down); 3) an Upstream Link Identifier (which may be at least be unique on the source switch); 4) a SW Version of some type; 5) a msg counter for a number of msgs of this type sent; and 6) a time stamp of action occurrence. For some embodiments, the Upstream Link Identifier may be a field in the message that identifies the upstream interface. As an alternative, the message may include an identifier of a group that is effected by the change in connectivity of an upstream interface. The upstream interface may be a single physical interface, or it could be a logical grouping of physical interfaces.
The particular contents of a message may depend on the particular type of message being sent. For some embodiments, there may be at least two types of messages, for example, broadcast or direct, with each having different contents. Regardless of the implementation, msg contents would most likely contain a common set of fields, such as those listed above. The field contents, however, would likely change depending up on the action and the source switch.
As illustrated in
For some embodiments, a user may be able to configure a particular set of downstream interfaces that should be brought down upon receiving a message indicating a loss of upstream connectivity. For example, as illustrated in
A second group of end stations 5102 may not have a secondary NIC and, thus should not be brought down in the event that upstream connectivity is lost. Such end stations (e.g., older “legacy” end stations) may not have the capability to switch over from a primary to secondary interface. As such, the downstream interfaces connecting to such end stations may be configured to be maintained (not brought down) by the user in the case of an upstream link losing connectivity. As illustrated, end stations in Group 2 may continue to use the same interface with SW3 and their traffic may be routed to the network core via a redundant link (LZ) between SW1 and SW2 to reach the core. While the bandwidth of the redundant link LZ may not be as high as the direct link to the core (link LC) and may not be meant to carry sustained traffic to the core, it may still allow the end station to remain active.
For some embodiments, the techniques described herein may be configurable by a user, for example, via a user command to run automatically, on demand, or be disabled. In some cases, a user may be allowed to define logical grouping of upstream interfaces to downstream interfaces. Further, for some embodiments, in addition to, or as an alternative to selecting groups, the techniques described herein could be configurable to be enabled in a variety of ways with varying scope, such as on a “per port” basis, via a particular connection (e.g., a VLAN), or switch-wide.
By propagating upstream connectivity losses detected by an upstream device to a downstream device connected, the downstream device may bring down a connection with an end station. Bringing down the connection may result in the end station promptly switching over to a redundant connection. Further, by maintaining the link between the intermediate switch and the upstream device, network traffic routed therethrough need not be re-routed.
While the foregoing is directed to embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow.
Tirumalai, Gokul S., Mitchell, Albert Hyde
Patent | Priority | Assignee | Title |
10097462, | Apr 02 2016 | VMware LLC | Throughput resilience during link failover |
10237157, | Jun 10 2015 | Amazon Technologies, Inc | Managing host failures in a traffic forwarding system |
10855516, | Sep 30 2015 | Orange | System for restoring services provided by a residential gateway |
10999169, | Nov 29 2019 | Amazon Technologies, Inc | Configuration and management of scalable global private networks |
11336528, | Nov 29 2019 | Amazon Technologies, Inc | Configuration and management of scalable global private networks |
11533231, | Nov 29 2019 | Amazon Technologies, Inc | Configuration and management of scalable global private networks |
11729077, | Nov 29 2019 | Amazon Technologies, Inc | Configuration and management of scalable global private networks |
9838302, | Jun 10 2015 | Amazon Technologies, Inc | Managing loss of network connectivity in traffic forwarding systems |
9866470, | Jan 24 2014 | Red Hat, Inc. | Multiple active link aggregators |
Patent | Priority | Assignee | Title |
5835696, | Nov 22 1995 | THE CHASE MANHATTAN BANK, AS COLLATERAL AGENT | Data router backup feature |
6032194, | Dec 24 1997 | Cisco Technology, Inc | Method and apparatus for rapidly reconfiguring computer networks |
6330229, | Nov 09 1998 | Hewlett Packard Enterprise Development LP | Spanning tree with rapid forwarding database updates |
6392990, | Jul 23 1999 | Glenayre Electronics, Inc. | Method for implementing interface redundancy in a computer network |
6442133, | May 24 2000 | NEC Corporation | Switchboard having a dual switching system |
6513129, | Jun 30 1999 | OBJECTIVE SYSTEMS INTEGRATORS, INC | System and method for managing faults using a gateway |
6535491, | Dec 24 1997 | Cisco Technology, Inc. | Method and apparatus for rapidly reconfiguring computer networks using a spanning tree algorithm |
6542934, | May 30 1997 | International Business Machines Corporation | Non-disruptively rerouting network communications from a secondary network path to a primary path |
6628661, | Aug 27 1998 | Intel Corporation | Spanning tree recovery in computer networks |
6678241, | Nov 30 1999 | Cisc Technology, Inc. | Fast convergence with topology switching |
6697339, | Mar 04 1999 | Hewlett Packard Enterprise Development LP | High availability spanning tree with rapid reconfiguration with alternate port selection |
6721275, | May 03 1999 | Hewlett Packard Enterprise Development LP | Bridged network stations location revision |
6801506, | Mar 31 1999 | Cisco Technology, Inc. | Method and apparatus for providing fast spanning tree re-starts |
6891808, | Mar 10 2000 | Anritsu Corporation | Spanning tree bridge and route change method using the same |
6956824, | Jun 14 2001 | Alcatel-Lucent Canada Inc | Extension of link aggregation protocols over the network |
6987740, | Sep 11 2000 | Cisco Technology, Inc. | STP root guard |
7016351, | Feb 29 2000 | Cisco Technology, Inc | Small group multicast in a computer network |
7197660, | Jun 26 2002 | Juniper Networks, Inc | High availability network security systems |
7453900, | Mar 05 2003 | Cisco Technology, Inc. | System and method for monitoring noise associated with a communication link |
7480283, | Mar 26 2002 | Genband US LLC; SILICON VALLEY BANK, AS ADMINISTRATIVE AGENT | Virtual trunking over packet networks |
7518986, | Nov 16 2005 | Juniper Networks, Inc. | Push-based hierarchical state propagation within a multi-chassis network device |
7580349, | Nov 02 2001 | AVAYA LLC | Content-aware dynamic network resource allocation |
7593319, | Oct 15 2002 | Garrettcom, Inc. | LAN switch with rapid fault recovery |
7616561, | Jan 19 2005 | Juniper Networks, Inc. | Systems and methods for routing data in a communications network |
20020112072, | |||
20020152320, | |||
20020167895, | |||
20020184387, | |||
20030016624, | |||
20030021226, | |||
20030097470, | |||
20030137932, | |||
20030158964, | |||
20030161260, | |||
20030179700, | |||
20030185225, | |||
20040001449, | |||
20040047336, | |||
20040098501, | |||
20040105390, | |||
20040133693, | |||
20040264364, | |||
20050063395, | |||
20050108401, | |||
20050111350, | |||
20060023630, | |||
20060159011, | |||
20060206602, | |||
20060227704, | |||
20060256712, | |||
20070041313, | |||
20070047436, | |||
20070047578, | |||
20070070886, | |||
20070076727, | |||
20070183347, | |||
20070280103, | |||
20080016402, | |||
20080181196, | |||
20090096592, | |||
20090109998, | |||
20100020680, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Apr 05 2006 | Cisco Technology, Inc. | (assignment on the face of the patent) | / | |||
Apr 05 2006 | TIRUMALAI, GOKUL S | Cisco Technology, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 017762 | /0523 | |
Apr 05 2006 | MITCHELL, ALBERT HYDE | Cisco Technology, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 017762 | /0523 |
Date | Maintenance Fee Events |
May 11 2018 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Jul 04 2022 | REM: Maintenance Fee Reminder Mailed. |
Dec 19 2022 | EXP: Patent Expired for Failure to Pay Maintenance Fees. |
Date | Maintenance Schedule |
Nov 11 2017 | 4 years fee payment window open |
May 11 2018 | 6 months grace period start (w surcharge) |
Nov 11 2018 | patent expiry (for year 4) |
Nov 11 2020 | 2 years to revive unintentionally abandoned end. (for year 4) |
Nov 11 2021 | 8 years fee payment window open |
May 11 2022 | 6 months grace period start (w surcharge) |
Nov 11 2022 | patent expiry (for year 8) |
Nov 11 2024 | 2 years to revive unintentionally abandoned end. (for year 8) |
Nov 11 2025 | 12 years fee payment window open |
May 11 2026 | 6 months grace period start (w surcharge) |
Nov 11 2026 | patent expiry (for year 12) |
Nov 11 2028 | 2 years to revive unintentionally abandoned end. (for year 12) |