A method for determining disjoint multi-paths in a data center fabric network system, according to one embodiment, includes creating a topology of a data center fabric network system, wherein the system comprises a plurality of switches; determining paths between the plurality of switches; calculating one or more disjoint multi-paths from a source device to a destination device in the system; and writing the one or more disjoint multi-paths into a forwarding table of at least one of: the source device and a nearest shared switch to the source device.
|
13. A method for determining disjoint multi-paths in a data center fabric network system, the method comprising:
receiving, at a switch controller, reports of nearest neighbor devices from each of a plurality of switches in a data center fabric network system;
creating a topology of the data center fabric network system using the switch controller, wherein the data center fabric network system comprises the plurality of switches and the switch controller coupled directly to the plurality of switches via secure communication links, the plurality of switches being restricted from executing a discovery protocol to determine paths to other switches in the system;
determining paths between the plurality of switches, provided that the plurality of switches do not themselves execute a discovery protocol to determine paths to other switches in the system;
storing the determined paths, using the switch controller, to a switch forwarding table in each switch of the plurality of switches;
calculating one or more disjoint multi-paths from a source device to a destination device in the system using the switch controller, wherein each of the one or more disjoint multi-paths comprise at least two paths that are disjoint from one another to as great of an extent as possible given the topology of the data center fabric network system resulting in at least one fully redundant path between the source device and the destination device; and
writing, using the switch controller, the one or more disjoint multi-paths into at least a forwarding table of a nearest shared switch to the source device that is part of at least one disjoint multi-path.
1. A method for determining disjoint multi-paths in a data center fabric network system, the method comprising:
creating a topology of a data center fabric network system using a switch controller, wherein the data center fabric network system comprises a plurality of switches and the switch controller coupled directly to the plurality of switches via secure communication links, the plurality of switches being restricted from executing a discovery protocol to determine paths to other switches in the data center fabric network system;
determining paths between the plurality of switches using path calculations performed by the switch controller;
calculating one or more disjoint multi-paths from a source device to a destination device using the switch controller based on the topology, the calculating comprising determining whether a fully disjoint multi-path is available from the source device to the destination device, the calculated one or more disjoint multi-paths being fully disjoint based on a determination that a fully disjoint multi-path is available from the source device to the destination device and being partially disjoint based on a determination that a fully disjoint multi-path is not available from the source device to the destination device;
writing, using the switch controller, the one or more disjoint multi-paths into a forwarding table of the source device, the source device being configured to communicate with the switch controller during the writing of the one or more disjoint multi-paths into the forwarding table; and
writing, using the switch controller, the one or more disjoint multi-paths into a switch forwarding table of a nearest shared switch to the source device so that all traffic sent from the source device is routed through the nearest shared switch.
2. The method as recited in
3. The method as recited in
calculating, using the switch controller, at least one disjoint multi-path between each switch and all other switches in the system; and
writing the at least one disjoint multi-path into a switch forwarding table of each switch using the switch controller.
4. The method as recited in
detecting a change in the system using the switch controller;
re-calculating at least one disjoint multi-path between each switch and all other switches in the system in response to the change in the system, the re-calculating comprising determining whether a fully disjoint multi-path is available from the source device to the destination device, the calculated at least one disjoint multi-path being fully disjoint based on a determination that a fully disjoint multi-path is available from the source device to the destination device and being partially disjoint based on a determination that a fully disjoint multi-path is not available from the source device to the destination device; and
writing the at least one disjoint multi-path into the switch forwarding table of each switch using the switch controller,
wherein the change in the system comprises one of: addition of a switch to the system, removal of a switch from the system, and change of a property of a switch in the system.
5. The method as recited in
6. The method as recited in
detecting a change in the system affecting a disjoint multi-path between the source device and a destination device using the switch controller;
re-calculating one or more disjoint multi-paths from the source device to the destination device in response to the change in the system, the re-calculating comprising determining whether a fully disjoint multi-path is available from the source device to the destination device, the calculated one or more disjoint multi-paths being fully disjoint based on a determination that a fully disjoint multi-path is available from the source device to the destination device and being partially disjoint based on a determination that a fully disjoint multi-path is not available from the source device to the destination device; and
writing, using the switch controller, the one or more re-calculated disjoint multi-paths into the forwarding table of at least one of: the source device and the nearest shared switch to the source device,
wherein the change in the system comprises one of: addition of a switch to the system, removal of a switch from the system, and change of a property of a switch in the system.
7. The method as recited in
8. The method as recited in
9. The method as recited in
10. The method as recited in
11. The method as recited in
12. The method as recited in
14. The method as recited in
calculating at least one disjoint multi-path between each switch and all other switches in the system; and
writing the at least one disjoint multi-path into a switch forwarding table of each switch.
15. The method as recited in
detecting a change in the system;
re-calculating at least one disjoint multi-path between each switch and all other switches in the system in response to the change in the system; and
writing the at least one disjoint multi-path into the switch forwarding table of each switch,
wherein the change in the system comprises one of: addition of a switch to the system, removal of a switch from the system, and change of a property of a switch in the system.
16. The method as recited in
detecting a change in the system affecting a disjoint multi-path between the source device and a destination device;
re-calculating one or more disjoint multi-paths from the source device to the destination device in response to the change in the system; and
writing the one or more re-calculated disjoint multi-paths into the forwarding table of at least one of: the source device and the nearest shared switch to the source device,
wherein the change in the system comprises one of: addition of a switch to the system, removal of a switch from the system, and change of a property of a switch in the system.
17. The method as recited in
|
This application is a continuation of copending U.S. patent application Ser. No. 13/407,658, filed Feb. 28, 2012, which is herein incorporated by reference.
The present invention relates to Ethernet switches and switching, and more particularly, this invention relates to disjoint multi-pathing for a data center network.
All fabric types (e.g., networking, storage, and clustering) used in data centers make use of support for multiple redundant paths through the network that help to enable high availability of critical applications. In particular, storage traffic, e.g., traffic over Fibre Channel (FC), internet small computer system interface (iSCSI), and/or FC over Ethernet (FCoE), has traditionally used this approach to insure continuous availability of all the storage resources in a data center. Similar requirements may be found in traditional Ethernet networks or server clusters, e.g., remote direct memory access (RDMA), RDMA over converged Ethernet (RoCE), internet wide area RDMA protocol (iWarp), or other similar RDMA protocols. This requirement may be achieved by implementing more than one physically separate network, each network having redundant switches and links. A significant disadvantage of this approach is the administrative overhead associated with configuring, managing, and maintaining at least two separate networks.
This may be somewhat overcome if a single network fabric is implemented which supports multi-pathing. Such fabrics use an alternative to spanning tree protocol (STP), which forces a single path through the network and thus under-utilizes network bandwidth (in the case of bridged domains, STP forces a single multi-cast tree to be available).
There are different types of multi-pathing, depending on whether the multiple paths through the network share common network devices or whether they are completely disjointed from each other. The highest levels of availability are provided by fabrics which support disjoint multi-pathing, preferably in a flat fabric architecture which facilitates virtual machine (VM) migration in highly virtualized data centers.
There are several possible approaches to support disjoint multi-pathing using open, industry standard protocols, all of which have drawbacks. For example, a single transparent interconnection of lots of links (TRILL) fabric covering the entire data center would not function properly unless enhancements to the currently available industry standards are made in order to support disjoint multi-pathing and a flat Layer 2 fabric. Further, there may be some concern as to implementation of TRILL that use proprietary type-length-value (TLV) encoding or which are not based on intermediate system to intermediate system (IS-IS). TRILL uses different paths for unicast and multicast traffic compared with shortest path bridging (SPB). For the vast majority of Internet protocol (IP) traffic, this should not make a difference. However, care should be taken for latency sensitive applications.
Other proposed standards such as SPB would use additional configuration steps to ensure that different paths are in fact fully disjoint and that the server adaptors select the correct paths to use. Further, by using either of TRILL or SPB, a method to ensure redundancy at the network interface controller (MC) or host bus adapter (HBA) and the first hop into the network would be necessary. Thus, the industry is currently lacking a mature standard-based solution to this problem.
In one embodiment, a method for determining disjoint multi-paths in a data center fabric network system includes creating a topology of a data center fabric network system, wherein the system includes a plurality of switches, determining paths between the plurality of switches, calculating one or more disjoint multi-paths from a source device to a destination device in the system, and writing the one or more disjoint multi-paths into a forwarding table of at least one of: the source device and a nearest shared switch to the source device.
Other aspects and embodiments of the present invention will become apparent from the following detailed description, which, when taken in conjunction with the drawings, illustrate by way of example the principles of the invention.
The following description is made for the purpose of illustrating the general principles of the present invention and is not meant to limit the inventive concepts claimed herein. Further, particular features described herein can be used in combination with other described features in each of the various possible combinations and permutations.
Unless otherwise specifically defined herein, all terms are to be given their broadest possible interpretation including meanings implied from the specification as well as meanings understood by those skilled in the art and/or as defined in dictionaries, treatises, etc.
It must also be noted that, as used in the specification and the appended claims, the singular forms “a,” “an,” and “the” include plural referents unless otherwise specified.
In one approach, a data center fabric using a single controller for managing the control plane, such as an OpenFlow controller, may support both shared and disjoint multi-pathing and may provide a faster time to market compared with alternative systems currently available. Accordingly, implementation of disjoint multi-pathing using OpenFlow Controllers (OFCs), or other controllers that facilitate consolidation of the control plane at a central place, may be configured to make use of novel software extensions to existing protocol, which are discussed in more detail herein.
According to one general embodiment, a data center fabric network system includes a controller, a plurality of switches connected to the controller via a plurality of communication links, a source device connected to at least a first switch of the plurality of switches, a destination device connected to at least a second switch of the plurality of switches, and wherein the controller is adapted for: creating a topology of the data center fabric network system, determining paths between the plurality of switches, calculating one or more disjoint multi-paths from the source device to the destination device based on the topology, and writing the one or more disjoint multi-paths into a forwarding table of at least one of: the source device, the first switch, and a nearest shared switch to the source device.
In another general embodiment, a computer program product includes a computer readable storage medium having computer readable program code embodied therewith, the computer readable program code including computer readable program code configured to create a topology of a data center fabric network system, wherein the system includes a plurality of switches, computer readable program code configured to determine paths between each switch and all other switches of the system, computer readable program code configured to calculate one or more disjoint multi-paths from a source device to a destination device in the system, and computer readable program code configured to write the one or more disjoint multi-paths into a forwarding table of at least one of the source device and a nearest shared switch to the source device.
In yet another general embodiment, a method for determining disjoint multi-paths in a data center fabric network system includes creating a topology of a data center fabric network system, wherein the system includes a plurality of switches, determining paths between the plurality of switches, calculating one or more disjoint multi-paths from a source device to a destination device in the system, and writing the one or more disjoint multi-paths into a forwarding table of at least one of: the source device and a nearest shared switch to the source device.
As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as “logic,” a “circuit,” “module,” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
Any combination of one or more computer readable medium(s) may be utilized. The computer readable medium may be a computer readable signal medium or a non-transitory computer readable storage medium. A non-transitory computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the non-transitory computer readable storage medium include the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a portable compact disc read-only memory (CD-ROM), a Blu-ray disc read-only memory (BD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a non-transitory computer readable storage medium may be any tangible medium that is capable of containing, or storing a program or application for use by or in connection with an instruction execution system, apparatus, or device.
A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a non-transitory computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device, such as an electrical connection having one or more wires, an optical fibre, etc.
Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fibre cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++, or the like, and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on a user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer or server may be connected to the user's computer through any type of network, including a local area network (LAN), storage area network (SAN), and/or a wide area network (WAN), or the connection may be made to an external computer, for example through the Internet using an Internet Service Provider (ISP).
Aspects of the present invention are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatuses (systems), and computer program products according to various embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, may be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer readable medium that may direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
In use, the gateway 101 serves as an entrance point from the remote networks 102 to the proximate network 108. As such, the gateway 101 may function as a router, which is capable of directing a given packet of data that arrives at the gateway 101, and a switch, which furnishes the actual path in and out of the gateway 101 for a given packet.
Further included is at least one data server 114 coupled to the proximate network 108, and which is accessible from the remote networks 102 via the gateway 101. It should be noted that the data server(s) 114 may include any type of computing device/groupware. Coupled to each data server 114 is a plurality of user devices 116. Such user devices 116 may include a desktop computer, laptop computer, handheld computer, printer, and/or any other type of logic-containing device. It should be noted that a user device 111 may also be directly coupled to any of the networks, in some embodiments.
A peripheral 120 or series of peripherals 120, e.g., facsimile machines, printers, scanners, hard disk drives, networked and/or local storage units or systems, etc., may be coupled to one or more of the networks 104, 106, 108. It should be noted that databases and/or additional components may be utilized with, or integrated into, any type of network element coupled to the networks 104, 106, 108. In the context of the present description, a network element may refer to any component of a network.
According to some approaches, methods and systems described herein may be implemented with and/or on virtual systems and/or systems which emulate one or more other systems, such as a UNIX system which emulates an IBM z/OS environment, a UNIX system which virtually hosts a MICROSOFT WINDOWS environment, a MICROSOFT WINDOWS system which emulates an IBM z/OS environment, etc. This virtualization and/or emulation may be enhanced through the use of VMWARE software, in some embodiments.
In more approaches, one or more networks 104, 106, 108, may represent a cluster of systems commonly referred to as a “cloud.” In cloud computing, shared resources, such as processing power, peripherals, software, data, servers, etc., are provided to any system in the cloud in an on-demand relationship, thereby allowing access and distribution of services across many computing systems. Cloud computing typically involves an Internet connection between the systems operating in the cloud, but other techniques of connecting the systems may also be used, as known in the art.
The workstation shown in
The workstation may have resident thereon an operating system such as the MICROSOFT WINDOWS Operating System (OS), a MAC OS, a UNIX OS, etc. It will be appreciated that a preferred embodiment may also be implemented on platforms and operating systems other than those mentioned. A preferred embodiment may be written using JAVA, XML, C, and/or C++ language, or other programming languages, along with an object oriented programming methodology. Object oriented programming (OOP), which has become increasingly used to develop complex applications, may be used.
Referring now to
In some approaches, tunnels (not shown) may be established between virtual networks 304, 306 and/or between at least one virtual network 304, 306 and one or more VMs 310.
Each virtual network 304, 306 may include any number of virtual machines (VMs) 308, 310. In one embodiment, Virtual Network A 304 includes one or more VMs 308, and Virtual Network B 306 includes one or more VMs 310. As shown in
Now referring to
The standby device 404 is provided in the DFP 400 so that should connectivity to the system master 402 be lost for any reason, or if the system master 402 fails for any reason, the standby device 404 is adapted to become the system master and provides all functionality of a system master.
The system master 402 is responsible for determining services used for a particular traffic flow (movement of data/requests/alerts or the like through the DFP 400) and redirects it through a series of members 406, as appropriate based on the particular traffic flow involved, as would be understood by one of skill in the art.
In one embodiment, the topology of the DFP 400, being a cell switched Clos, provides high availability of any links/devices on the DFP 400. Connectivity details and cross-sectional bandwidth may be implementation specific and may be determined on an application specific basis, as would be understood by one of skill in the art.
From a networking perspective, the entire domain of the DFP 400 from outside may appear like an end host, where some services are being applied to traffic flows. The backplane of the topology, being a cell switched interface as described before, removes the possibility of endless looping and it efficiently uses the available backplane bandwidth by spreading cells over multiple links for a single destination, according to one embodiment. Individual members 406 may be added/removed in the domain of the DFP 400 without disrupting other services, devices, etc.
According to one embodiment, up to about 800 members may be grouped together in the DFP 400. This allows for about 32,768 Ethernet ports in a single domain of a DFP 400, in one example. In a further embodiment, with the architecture described herein according to various embodiments, constant latency of about 2.5 microseconds or less is possible along with the large number of Ethernet ports possible in one traffic management domain. In one example, the DFP 400 may be envisioned as a single switch having over 32,000 Ethernet ports.
Accordingly, in addition to integrating the plurality of Ethernet ports into the DFP 400, all of the requirements of a data center are also integrated into the DFP 400, in some embodiments. Therefore, if a new extremely low latency, flat data center is being initiated, the DFP 400 helps with virtualization, because most of the virtualized data centers, especially hypervisor-based servers, run a plurality of VMs. By having a “flat switch” architecture, the plurality of VMs may move data across the architecture easily and efficiently. It is very difficult for VMs, generally, to move across layer 3-domains, between one subnet to another subnet, internet protocol (IP) subnet to IP subnet, etc. But if the architecture is similar to a large flat switch, in a very large layer 2-domain, then the VMs are aided in their attempt to move data across the architecture.
Now referring to
Each of the switches and devices of the system 500 may be connected to one another via connections 516, which may allow communications via Ethernet, fibre channel (FC), peripheral component interconnect (PCI), etc.
In one embodiment, the controller 502 is adapted for: creating a topology of the system 500, determining paths between all switches of the plurality of switches, calculating one or more disjoint multiple paths (multi-paths) from the source device 512 to the destination device 514 based on the topology, and writing the one or more disjoint multi-paths into a forwarding table of at least one of the source device 512, the first switch, and a nearest shared switch to the source device 512 (which in
If the source device is a switch, then the one or more disjoint multi-paths may be stored into the forwarding table of the source device 512, as it would be capable of using the information to properly forward data along the one or more disjoint paths. If not, then the one or more disjoint multi-paths may be stored into the forwarding table of switches 504 and/or 510, depending on whether one or each of switches 504, 510 are members of the disjoint path between the source device 512 and the destination device 514.
According to embodiments described herein, each of the one or more disjoint multi-paths may comprise at least two paths that are disjoint from one another resulting in at least one fully redundant path. In more embodiments, each disjoint multi-path may comprise at least two paths that are partially disjoint from one another (possibly sharing one or more elements in the paths) resulting in at least one partially redundant path (that may share one or more elements in the path).
In another approach, a control plane for each of the plurality of switches may be controlled by the controller 502, thereby reducing processing demands on each of the plurality of switches in the system 500 and centralizing control over the switching network.
As shown in
In one embodiment, the controller may be further adapted to store the determined paths to a forwarding table in each switch, thereby allowing each switch to forward data along designated paths through the system 500.
According to one embodiment, the controller 502 may be adapted to operate as an OpenFlow controller. This allows the controller 502 to operate with certain advantages over other operational systems, as explained in more detail herein. The OpenFlow controller may operate in conjunction with OpenFlow-enabled switches, such as the IBM BNT RACKSWITCH 8264, which is a 40G switch.
In some advantageous embodiments, the plurality of switches do not themselves execute a discovery protocol to determine paths to other switches in the system 500. This frees up processing capacity for each switch, and reduces processing overheard for the entire system 500, due to the centralized nature of the path calculations being performed on the controller 502.
In another embodiment, the controller 502 may be further adapted for calculating at least one disjoint multi-path between each switch and all other switches in the system 500, and writing the at least one disjoint multi-path into a forwarding table of each switch in the system 500.
For example, for switch 504, the disjoint multi-paths to destination device 514 are 504-506-514 and 504-508-514. These disjoint multi-paths may be stored to the forwarding table for this switch, such that this switch is capable of directing traffic on disjoint paths whenever applicable, as would be known to one of skill in the art. Then, the same calculations are carried out for all other switches in the system 500, and the disjoint multi-paths are stored to a forwarding table for each switch in the system 500, thereby allowing each switch to forward traffic on disjoint paths whenever applicable.
In another embodiment, the controller 502 may be further adapted for detecting a change in the system 500 (such as the addition of a switch to the system 500, the removal of a switch from the system 500, change of a property of a switch in the system 500, etc.), re-calculating at least one disjoint multi-path between each switch and all other switches in the system 500 in response to the change in the system 500, and writing the at least one disjoint multi-path into a forwarding table of each switch.
According to some embodiments, the controller 502 may be further adapted for: detecting a change in the system 500 affecting a disjoint multi-path between the source device 512 and a destination device 514, re-calculating one or more disjoint multi-paths from the source device 512 to the destination device 514 in response to the change in the system 500, and writing the one or more re-calculated disjoint multi-paths into the forwarding table of at least one of: the source device 512, the nearest shared switch to the source device 512, and the first switch, as discussed previously.
According to more approaches, the one or more disjoint multi-paths from the source device 512 to the destination device 514 may be calculated in response to a request from the source device 512 or the nearest shared switch to the source device 512. This request may be initiated due to data intending to be sent from the source device 512 to somewhere in the system 500.
As shown in
Referring now to
First, each switch in the fabric network 500 is connected to an out of band management network. In some preferred embodiments, an OpenFlow controller 502 is connected to the switches over a secure communication link 518, using switches which support OpenFlow protocols. This implementation is relatively easy, since many data center networks currently make use of inexpensive, relatively low data rate (1 Gbit/s) Ethernet outband networks for management of more expensive switch infrastructures.
Second, all switches in the fabric network 500 are instructed by the OpenFlow controller 502 to turn off Layer 2 forwarding standards, such as STP, etc. Third, all switches in the fabric network 500 enable reporting of their nearest adjacent neighbors to the OpenFlow controller 502.
Under these conditions, in some embodiments, it becomes possible for the OpenFlow controller 502 to calculate the topology of the fabric network 500 and then load appropriate forwarding tables into the switches. The OpenFlow controller 502 may contain a topology manager and may be capable of network topology discovery (similar to the IS-IS and “hello” protocols used by TRILL and other competing approaches). Through extensions of the link layer discovery protocol (LLDP) and protocol data units (PDUs), the OpenFlow controller 502 may create a topology map of the network 500. The controller 502 then calculates possible paths through the network 500, which are communicated to OpenFlow enabled switches as forwarding table entries in each switch.
The OpenFlow controller 502 may calculate disjoint multi-paths from a given source 512 to a given destination 514 (depending on the size of the network 500 and speed of the calculation, there are several alternative embodiments; for example, the controller 502 may calculate all possible disjoint multi-paths upon initialization of the network 500, it may only calculate paths from a specific source to a specific destination upon receiving a request from a server, etc.). The disjoint multi-paths may be calculated using various algorithms, as known in the art.
For example, the controller 502 may begin the calculation with one initial path and search all other possible paths in a serial fashion, discarding those which share a common element with the initial path. Alternately, using a multi-thread processor, the controller 502 may begin with one initial path and search other possible paths in parallel to identify those paths which have a first switch in common with the initial path, a second switch in common with the initial path, etc., until the only remaining paths have no elements in common with the initial path (are disjoint of one another). Of course, other variations would be apparent to those skilled in the art, such as graph theory algorithms for calculating paths through a network.
The resulting disjoint multi-paths are then written into the forwarding tables of the switches. In the event a switch is added, removed, or changed in the fabric network 500, this calculation of the topology map and disjoint multi-paths may be repeated, and a new set of forwarding tables may be written to the switches by the OpenFlow controller 502. It is expected that reconvergence times will be faster under these conditions when compared with alternative solutions, since the recalculation is centralized in the OpenFlow controller 502 rather than being distributed across each switch in the network 500.
Note that since topology discovery is enabled directly from OpenFlow, in some approaches, the switches do not nun other discovery protocols such as MSTP, TRILL, etc. This is an advantage of using the centralized controller 502 for topology discovery.
All Ethernet address resolution protocols (ARPs) and fibre channel over Ethernet interne protocol (HP) requests are routed to the OpenFlow controller 502, with request for disjoint multi-pathing support as requested. If disjoint multi-pathing is requested (for example, through a server element manager interface, such as TIVOLI interfacing to the OpenFlow controller 502, an enhancement to ARP which requests disjoint multi-pathing, etc.), the OpenFlow controller 502 may select appropriate disjoint paths through the network. Thus, the OpenFlow controller 502 may also include a mechanism for detecting that a given physical server is requesting disjoint multi-pathing as described above. The OpenFlow controller 502 may also include a mechanism for setting up disjoint paths through the appropriate flow tables in the switches, and a mechanism for responding to the server's request for disjoint paths. In this manner, a single network fabric with one management interface may be created which supports multi-pathing features that were formerly available only through management of multiple separate networks.
There are several alternative embodiments presented below. For example, while this approach describes a centralized OpenFlow control plane, it is also possible for the OpenFlow control plane to be distributed, or to be shared among redundant OpenFlow controllers (two or more controllers) attached to the network (subject to any limitations of the OpenFlow protocol). In another embodiment, the OpenFlow controller may not be restricted to equal cost or shortest-hop multi-path calculations; instead, multiple paths (disjoint or not) may be calculated using some other basis (for example, multiple paths with significantly different path lengths may be used). Alternatives such as TRILL and SPB do not support this option in their multi-path calculations.
In order for the server to place appropriate traffic onto the disjoint multi-paths through the network, this traffic is directed to appropriate ports of the server network adapter (NIC) or host bus adapter (HBA). Today, the OpenFlow specification mainly addresses the control plane rather than the network management functions. Most servers provide at least two physical connections which are multi-homed to the network. The server port may be mapped to the first network switch through a new interface within the network element manager or with the management plane in a Software Defined Network (SDN) implementation. One possible implementation may be that the host requesting disjoint multi-paths advertises to the controller the address of the two physical connections. The controller in turn indicates to the host the disjoint multi-paths.
This is represented in
Conversely, if the destination device 514 (Q) requests disjoint multi-paths from the controller 502, the same actions may be taken. For example, the destination device 514 (Q) may request the disjoint multi-paths from the controller 502 by indicating its own ports (Q→Controller: My addresses (Q1, Q2)). In response, the controller 502 indicates the disjoint multi-paths to the destination device 514 (Controller→Q: Disjoint Paths (P1-Q1, P2-Q2)).
In the case where a virtual switch provides multiple data paths from the hypervisor to virtual NICs or HBAs, the virtual switch may participate in the OpenFlow controller topology discovery.
Note that all other Layer 2 routing in the network 500 (such as TRILL and SPB) is disabled. OpenFlow allows the potential to route traffic and compute topology maps based on Layer 2, Layer 3, or a combination of both parameters. If a preferred embodiment uses only Layer 2 features for OpenFlow topology discovery and routing, the resulting calculations will be transparent to other network protocols which rely exclusively on Layer 3, such as Distributed Overlay Virtual Ethernet (DOVE). This is an advantage, since DOVE (or similar Layer 3 variants such as VXLAN) use equal cost multi-pathing and have mechanisms to isolate multi-tenant traffic, but do not provide the control required to insure disjoint multi-pathing.
In another embodiment, the OpenFlow controller may interact with DOVE to enhance its performance. For example, DOVE may treat OpenFlow as another lower level network option and run on top of OpenFlow. When a DOVE switch queries the learning service for the location of a VM, the learning service (or the DOVE switch) may, at the same time, make a request to the OpenFlow controller to implement a disjoint multi-path. Another addition would be the ability of DOVE to request that the OpenFlow controller set a path for certain traffic flows so that they go through a firewall, some other security appliance, a traffic conditioning device, etc. (the learning service may make this request). In this case, OpenFlow may be enhanced to inspect deeper into the packets and choose a path based on inner headers, rather than on the encapsulated headers, as not all traffic between two DOVE switches have the same security policy or other such requirements.
Now referring to
Now referring to
Each of the steps of the method 800 may be performed by any suitable component of the operating environment. For example, in one embodiment, the method 800 may be partially or entirely performed by the controller, in one approach, as described herein in more detail.
Optional operations described below are those operations which may be omitted, or may be performed once while other operations are performed multiple tunes, or may be performed by other entities are systems and relied upon by other operations, according to various embodiments.
As shown in
In operation 804, paths are determined between all switches of the plurality of switches, thereby allowing discovery of paths through the network.
In optional operation 806, the determined paths are stored to a forwarding table in each switch of the plurality of switches. This allows the switches to forward data to the appropriate ports.
In operation 808, one or more disjoint multi-paths are calculated from a source device to a destination device in the system. The source device and/or the destination device may be a switch, a host, a server, etc. Also, the one or more disjoint multi-paths may be fully or partially disjoint, depending on the topology of the network.
In operation 810, the one or more disjoint multi-paths are written into a forwarding table of at least one of the source device and a nearest shared switch to the source device.
In one approach, the plurality of switches do not themselves execute a discovery protocol to determine paths to other switches in the system.
In one embodiment, the method 800 may further comprise calculating at least one disjoint multi-path between each switch and all other switches in the system and writing the at least one disjoint multi-path into a forwarding table of each switch. This may be performed at initialization of the network, according to preferred approaches. In this way, when a request to send data over disjoint paths is received, no further calculations are necessary to report the disjoint paths, as they are already calculated.
In further approaches, the method 800 may include detecting a change in the system, re-calculating at least one disjoint multi-path between each switch and all other switches in the system in response to the change in the system, and writing the at least one disjoint multi-path into a forwarding table of each switch. In some approaches, the change in the system may comprise one of addition of a switch to the system, removal of a switch from the system, and change of a property of a switch in the system. This is applicable to the forwarding table of each switch in the network.
In a further embodiment, the method 800 may include detecting a change in the system affecting a disjoint multi-path between the source device and a destination device, re-calculating one or more disjoint multi-paths from the source device to the destination device in response to the change in the system, and writing the one or more re-calculated disjoint multi-paths into the forwarding table of at least one of: the source device and the nearest shared switch to the source device. This is applicable to when only a calculation has been performed in response to a request for disjoint multi-paths.
For example, in one approach, the one or more disjoint multi-paths from the source device to the destination device may be calculated in response to a request from the source device or the nearest shared switch to the source device.
In one embodiment, each of the one or more disjoint multi-paths may comprise at least two paths that are disjoint from one another resulting in at least one fully redundant path.
In another approach, a control plane for each of the plurality of switches may be controlled by a controller of the system, which may be adapted to operate as an OpenFlow controller.
Of course, any aspect of the method 800 may be executed as a computer program product.
For example, in one embodiment, a computer program product may comprise a computer readable storage medium having computer readable program code embodied therewith. The computer readable program code may comprise computer readable program code configured to create a topology of a data center fabric network system, wherein the system comprises a plurality of switches; computer readable program code configured to determine paths between each switch and all other switches of the system; computer readable program code configured to calculate one or more disjoint multi-paths from a source device to a destination device in the system; and computer readable program code configured to write the one or more disjoint multi-paths into a forwarding table of at least one of: the source device and a nearest shared switch to the source device.
Of course, any of the other previously described embodiments and/or approaches may be implemented in the computer program product, as would be understood by one of skill in the art upon reading the present descriptions.
While various embodiments have been described above, it should be understood that they have been presented by way of example only, and not limitation.
Thus, the breadth and scope of an embodiment of the present invention should not be limited by any of the above-described exemplary embodiments, but should be defined only in accordance with the following claims and their equivalents.
DeCusatis, Casimer M., Kamble, Keshav G., Recio, Renato J., Kumbhare, Abhijit P., Singh, Simarbir
Patent | Priority | Assignee | Title |
10103974, | Jul 12 2016 | National Tsing Hua University | Software-defined network controller and multipath routing method |
9455899, | Feb 28 2012 | KYNDRYL, INC | Disjoint multi-pathing for a data center network |
Patent | Priority | Assignee | Title |
7349350, | Sep 23 2003 | Intel Corporation | Determining two node-disjoint paths using on-demand flooding |
7567547, | Sep 07 2006 | Xerox Corporation | Method and system for loop-free ad-hoc routing |
8432913, | Apr 22 2008 | NEC Corporation | Relay device, network system, route switching method, and recording medium |
20030033427, | |||
20030130821, | |||
20050071482, | |||
20060062199, | |||
20080189769, | |||
20080285439, | |||
20110019539, | |||
20110261722, | |||
20110261825, | |||
20110267954, | |||
20110267979, | |||
20110286324, | |||
20110286326, | |||
20110305168, | |||
20110310734, | |||
20120023231, | |||
20130039214, | |||
20130071116, | |||
20130208621, | |||
20130223440, | |||
WO2009042919, | |||
WO2011118586, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Feb 27 2012 | DECUSATIS, CASIMER M | International Business Machines Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036499 | /0924 | |
Feb 28 2012 | KAMBLE, KESHAV G | International Business Machines Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036499 | /0924 | |
Feb 28 2012 | KUMBHARE, ABHIJIT P | International Business Machines Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036499 | /0924 | |
Feb 28 2012 | KUMBHARE, ABHIJIT | International Business Machines Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036809 | /0368 | |
Apr 10 2012 | SINGH, SIMARBIR | International Business Machines Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036499 | /0924 | |
May 30 2012 | International Business Machines Corporation | (assignment on the face of the patent) | / | |||
Jun 18 2012 | RECIO, RENATO J | International Business Machines Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036499 | /0924 | |
Sep 30 2021 | International Business Machines Corporation | KYNDRYL, INC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 057885 | /0644 |
Date | Maintenance Fee Events |
Apr 17 2019 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Apr 18 2023 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Date | Maintenance Schedule |
Nov 03 2018 | 4 years fee payment window open |
May 03 2019 | 6 months grace period start (w surcharge) |
Nov 03 2019 | patent expiry (for year 4) |
Nov 03 2021 | 2 years to revive unintentionally abandoned end. (for year 4) |
Nov 03 2022 | 8 years fee payment window open |
May 03 2023 | 6 months grace period start (w surcharge) |
Nov 03 2023 | patent expiry (for year 8) |
Nov 03 2025 | 2 years to revive unintentionally abandoned end. (for year 8) |
Nov 03 2026 | 12 years fee payment window open |
May 03 2027 | 6 months grace period start (w surcharge) |
Nov 03 2027 | patent expiry (for year 12) |
Nov 03 2029 | 2 years to revive unintentionally abandoned end. (for year 12) |