An example embodiment of the present invention provides processes relating to a connection/communication protocol and a memory-addressing scheme for a distributed shared memory system. In the example embodiment, a logical node identifier comprises bits in the physical memory addresses used by the distributed shared memory system. Processes in the embodiment include logical node identifiers in packets which conform to the protocol and which are stored in a connection control block in local memory. By matching the logical node identifiers in a packet against the logical node identifiers in the connection control block, the processes ensure reliable delivery of packet data. Further, in the example embodiment, the logical node identifiers are used to create a virtual server consisting of multiple nodes in the distributed shared memory system.

Patent
   RE44610
Priority
Apr 26 2007
Filed
May 10 2012
Issued
Nov 26 2013
Expiry
Apr 26 2027
Assg.orig
Entity
Large
62
9
all paid
0. 16. A method comprising:
receiving, at a first node in a distributed shared memory system, a message from a second node in the distributed shared memory system, the distributed shared memory system comprising a plurality of interconnected nodes each having a unique logical node identifier, wherein the message indicates a memory operation related to a local memory of the first node and identifies a memory address;
if a first plurality of contiguous bits of the memory address equal a logical node identifier of the first node, changing the first plurality of contiguous bits to a predetermined value;
if the first plurality of contiguous bits of the memory address equal the predetermined value, changing the first plurality of contiguous bits to the logical node identifier of the first node; and
forwarding the message to a processor of the first node for processing.
0. 21. A method comprising:
receiving, at a first node in a distributed shared memory system, a message from a processor of the first node identifying a memory operation related to a local memory of a second node in the distributed shared memory system, the distributed shared memory system comprising a plurality of nodes each having a unique logical unit identifier, the plurality of nodes being interconnected by a switch fabric, wherein the message identifies a memory address;
if a first plurality of contiguous bits of the memory address equal a logical node identifier of the first node, changing the first plurality of contiguous bits to a predetermined value;
if the first plurality of contiguous bits of the memory address equal the predetermined value, changing the first plurality of contiguous bits to the logical node identifier of the first node; and
forwarding the message to the second node for processing.
4. A method, comprising:
receiving, at a distributed memory logic circuit of a first node, a packet from a distributed memory logic circuit of a second node, wherein the packet includes a source logical node identifier and wherein the first and second nodes are connected by a network switch fabric and are parts of a distributed shared memory system;
determining whether a destination switch fabric address included in the packet matches a switch fabric address for the first node;
using the source logical node identifier as an index into a connection control block to locate an entry for the a connection between the first and second nodes, resulting in a located entry of the connection control block, wherein the connection control block is stored in a local memory on the first node;
determining whether a destination logical node identifier included in the packet matches a logical node identifier for the first node, wherein the logical node identifier for the first node is identified in the located entry of the connection control block; and
accepting data in the packet for further processing by the first node.
11. A distributed memory logic circuit encoded with executable logic, the logic when executed operable to:
receive, at the distributed memory logic circuit of a first node, a packet from a distributed memory logic circuit of a second node, wherein the packet includes a source logical node identifier and wherein the first and second nodes are connected by a network switch fabric and are parts of a distributed shared memory system;
determine whether a destination switch fabric address included in the packet matches a switch fabric address for the first node;
use the source logical node identifier as an index into a connection control block to locate an entry for a connection between the first and second nodes, resulting in a located entry of the connection control block, wherein the connection control block is stored in a local memory on the first node;
determine whether a destination logical node identifier included in the packet matches a logical node identifier for the first node, wherein the logical node identifier for the first node is identified in the located entry of the connection control block; and
accept data in the packet for further processing by the first node.
0. 22. A distributed shared memory system, comprising:
a network switch fabric; and
a plurality of nodes interconnected by the network switch fabric, each given node of the plurality of nodes comprising:
a logical node identifier of a plurality of contiguous bits;
a local memory;
a distributed shared memory management chip operative to share the local memory of the given node with others of the plurality of nodes in the distributed shared memory system to create a shared memory accessible using binary addresses comprising a plurality of bits, wherein a set of contiguous most-significant bits of the binary addresses collectively represent a logical node identifier of a node of the plurality of nodes; and
one or more processors each operative to access the local memory of the given node, the local memory accessed using binary addresses having the set of contiguous most-significant bits collectively set to a predetermined value,
wherein the distributed shared memory management chip is further operative to map the predetermined value to the logical node identifier of the given node in memory management traffic transmitted between the plurality of nodes that include one or more binary addresses of the shared memory.
1. A method, comprising:
receiving, at a distributed memory logic circuit of a first node, data for a packet destined to a distributed memory logic circuit of a second node, wherein the first and second nodes are connected by a network switch fabric and are parts of a distributed shared memory system, and wherein the data for the packet includes a physical memory address in which one or more bits in the physical memory address comprise a destination logical node identifier for the second node;
using the destination logical node identifier as an index into a connection control block to locate an entry for a connection between the first and second nodes, resulting in a located entry of the connection control block, wherein the connection control block is stored in a local memory on the first node;
building a the packet in a format of a connection and communication protocol using the data, the destination logical node identifier, and a logical node identifier for the first node, wherein the logical node identifier for the first node is included in the located entry of the connection control block entry;
adding, to the packet, a header that includes a switch fabric address for the second node, wherein the switch fabric address is identified in the located entry of the connection control block; and
transmitting the packet on a link to the switch fabric.
8. A distributed memory logic circuit encoded with executable logic, the logic when executed operable to:
receive, at the distributed memory logic circuit of a first node, data for a packet destined to a distributed memory logic circuit of a second node, wherein the first and second nodes are connected by a network switch fabric and are parts of a distributed shared memory system,; and wherein the data for the packet includes a physical memory address in which one or more bits in the physical memory address comprise a destination logical node identifier for the second node;
use the destination logical node identifier as an index into a connection control block to locate an entry for a connection between the first and second nodes, resulting in a located entry of the connection control block, wherein the connection control block is stored in a local memory on the first node;
build a the packet in a format of a connection and communication protocol using the data, the destination logical node identifier, and a logical node identifier for the first node, wherein the logical node identifier for the first node is included in the located entry of the connection control block entry;
add, to the packet, a header that includes a switch fabric address for the second node, wherein the switch fabric address is identified in the located entry of the connection control block; and
transmit the packet on a link to the switch fabric.
15. A distributed shared memory system comprising:
a network switch fabric;
two or more nodes in a distributed shared memory system connected by a the network switch fabric; and wherein, each of the two or more nodes comprises comprising:
one or more processors,;
local memory; and
a distributed shared memory logic circuit,
wherein the distributed memory logic circuit is encoded with executable logic, the logic that
when executed, is operable to:
receive, at the distributed memory logic circuit of a local node, data for a packet destined to a distributed memory logic circuit of a remote node of the two or more nodes in the distributed shared memory system, wherein the data for the packet includes a physical memory address in which one or more bits in the physical memory address comprise a destination logical node identifier for the remote node,
use the destination logical node identifier as an index into a connection control block to locate an entry for a connection between the local node and the remote node, resulting in a local entry of the connection control block, wherein the connection control block is stored in local memory on the local node,
build a the packet in a format of a connection and communication protocol using the data, the destination logical node identifier, and a logical node identifier for the local node, wherein the logical node identifier for the local node is included in the located entry of the connection control block entry,
add, to the packet, a header that includes a switch fabric address for the remote node, wherein the switch fabric address is identified in the located entry of the connection control block,
transmit the packet on a link to the network switch fabric, receive, at the distributed memory logic circuit of the local node, a second packet from a distributed memory logic circuit of the remote node or another remote node of the two or more nodes in the distributed shared memory system, wherein the second packet includes a source logical node identifier,
determine whether a destination switch fabric address included in the second packet matches a switch fabric address for the local node,
use the source logical node identifier as an index into the connection control block to locate an entry for a connection between the local and remote node, resulting in a second located entry of the connection control block, determine whether a destination logical node identifier included in the second packet matches a the logical node identifier for the local node, wherein the logical node identifier for the local node is identified in the second located entry of the connection control block, and
accept data in the packet for further processing by the local node.
2. A method as in claim 1, wherein the distributed shared memory system is a cache coherent non-uniform memory access system.
3. A method as in claim 1, wherein the a distributed memory logic circuit in the first node sets the destination logical node identifier to zero if the destination logical node identifier in the physical memory address equals the logical node identifier for the first node.
5. The method of claim 4, wherein the packet is discarded if the destination switch fabric address included in the packet does not match the switch fabric address for the first node.
6. The method of claim 4, wherein the packet is discarded if the destination logical node identifier does not match the logical node identifier for the first node identified in the located entry of the connection control block.
7. The method of claim 4, wherein the distributed shared memory system is a cache coherent non-uniform memory access system.
9. The distributed memory logic circuit of in claim 8, wherein the distributed shared memory system is a cache coherent non-uniform memory access system.
10. The distributed memory logic circuit of claim 8, wherein the distributed memory logic circuit of the first node sets logic is further operable to set the destination logical node identifier to zero if the destination logical node identifier in the physical memory address equals the logical node identifier for the first node.
12. The distributed memory logic circuit of claim 11, wherein the packet is discarded if the destination switch fabric address included in the packet does not match the switch fabric address for the first node.
13. The distributed memory logic circuit of claim 11, wherein the packet is discarded if the destination logical node identifier does not match the logical node identifier for the first node identified in the located entry of the connection control block.
14. The distributed memory logic circuit of claim 11, wherein the distributed shared memory system is a cache coherent non-uniform memory access system.
0. 17. The method of claim 16, wherein the predetermined value is zero.
0. 18. The method of claim 16, wherein each node of the plurality of interconnected nodes internally accesses a respective local memory having memory addresses with a first plurality of contiguous bits set to the predetermined value.
0. 19. The method of claim 16, wherein a given node of the plurality of interconnected nodes accesses a local memory of another node of the plurality of interconnected nodes that has a logical unit identifier equal to the predetermined value using the given node's own respective logical node identifier for the another node.
0. 20. The method of claim 16, wherein the memory operation is one of a read command, a write command, or a probe.
0. 23. The distributed shared memory system of claim 22, wherein the distributed shared memory management chip of each node of the plurality of nodes is further operative to:
if the set of contiguous most-significant bits of a given binary address equal the logical node identifier of the given node, change the set of contiguous most-significant bits of the given binary address to the predetermined value; and
if the set of contiguous most-significant bits of the given binary address equal the predetermined value, change the set of contiguous most-significant bits of the given binary address to the logical node identifier of the given node.

identities identifies the packet's destination node. This is the connection identifier (i.e., remote LNID) at the source node. This field is 16 bits wide.

In particular embodiments, the DSM system uses a software data structure called the connection control block (CCB), stored in local memory such as the local main memory shown in FIG. 1, to facilitate implementation of the RDP protocol. The RDM uses a received packet's source LNID as an index into the CCB to find an entry for the connection corresponding to the packet. FIG. 6 is a diagram showing the format of a CCB entry for a single connection, which format might be used in sonic some embodiments of the present invention. As shown in FIG. 6, each entry records the fabric address for two paths, Path 0 and Path 1, which may correspond to the two fabric interface ports shown connected to the RDM in FIG. 2. In other embodiments, there might be more than two paths, corresponding to more than two fabric interface ports. It will be appreciated that the CCB entry has a field called MY_LNID, which identifies the LNID for the RDM's node.

For an RDP connection between a pair of nodes, the node at each end uses an LNID to refer to the node at the other end. Within a multi-node virtual server (VS), every node is assigned a unique LNID, possibly by some management entity for the DSM system. For example, within a three-node VS, the LNID values might be 0, 1, and 2, or 1, 3, and 4, i.e., they not need to be sequentially incrementing from 0. In addition, every server (multi-node virtual server or standalone server) assigns a unique LNID to each node that communicates with it. For example, a standalone server node that communicates with the virtual server described above might be assigned an LNID value of 16 by the VS. If that same node communicates with another server, it may be assigned the same LNID or a different LNID by that server. Therefore, LNID assignments are unique from the standpoint of a given server, but they are not unique across servers.

An example of LNID assignments is shown in FIG. 7. In the example, a virtual computing environment (VCE) consists of two virtual servers (A and B), an application server (C), and a virtual I/O server (D). In this example, virtual server A assigns LNID values 0, 1, and 2 to each of its own nodes (VS nodes A0, A1, and A2, respectively) and an LNID value of 16 to virtual I/O server D. Virtual server B assigns values of 1 and 5 to each of its own nodes (VS nodes B1 and B5, respectively) and an LNID value of 18 to virtual I/O server D. Application server C assigns an LNID value of 3 to virtual I/O server D. Virtual I/O server D assigns LNID values 0, 2, and 4, to VS nodes A0, A1 and A2, respectively, and LNID values of 6 and 8 to VS nodes B1 and B5. Finally, virtual I/O server D assigns a value of 10 to application server C. These various assignments are collected and summarized in Table 7.1 in FIG. 7.

Table 7.2 shows the SrcLNID and DstLNID values used in the headers of RDP packets exchanged between different node pairs. For example, VS nodes A0 and A1 both belong to virtual server A, so a packet from A0 to A1 will have a SrcLNID value of 0 (LNID assigned to A1 by VS A), and a DstLNID value of 1 (LNID assigned to A1 by VS A). As another example, a packet from A1 to I/O server D will have a SrcLNID value of 2 (LNID assigned to A1 by I/O server D) and a DstLNID value of 16 (LNID assigned by VS A to I/O server D).

FIG. 8 is a diagram showing a flowchart of an example process for building an RDP packet for transmission over the switched fabric network, which process might be used with an embodiment of the present invention. In the process's first step 801, the node's Reliable Delivery Manager (RDM) receives a DestLNID and data for an RDP packet from the node's CMM or DMM. The RDM uses the packet's DestLNID to look up the entry corresponding to the DestLNID in the Connection Control Block (CCB), in step 802. If there is no corresponding entry, the RDM sends an error message to the CMM or DMM, as the case may be. Then in step 803, the RDM builds an RDP header for an RDP packet for the data, using the DestLNID and the CCB entry's MY_LNID value. In step 804, the RDM builds a fabric header for the RDP packet, using information in the CCB entry's remote fabric address. Once the RDP packet is complete, the RDM sends the packet to the fabric link for transmission to the remote node, in step 805.

FIG. 9 is a diagram showing a flowchart of an example process for validating an RDP packet received over the switched fabric network, which process might be used with an embodiment of the present invention. In the process's first step 901, a node's RDM receives an RDP packet over the switched fabric network. The RDM then checks to see whether the packet's destination fabric address (e.g., the 6-byte MAC DA in an Ethernet header or the Destination Local ID in an Infiniband LRH) matches the node's fabric address, in step 902. If not, the RDM discards the packet. Otherwise, the RDM goes to step 903 and determines whether the packet is an RDP packet. If not, the RDM will process the packet as a non-RDP packet, in step 904. Otherwise, if the packet is an RDP packet, the RDM uses the packet's SrcLNID to look up the entry corresponding to the SrcLNID in the Connection Control Block (CCB), in step 905. If there is no corresponding entry, the RDM discards the packet. Then the RDM goes to step 906 and checks to make sure that the packet's source fabric address (e.g., the 6-byte MAC SA in an Ethernet header or the Source Local ID in an Infiniband LRH) matches the CCB entry's remote fabric address (e.g., for Path 0 or Path 1). If not, the RDM discards the packet. Otherwise, the RDM checks to determine whether the packet's DestLNID matches the CCB entry's MY_LNID, in step 907. If not, the RDM discards the packet. But if there is a match, the RDM forwards the packet to the CMM or DMM for further processing.

As indicated earlier, the DSM system also uses LNIDs in its memory-addressing scheme. In particular embodiments, the physical memory address width is 40-bits (e.g., in DSM systems that use the present generation of Opteron CPUs), though it will be appreciated that there are numerous other suitable widths. FIG. 10 is a diagram showing the format of a 40-bit physical memory address in a 16-node DSM system and the format of a 40-bit physical memory address in a 256-node DSM system. As shown in FIG. 10, the four most significant bits comprise an LNID in the 16-node DSM system and the eight most significant bits comprise an LNID in the 256-node DSM system.

In particular embodiments of the DSM system, the physical address space for a virtual server is arranged so that the local node's memory always starts at address 0 (zero). One reason for using this arrangement is compatibility with legacy system software, in particular embodiments. Specifically, with local memory starting at address 0, system software (e.g., boot code) accesses local memory the same way that it does on a standard server. Another reason for using this arrangement is that it simplifies the address lookup in the CMM. For a memory read/write request from a local processor, an address in the lower 1/16th or 1/256th segment of the 40-bit address space is always local and all other addresses map to memory in other nodes.

To see how the arrangement works, consider the example of a virtual server consisting of three nodes: 0, 1, and 2. In a 16-node DSM system, the total addressable memory space for this virtual server would be 1 terabyte (2^40) and each node would be allocated a segment which is 1/16 of that space (64GB or 2^36). From a global view, the first 64GB segment of the physical address space starting at address 0 would be allocated to node 0 (i.e., the node whose LNID equals 0), the next 64GB segment to node 1, and the following segment to node 2. The remaining 13 segments would be unused since LNIDs 4-15 are not used.

FIG. 11 shows this physical address space from the local view of each of the three nodes in the virtual server. The local view of node 0 would be the same as the global view and is shown in FIG. 11 under the label “Node 0”, with Local Memory (0) first. Node 1 Memory second, and Node 2 Memory third. The local view of node 1 would be as shown under the label “Node 1”, with Local Memory (1) first, Node 0 Memory second, and Node 2 Memory third. And the local view of node 2 would be as shown under the label “Node 2”, with Local Memory (2) first, Node 1 Memory second, and Node 0 Memory third.

It will be appreciated that in order to accomplish this arrangement, the locations of the local segment and the node 0 segment are swapped in the address map. And since MY_LNID, as defined above, is the LNID assigned to the local node, this is equivalent to swapping MY_LNID with LNID 0 in the address map. However, such a swapping would create confusion in the DSM system if it were applied to memory traffic leaving the node over the switched fabric. Therefore, the node's CMM reverses the swapping for traffic leaving the node.

FIG. 12 is a diagram showing a flowchart of an example process for altering a physical memory address, by the swapping a described above, prior to transmission over a HyperTransport bus. In the process's first step 1201, a node's CMM receives a memory operation (e.g., a read, write, or probe) pertaining to a physical memory address from the RDM on the DSM-management chip. In step 1202, the CMM determines whether the four (or eight) most significant bits in the physical address are equal to: (1) the MY_LNID value for the node; or (2) zero. If so, the CMM goes to step 1203, where: (1) if those bits are equal to the MY_LNID value, the CMM sets the bits to zero (e g., by changing to zero the four (or eight) most significant bits in the physical memory address) before transmission of the operation over the HyperTransport bus; and (2) if those bits are equal to zero, the CMM sets those bits to MY_LNID (e.g., by changing to MY_LNID the four (or eight) most significant bits in the physical memory address) before transmission of the operation over the HyperTransport bus. Otherwise, if those bits are not equal to MY_LNID or zero, the CMM goes to step 1204 and allows the memory operation to proceed without processing relating to LNID swapping.

FIG. 13 is a diagram showing a flowchart of an example process for altering a physical memory address, by reversing the swapping as described above, prior to transmission over a switched fabric. In the process's first step 1301, a node's CMM receives a memory operation (e.g., a read, write, or probe) pertaining to a physical memory address from one of the node's CPUs over the HyperTransport (e.g., ccHT) bus that connects the node's CPUs to the node's DSM-management chip. In step 1302, the CMM determines whether the four (or eight) most significant bits in the physical address are equal to (1) the MY_LNID value for the node; or (2) zero. If so, the CMM goes to step 1303, where: (1) if those bits are equal to the MY_LNID value, the CMM sets the DstLNID value to zero (e g., by changing to zero the four (or eight) most significant bits in the physical memory address) before transmission of the operation to the RDM; and (2) if those bits are equal to zero, the CMM sets the DstLNID value to MY_LNID (e.g. by changing to MY_LNID the four (or eight) most significant bits in the physical memory address) before transmission of the operation to the RDM. Otherwise, if those bits are not equal to MY_LNID or zero, the CMM goes to step 1304 and allows the memory operation to proceed without processing relating to LNID swapping, if the physical memory address is not for exported local memory. (If the physical memory address is for exported local memory, a probe operation to another physical memory address might result, feeding back into the process at step 1301.)

Particular embodiments of the above-described processes might be comprised of instructions that are stored on storage media. The instructions might be retrieved and executed by a processing system. The instructions are operational when executed by the processing system to direct the processing system to operate in accord with the present invention. Some examples of instructions are software, program code, firmware, and microcode. Some examples of storage media are memory devices, tape, disks, integrated circuits, and servers. The term “processing system” refers to a single processing device or a group of inter-operational processing devices. Some examples of processing devices are integrated circuits and logic circuitry. Those skilled in the art are familiar with instructions, storage media, and processing systems.

Those skilled in the art will appreciate variations of the above-described embodiments that fall within the scope of the invention. In this regard, it will be appreciated that there are many other possible orderings of the steps in the processes described above and many other possible modularizations of those orderings. Also, it will be appreciated that the above processes relating to memory-addressing will work with physical memory addresses that exceed 40-bits in width and DSM systems that have more than 256 nodes. Further, it will be appreciated that the DSM system will work with nodes whose CPUs are not Opterons having a ccHT bus. As a result, the invention is not limited to the specific examples and illustrations discussed above, but only by the following claims and their equivalents.

Akkawi, Isam, Krakirian, Shahe Hagop

Patent Priority Assignee Title
10021806, Oct 28 2011 III Holdings 2, LLC System and method for flexible storage and networking provisioning in large scalable processor installations
10050970, Oct 30 2009 III Holdings 2, LLC System and method for data center security enhancements leveraging server SOCs or server fabrics
10135731, Oct 30 2009 III Holdings 2, LLC Remote memory access functionality in a cluster of data processing nodes
10140245, Oct 30 2009 III Holdings 2, LLC Memcached server functionality in a cluster of data processing nodes
10187452, Aug 23 2012 Hewlett Packard Enterprise Development LP Hierarchical dynamic scheduling
10205772, Aug 23 2012 Hewlett Packard Enterprise Development LP Saving and resuming continuation on a physical processor after virtual processor stalls
10353736, Aug 29 2016 Hewlett Packard Enterprise Development LP Associating working sets and threads
10579274, Jun 27 2017 Hewlett Packard Enterprise Development LP Hierarchical stalling strategies for handling stalling events in a virtualized environment
10579421, Aug 29 2016 Hewlett Packard Enterprise Development LP Dynamic scheduling of virtual processors in a distributed system
10620992, Aug 29 2016 Hewlett Packard Enterprise Development LP Resource migration negotiation
10623479, Aug 23 2012 Hewlett Packard Enterprise Development LP Selective migration of resources or remapping of virtual processors to provide access to resources
10645150, Aug 23 2012 Hewlett Packard Enterprise Development LP Hierarchical dynamic scheduling
10783000, Aug 29 2016 Hewlett Packard Enterprise Development LP Associating working sets and threads
10817347, Aug 31 2017 Hewlett Packard Enterprise Development LP Entanglement of pages and guest threads
10877695, Oct 30 2009 III Holdings 2, LLC Memcached server functionality in a cluster of data processing nodes
11023135, Jun 27 2017 Hewlett Packard Enterprise Development LP Handling frequently accessed pages
11159605, Aug 23 2012 Hewlett Packard Enterprise Development LP Hierarchical dynamic scheduling
11175927, Nov 14 2017 Hewlett Packard Enterprise Development LP Fast boot
11240334, Oct 01 2015 Hewlett Packard Enterprise Development LP Network attached memory using selective resource migration
11403135, Aug 29 2016 Hewlett Packard Enterprise Development LP Resource migration negotiation
11449233, Jun 27 2017 Hewlett Packard Enterprise Development LP Hierarchical stalling strategies for handling stalling events in a virtualized environment
11467883, Mar 13 2004 III Holdings 12, LLC Co-allocating a reservation spanning different compute resources types
11494235, Nov 08 2004 III Holdings 12, LLC System and method of providing system jobs within a compute environment
11496415, Apr 07 2005 III Holdings 12, LLC On-demand access to compute resources
11513836, Aug 29 2016 Hewlett Packard Enterprise Development LP Scheduling resuming of ready to run virtual processors in a distributed system
11522811, Apr 07 2005 III Holdings 12, LLC On-demand access to compute resources
11522952, Sep 24 2007 The Research Foundation for The State University of New York Automatic clustering for self-organizing grids
11526304, Oct 30 2009 III Holdings 2, LLC Memcached server functionality in a cluster of data processing nodes
11533274, Apr 07 2005 III Holdings 12, LLC On-demand access to compute resources
11537434, Nov 08 2004 III Holdings 12, LLC System and method of providing system jobs within a compute environment
11537435, Nov 08 2004 III Holdings 12, LLC System and method of providing system jobs within a compute environment
11630704, Aug 20 2004 III Holdings 12, LLC System and method for a workload management and scheduling module to manage access to a compute environment according to local and non-local user identity information
11650857, Mar 16 2006 III Holdings 12, LLC System and method for managing a hybrid computer environment
11652706, Jun 18 2004 III Holdings 12, LLC System and method for providing dynamic provisioning within a compute environment
11656878, Nov 14 2017 Hewlett Packard Enterprise Development LP Fast boot
11656907, Nov 08 2004 III Holdings 12, LLC System and method of providing system jobs within a compute environment
11658916, Mar 16 2005 III Holdings 12, LLC Simple integration of an on-demand compute environment
11709709, Nov 08 2004 III Holdings 12, LLC System and method of providing system jobs within a compute environment
11720290, Oct 30 2009 III Holdings 2, LLC Memcached server functionality in a cluster of data processing nodes
11762694, Nov 08 2004 III Holdings 12, LLC System and method of providing system jobs within a compute environment
11765101, Apr 07 2005 III Holdings 12, LLC On-demand access to compute resources
11803306, Jun 27 2017 Hewlett Packard Enterprise Development LP Handling frequently accessed pages
11831564, Apr 07 2005 III Holdings 12, LLC On-demand access to compute resources
11861404, Nov 08 2004 III Holdings 12, LLC System and method of providing system jobs within a compute environment
11886915, Nov 08 2004 III Holdings 12, LLC System and method of providing system jobs within a compute environment
11907768, Aug 31 2017 Hewlett Packard Enterprise Development LP Entanglement of pages and guest threads
9054990, Oct 30 2009 Silicon Valley Bank System and method for data center security enhancements leveraging server SOCs or server fabrics
9069929, Oct 31 2011 Silicon Valley Bank Arbitrating usage of serial port in node card of scalable and modular servers
9077654, Oct 30 2009 Silicon Valley Bank System and method for data center security enhancements leveraging managed server SOCs
9092594, Oct 31 2011 Silicon Valley Bank Node card management in a modular and large scalable server system
9465771, Sep 24 2009 Silicon Valley Bank Server on a chip and node cards comprising one or more of same
9479463, Oct 30 2009 III Holdings 2, LLC System and method for data center security enhancements leveraging managed server SOCs
9509552, Oct 30 2009 III Holdings 2, LLC System and method for data center security enhancements leveraging server SOCs or server fabrics
9585281, Oct 28 2011 Silicon Valley Bank System and method for flexible storage and networking provisioning in large scalable processor installations
9648102, Dec 27 2012 Silicon Valley Bank Memcached server functionality in a cluster of data processing nodes
9680770, Oct 30 2009 Silicon Valley Bank System and method for using a multi-protocol fabric module across a distributed server interconnect fabric
9749326, Oct 30 2009 III Holdings 2, LLC System and method for data center security enhancements leveraging server SOCs or server fabrics
9792249, Oct 31 2011 III Holdings 2, LLC Node card utilizing a same connector to communicate pluralities of signals
9866477, Oct 30 2009 III Holdings 2, LLC System and method for high-performance, low-power data center interconnect fabric
9876735, Oct 30 2009 Silicon Valley Bank Performance and power optimized computer system architectures and methods leveraging power optimized tree fabric interconnect
9929976, Oct 30 2009 III Holdings 2, LLC System and method for data center security enhancements leveraging managed server SOCs
9965442, Oct 31 2011 III Holdings 2, LLC Node card management in a modular and large scalable server system
Patent Priority Assignee Title
5774731, Mar 22 1995 Hitachi, Ltd.; Hitachi ULSI Engineering Co., Ltd. Exclusive control method with each node controlling issue of an exclusive use request to a shared resource, a computer system therefor and a computer system with a circuit for detecting writing of an event flag into a shared main storage
6160814, May 31 1997 TEXAS INSRRUMENTS INCORPORATED Distributed shared-memory packet switch
6757790, Feb 19 2002 EMC IP HOLDING COMPANY LLC Distributed, scalable data storage facility with cache memory
6877030, Feb 28 2002 Hewlett Packard Enterprise Development LP Method and system for cache coherence in DSM multiprocessor system without growth of the sharing vector
6922766, Sep 04 2002 Cray Inc. Remote translation mechanism for a multi-node system
20010037435,
20030076831,
20040030763,
20040148472,
////
Executed onAssignorAssigneeConveyanceFrameReelDoc
Oct 14 2011THE FLORIDA STATE UNIVERSITY FOUNDATION, INCORPORATEDIntellectual Ventures Holding 80 LLCASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0283470991 pdf
May 10 2012Intellectual Ventures Holding 80 LLC(assignment on the face of the patent)
Aug 27 2015Intellectual Ventures Holding 80 LLCINTELLECTUAL VENTURES FUND 81 LLCMERGER SEE DOCUMENT FOR DETAILS 0375750812 pdf
Aug 27 2015Intellectual Ventures Holding 80 LLCIntellectual Ventures Holding 81 LLCCORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE NAME PREVIOUSLY RECORDED AT REEL: 037575 FRAME: 0812 ASSIGNOR S HEREBY CONFIRMS THE MERGER 0385160869 pdf
Date Maintenance Fee Events
Oct 16 2017M1552: Payment of Maintenance Fee, 8th Year, Large Entity.
Oct 14 2021M1553: Payment of Maintenance Fee, 12th Year, Large Entity.


Date Maintenance Schedule
Nov 26 20164 years fee payment window open
May 26 20176 months grace period start (w surcharge)
Nov 26 2017patent expiry (for year 4)
Nov 26 20192 years to revive unintentionally abandoned end. (for year 4)
Nov 26 20208 years fee payment window open
May 26 20216 months grace period start (w surcharge)
Nov 26 2021patent expiry (for year 8)
Nov 26 20232 years to revive unintentionally abandoned end. (for year 8)
Nov 26 202412 years fee payment window open
May 26 20256 months grace period start (w surcharge)
Nov 26 2025patent expiry (for year 12)
Nov 26 20272 years to revive unintentionally abandoned end. (for year 12)