Embodiments of the invention may improve the performance of multi-processor systems in processing information received via a network. For example, some embodiments may enable configuration of a system such that information received can be distributed among multiple processors for efficient processing. A user may select from among multiple configuration options, each configuration option being associated with a particular mode of processing information received. By selecting a configuration option, the user may specify how received information is processed to capitalize on the system's characteristics, such as by aligning processors on the system with certain NICs. As such, the processor(s) aligned with a NIC may perform networking-related tasks associated with information received by that NIC. If initial alignment causes one or more processors to become over-burdened, processing tasks may be dynamically re-distributed to other processors.
|
1. A computer, coupled to at least one computer network, comprising:
a plurality of processors, comprising a plurality of non-uniform memory access nodes;
at least one network interface controller operable to receive packets of information transmitted via the at least one computer network over at least one network connection, each packet being associated with a particular network connection, and to provide packets associated with each network connection to a particular one of the plurality of non-uniform memory access nodes; and
at least one storage facility storing first programmed instructions defining a query facility usable to identify a non-uniform memory access node to which packets associated with a particular network connection are provided by the at least one network interface controller;
wherein an application accesses the query facility to determine the non-uniform memory access node on which networking-related processing for the application is performed, and wherein the application executes on the determined non-uniform memory access node.
8. A computer implemented process, performed by a computer, coupled to at least one computer network, the computer including a plurality of processors including a plurality of non-uniform memory access nodes, and at least one network interface controller, and at least one storage facility storing first programmed instructions defining a query facility, and an application executing on the computer, the computer-implemented process comprising:
the at least one network interface controller receiving packets of information transmitted via the at least one computer network over at least one network connection, each packet being associated with a particular network connection;
the at least one network controller providing packets associated with each network connection to a particular one of the plurality of non-uniform memory access nodes; and
the application accessing the query facility to determine a non-uniform memory access node on which networking-related processing for the application is performed;
the query facility identifying to the application a non-uniform memory access node to which packets associated with a particular network connection for the application are provided by the at least one network interface controller; and
the application executing on the determined non-uniform memory access node to process packets received over the network connection for the application.
15. An article of manufacture, comprising:
a computer storage device or a computer memory device;
computer program instructions stored on the computer storage device or computer memory device that, when executed by a computer, coupled to at least one computer network, the computer including a plurality of processors including a plurality of non-uniform memory access nodes, and at least one network interface controller, and at least one storage facility storing first programmed instructions defining a query facility, and an application executing on the computer, the computer performs a process comprising:
the at least one network interface controller receiving packets of information transmitted via the at least one computer network over at least one network connection, each packet being associated with a particular network connection;
the at least one network controller providing packets associated with each network connection to a particular one of the plurality of non-uniform memory access nodes; and
the application accessing the query facility to determine a non-uniform memory access node on which networking-related processing for the application is performed;
the query facility identifying to the application a non-uniform memory access node to which packets associated with a particular network connection for the application are provided by the at least one network interface controller; and
the application executing on the determined non-uniform memory access node to process packets received over the network connection for the application.
2. The computer of
3. The computer of
4. The computer of
5. The computer of
6. The computer of
7. The computer of
9. The computer-implemented process of
10. The computer-implemented process of
11. The computer-implemented process of
12. The computer-implemented process of
13. The computer-implemented process of
14. The computer-implemented process of
16. The article of manufacture of
17. The article of manufacture of
18. The article of manufacture of
19. The article of manufacture of
20. The article of manufacture of
|
This application claims priority to, and the benefits of, under 35 U.S.C. § 120, and is a continuation of, U.S. patent application Ser. No. 14/696,474, entitled “IMPROVING PERFORMANCE OF MULTIPROCESSOR COMPUTER SYSTEMS”, filed Apr. 26, 2015, pending, which is a continuation of U.S. patent application Ser. No. 13/219,373, entitled “IMPROVING PERFORMANCE OF MULTIPROCESSOR COMPUTER SYSTEMS”, filed Aug. 26, 2011, issued as U.S. Pat. No. 9,021,138 on Apr. 28, 2015, which are hereby incorporated by reference.
Some computer systems employ multiple processors. These processors may, for example, be used to process information received from a remote computer via a network. Typically, the information is received at the computer by one or more network interface controllers (NICs) that receive information transmitted in accordance with a prescribed network communication protocol (e.g., TCP, or Transmission Control Protocol). The received information may, for example, include requests to perform networking-related operations that may be distributed across the multiple processors for execution. Allowing multiple processors to simultaneously perform networking-related operations can improve the computer's performance. One known architecture which employs multiple processors to perform networking-related operations is a “symmetric multiprocessor” (SMP) system architecture.
In a conventional technique for processing information received via a network on an SMP system, a NIC on the computer receives a data packet and stores it in memory with a packet descriptor that includes pointer information identifying the location of the data in memory. If a previous packet is not currently being processed, an interrupt is generated to one of the processors, and an interrupt service routine (ISR) executes, suspending further interrupts from the NIC while a deferred procedure call (DPC) is requested to run on a processor selected to handle the packet. As the DPC executes, one or more data packets and descriptors are retrieved from memory to build an array, and the processor then initiates processing the packet. After the DPC completes, further interrupts from the NIC are re-enabled, so that additional packets may be distributed to one or more other of the processors.
In some conventional multi-processor architectures, one or more processors may have associated memory. For example, in a Non-Uniform Memory Access (NUMA) system, one or more processors may comprise a NUMA node having an associated set of memory addresses that are accessed most efficiently by processors in the NUMA node. In this respect, a particular NUMA node on a system may be thought of as being “closest” to the set of memory addresses associated with the node, and “further away” from other sets of memory addresses that are each associated with other NUMA nodes on the system. In a NUMA-enabled system, access by a first NUMA node to a memory address associated with a second NUMA node is slower than access by the second NUMA node would be.
Some embodiments of the present invention provide methods and apparatus for improving the performance and/or scalability of multi-processor computer systems, such as those which employ NUMA and/or other system architectures. For example, some embodiments of the invention provide tools and/or techniques for configuring a system so that processing tasks associated with information (e.g., packets and/or other traffic) received via a network may be distributed among multiple processors on the system, so that the tasks may be performed efficiently. For example, in some embodiments of the invention, multiple configuration options (e.g., implemented as a set of “profiles”) may be presented to a user (e.g., system administrator) for selection, with each configuration option being associated with a particular mode of processing information received via a network. In this respect, the inventors have recognized that there may not be one single configuration scheme that is suitable and optimal for every type of system. As such, some embodiments of the invention provide multiple configuration options, so that users may select one or more configuration options that capitalize on a particular system's characteristics and/or capabilities.
In one example configuration option described herein, specific processors may be aligned with certain NICs on the system, so that the processor(s) aligned with a particular NIC perform(s) networking-related operations associated with information received via a network by the NIC. A processor may be aligned with a NIC in any of numerous ways. For example, in one technique described below, a processor may be aligned with a NIC by assigning execution of ISRs and DPCs generated by the NIC to the processor. Of course, any suitable criteria may define alignment of a processor to a NIC, as embodiments of the invention are not limited to any particular manner of implementation.
An example configuration option may enable a user to specify that if an initial alignment of processors to NICs causes one or more processors to be over-burdened with a disproportionate amount of networking-related tasks while other processors are under-utilized, tasks may be dynamically re-distributed to other processors on the system to achieve a more even distribution of the overall processing burden. As a result, a system may perform networking-related tasks more efficiently.
In some embodiments of the invention, configuration of a multi-processor system may account for the characteristics of an architecture employed by the system. For example, embodiments of the invention used with a NUMA system may cause processing tasks to be distributed across NUMA nodes on the system. If a subsequent redistribution of processing tasks is desired, redistribution may be performed in a manner which does not result in a processor assigned to one NUMA node attempting to access a memory address associated with another NUMA node, which can detract significantly from system performance.
Some embodiments of the invention may provide other features designed to benefit NUMA systems. For example, some embodiments of the invention may enable applications executing on NUMA systems to determine the NUMA node on which networking-related processing for the application is performed, so that the application may execute on the same NUMA node. As a result, attempts to access memory addresses associated with another NUMA node may be prevented.
The foregoing is a non-limiting summary of the invention, which is defined by the attached claims.
The accompanying drawings are not intended to be drawn to scale. In the drawings, each identical or nearly identical component that is illustrated in the various figures is represented by a like numeral. For purposes of clarity, not every component may be labeled in every drawing. In the drawings:
Some embodiments of the present invention may improve the performance of multi-processor computer systems in processing information received via a network. For example, some embodiments of the invention may enable configuration of a system such that information received via a network may be distributed among multiple processors for efficient processing. A user (e.g., system administrator) may select from among multiple configuration options, each of which may generally be considered a “policy,” or profile, defining a mode of processing information received via a network. By selecting one or more configuration options, the user may specify how information received via the network is processed to capitalize on the system's characteristics. For example, in some embodiments, a multi-processor system may be configured such that specific processors on the system are aligned with certain NICs, so that each processor aligned with a NIC performs networking-related operations associated with information received by the NIC. For example, in some embodiments, a processor aligned with a NIC may process ISRs and DPCs (or any other suitable operation type(s)) generated by the NIC, and not operations generated by any other NIC on the system.
In some embodiments of the invention, if an initial alignment of processors to NICs causes one or more processors to become over-burdened while other processors are under-utilized, processing tasks may be dynamically re-distributed so as to achieve a more even distribution of the overall processing burden across processors on the system. As a result, the system may efficiently utilize the processing resources at its disposal to perform networking-related operations.
Some embodiments of the invention provide specific benefits for systems which employ NUMA architectures. For example, some embodiments may enable networking-related tasks to be distributed across processors so that multiple NUMA nodes are utilized. If a redistribution of processing tasks is subsequently performed, redistribution may be performed in a manner which does not result in a processor assigned to one NUMA node attempting to access a memory address associated with another NUMA node, as described below.
An example multi-processor system 200 on which embodiments of the invention may be implemented is depicted conceptually in
An example process 100 which may be performed to configure system 200 to perform networking-related tasks is shown in
At the start of process 100, one or more processors are aligned with one or more NICs in the system in act 105.
In some embodiments of the invention, different configuration options may specify different criteria for aligning processors with NICs. For example, in some embodiments, a NIC having greater bandwidth may be aligned with a proportionally greater number of processors than other NICs having lesser bandwidth, to enable a system to capitalize on a high-speed connection. In this respect,
Some embodiments may limit a number of processors aligned with a particular NIC to accommodate the NIC's ability to handle a predetermined quantity of processors. For example, if NIC 225 were only capable of supporting four processors, then only four (or fewer) processors might be aligned with NIC 225, and the remaining nine processors may be distributed among NICs 215 and 220. Any suitable manner of aligning processors with NICs, recognizing any suitable constraint(s), may be employed, as embodiments of the invention are not limited in this respect.
Certain implementations may include processors that are grouped into NUMA nodes. As such, some embodiments of the invention provide configuration options which enable processors to be aligned with NICs such that the available NUMA nodes on the system are fully utilized for networking-related tasks. For example, some embodiments may enable alignment of processors in each NUMA node on the system to a particular NUMA-enabled NIC, thereby enabling traffic to be distributed across NUMA nodes on the system. An example is depicted in
In some embodiments, processors are assigned to NUMA-enabled NICs prior to being assigned to non-NUMA-enabled NICs, to maximize the probability that each NUMA-enabled NIC is aligned with at least one processor from each NUMA node, thereby enabling traffic to be distributed across the NUMA nodes on the system. For example, processors may be assigned to NUMA-enabled NICs until each NIC is assigned the maximum number of processors it can support. In these embodiments, processors may then be assigned to non-NUMA enabled NICs. Thus, in the example shown in
As a result of the example configuration shown in
Referring again to
An example process 600 for redistributing processing tasks upon a determination that one or more processors is over-burdened is depicted in
At the start of process 600, a weighted average utilization is computed for a particular processor in act 605. For example, act 605 may include a processor calculating a weighted average utilization, based on busy and idle cycle counts, using techniques which are known in the art for measuring processor load. Any suitable technique may be employed to calculate weighted average utilization, as embodiments of the invention are not limited in this respect.
Process 600 then proceeds to act 610, wherein a weighted activity level is computed for the particular processor. This also may be performed in any suitable fashion. For example, in some embodiments, the weighted activity level is based on a number of Internet Protocol (IP) datagrams handled by the processor during a given interval, including IP datagrams corresponding to locally-destined segments and datagrams destined for remote devices. Of course, embodiments of the invention are not so limited, as a weighted average activity level may be calculated using any suitable technique(s).
At the completion of act 610, process 600 proceeds to act 615, wherein a determination is made whether the processor is a “processor to free.” In this respect, a processor may be designated as one that is to be freed of its processing responsibilities for any of various reasons. For example, for power conservation reasons, a processor may be designated as one which is to be freed of its responsibilities so that power is not consumed to keep it running.
If it is determined in act 615 that the particular processor is not a processor to free, then process 600 proceeds to act 620, wherein it is determined whether the processor is busy. This determination may, for example, be based on an analysis of information gathered in acts 605 and/or 610, and/or other information. Analysis may be performed in any suitable manner, such as by determining whether information collected in acts 605 and/or 610 indicates that the processor's weighted average utilization and/or weighted activity level exceed(s) one or more predetermined thresholds. For example, in some embodiments, a processor may be deemed “busy” after its weighted average utilization exceeds ninety percent, until its weighted average utilization is less than seventy percent. Of course, any suitable measurement(s) may be used to determine whether a processor is busy.
In the example process 600, if it is determined in act 620 that the processor is not busy, process 600 completes. As a result, the processor executing example process 600 is not identified as one from which processing tasks should be shifted.
If, however, it is determined in act 620 that the processor is busy, then example process 600 proceeds to act 625, wherein it is determined whether the processor is currently handling more than one active partition. A partition may, for example, represent one or more connections over which information is received at the processor. If it is determined that the processor is not currently handling more than one active partition, then process 600 completes. If, however, it is determined that the processor is handling more than one active partition, or if it is determined in act 615 that the processor is a processor to free, then process 600 proceeds to act 630, wherein a determination is made whether there are eligible “target” processors to which tasks currently handled by the processor in question may be shifted.
An example process 700 for determining whether there are any eligible target processors is depicted in
At the start of example process 700, a determination is made in act 705 whether a processor being examined is assigned to the same adapter as the processor that executes example process 700, indicating that the processor being examined is aligned with the same NIC as the processor which executes example process 700. If not, it is determined that the processor being examined is ineligible as a target processor to which to shift networking-related tasks, and example process 700 completes.
If, however, the processor being examined is assigned to the same adapter, then process 700 proceeds to act 710, wherein a determination is made whether a user has invoked a NUMA profile. A NUMA profile having been invoked may indicate that the system employs a NUMA architecture, such that particular processors may be grouped into one or more NUMA nodes. If it is determined that a user has invoked a NUMA profile, then example process 700 proceeds to act 715, wherein a determination is made whether the processor being examined resides in the same NUMA node as the processor that performs example process 700. In this respect, the inventors have recognized the undesirability of shifting processing tasks from a processor in one NUMA node to a processor in another NUMA node, since doing so could cause the processor to which the task is shifted to access the memory of the other NUMA node to complete an networking-related task, which can detract from system performance. As a result, some embodiments of the invention keep processing tasks from being shifted from a processor in one NUMA node to a processor in another NUMA node. In this respect, if it is determined in act 715 that the processor being examined does not reside in the same NUMA node as the processor that performs example process 700, then the processor being examined is determined to be ineligible as a target processor, and example process 700 completes.
If it is determined in act 715 that the processor being examined is in the same NUMA node, or if it is determined in act 710 that the user has not invoked a NUMA profile, then example process 700 proceeds to act 720, wherein a determination is made whether the processor being examined is busy. This determination may, for example, be based on an analysis of information, collected for the processor being examined, similar to that which is collected in acts 605 and 610 (
If it is determined that the processor being examined is busy, then the processor is determined to be ineligible, and example process 700 completes. If it is determined that the processor is not too busy, then it is determined to be an eligible processor. In some embodiments, the processor may, for example, be added to a list or other collection of target processors to which networking-related tasks may be shifted. Example process 700 then completes.
Returning to example process 600 (
At the start of process 800, a determination is made in act 805 whether a NIC being examined is currently aligned with its maximum number of processors. For example, each NIC on the system may have a maximum number of processors which it can support, and/or a configuration option or policy may specify that a particular NIC be aligned with only a certain number of processors. A maximum number of processors may be specified in any suitable manner.
If it is determined in act 805 that the NIC being examined is currently aligned with a maximum number of processors, then process 800 proceeds to act 810, wherein a determination is made whether a NUMA profile is invoked, indicating that the system employs a NUMA architecture. A NUMA profile may be invoked, for example, by a user or other entity. If it is determined that a NUMA profile has not been invoked, then example process 800 completes with a conclusion that there are no processors to acquire to which to shift networking-related tasks.
If, however, it is determined in act 810 that a NUMA profile is invoked, then process 800 proceeds to act 815, wherein a determination is made whether there is an idle processor on a NUMA node, different than that to which the NIC is presently aligned, with more than one processor assigned. In some embodiments, an idle processor may be one with a weighted average utilization of less than twenty percent, although any suitable measurement of processor utilization and/or activity may be employed. If it is determined that there is an idle processor on a NUMA node, then example process 800 completes with a conclusion that there are no processors to acquire.
If, however, it is determined in act 815 that there is at least one idle processor on another NUMA node, then process 800 proceeds to act 820, wherein the least active processor on the other NUMA node is released for use. A “least active” processor may be determined using any suitable technique, as embodiments of the invention are not limited in this respect. For example, the least active processor may be identified using information similar to that which is collected in act 605 and 610 for each processor on the other NUMA node.
At the completion of act 820, example process 800 proceeds to act 830, wherein a determination is made whether there is an idle processor on the same NUMA node as that to which the NIC is presently aligned. If not, it is determined that there are no processors to acquire, and example process 800 completes.
If it is determined in act 830 that there is at least one idle processor on the same NUMA node, then example process 800 proceeds to act 835, wherein the least active processor on the same NUMA node is acquired for use, and process 800 then completes. As noted above with respect to act 815, the least active processor may be identified using any suitable criteria and/or technique, as embodiments of the invention are not limited in this respect.
Returning to act 805, if it is determined in that the NIC is not presently using its maximum processor count, then process 800 proceeds to act 825, wherein it is determined whether the user has invoked a NUMA profile. If so, example process 800 proceeds to act 830, and proceeds as described above. If not, process 800 proceeds to act 840, wherein a determination is made whether an idle processor is available. If it is determined that an idle processor is available, then the least active processor (identified using any suitable technique) is acquired, and process 800 then completes. If it is determined in act 840 that no idle processor is available, then no processor is acquired, and example process 800 completes.
Returning again to act 635 (
At the completion of either of acts 640 or 645, process 600 proceeds to act 650, wherein a highest average activity level for all processors on the system is determined. This may, for example, reflect a recognition that shifting processing tasks from one processor to another, only to make the receiving processor the busiest processor on the system, is undesirable. Consequently, in some embodiments of the invention, the most heavily utilized processor on the system is first identified, and processing tasks are shifted only if the activity level on the receiving processor will not exceed that of the busiest processor.
Process 600 then proceeds to act 655, wherein a determination is made whether any partition assigned to the processor being examined is movable. An example process 900 for making this determination is shown in
At the start of example process 900, a determination is made in act 905 whether the natural processor for a particular partition belongs to the same adapter as that to which the processor is assigned. In this respect, a partition's natural processor may, for example, handle timer processing for the partition. If it is determined in act 905 that the partition's natural processor belongs to the current adapter, then example process 900 proceeds to act 910, wherein a determination is made in act 910 whether the partition can move to the natural processor. This determination may involve evaluating any suitable criteria, including (for example) the present level of activity on the natural processor.
If it is determined in act 910 that the partition can be moved to the natural processor, then example process 900 proceeds to act 915, wherein it is determined whether the natural processor is the processor to free (as described above in relation to act 615,
If it is determined in act 915 that the natural processor is the processor to free, then the partition is assigned a weight represented by a 2-tuple comprising a four and the partition's activity level.
If it is determined in act 910 that the partition cannot move to the natural processor, or if it is determined in act 905 that the partition's natural processor is not assigned to the current adapter, then example process 900 proceeds to act 920, wherein it is determined whether the current processor is the processor to free. If it is determined in act 920 that the current processor is the processor to free, then example process 900 proceeds to act 930, wherein it is determined whether the partition can move to the least active partition that isn't a processor to free. If so, the partition is assigned a weight represented by a 2-tuple comprising the number three and the partition's activity level, and example process 900 completes.
If it is determined in act 930 that the partition cannot move to the processor to which the least active partition that is not a processor to free has been assigned, then example process 900 proceeds to act 935, wherein it is determined whether the partition can move to the least active processor. If so, the partition is assigned a weight represented by a 2-tuple comprising the number two and the partition's activity level. If not, it is determined that the current partition cannot be moved, and example process 900 completes.
If it is determined in act 920 that the current processor is not the processor to free, then example process 900 proceeds to act 925, wherein it is determined whether the partition being examined is the most current on the current adapter. If not, example process 900 proceeds to act 930, and continues as described above. If it is determined that the partition being examined is the most active on the current adapter, then it is determined that the current partition cannot be moved, and example process 900 completes.
Referring again to process 600 in
At the completion of act 660, or if it is determined in act 655 that there are no movable partitions, then example process 600 completes.
Another configuration option may enable an application on a NUMA system to identify the NUMA node(s) on which networking-related tasks are being performed, so that application-related processing may be performed on the same node(s). In this respect, as noted above, a process running on one NUMA node which attempts access to the memory of another NUMA node (e.g., to complete a networking-related task, to perform application processing, etc.) can significantly detract from overall system performance. Consequently, some embodiments of the invention enable an application to identify a NUMA node and processor that presently services a connection over which information being processed by the application is received, so that the application itself may execute on the same node, thereby avoiding attempts to access memory addresses associated with another NUMA node. Of course, it should be appreciated that embodiments of the invention are not limited to enabling an application to identify which NUMA node handles a connection. For example, some embodiments may enable an application to specify a NUMA node on which it executes (or previously executed), so that networking-related tasks may be moved to that NUMA node for processing. Embodiments of the invention may be implemented in any of numerous ways.
An example process 1000 for identifying a node handling a particular connection is shown in
At the completion of act 1005, an identification of the node is provided to the application in act 1010. An application may use this information to, for example, “wake up” on the same node, to prevent attempts to access memory of one or more other NUMA nodes. Example process 1000 then completes.
The invention is operational with numerous other general purpose or special purpose computing system environments or configurations. Examples of well known computing systems, environments, and/or configurations that may be suitable for use with the invention include, but are not limited to, personal computers, server computers, hand-held or laptop devices, multiprocessor systems, microprocessor-based systems, set top boxes, programmable consumer electronics, network PCs, minicomputers, mainframe computers, distributed computing environments that include any of the above systems or devices, and the like.
The computing environment may execute computer-executable instructions, such as program modules. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types. The invention may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in both local and remote computer storage media including memory storage devices.
With reference to
Computer 1210 typically includes a variety of computer readable media. Computer readable media can be any available media that can be accessed by computer 1210 and includes both volatile and nonvolatile media, removable and non-removable media. By way of example, and not limitation, computer readable media may comprise computer storage media and communication media. Computer storage media includes both volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can accessed by computer 1210. Communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of the any of the above should also be included within the scope of computer readable media.
The system memory 1230 includes computer storage media in the form of volatile and/or nonvolatile memory such as read only memory (ROM) 1231 and random access memory (RAM) 1232. A basic input/output system 1233 (BIOS), containing the basic routines that help to transfer information between elements within computer 1210, such as during start-up, is typically stored in ROM 1231. RAM 1232 typically contains data and/or program modules that are immediately accessible to and/or presently being operated on by processing unit 1220. By way of example, and not limitation,
The computer 1210 may also include other removable/non-removable, volatile/nonvolatile computer storage media. By way of example only,
The drives and their associated computer storage media discussed above and illustrated in
The computer 1210 may operate in a networked environment using logical connections to one or more remote computers, such as a remote computer 1280. The remote computer 1280 may be a personal computer, a server, a router, a network PC, a peer device or other common network node, and typically includes many or all of the elements described above relative to the computer 1210, although only a memory storage device 1281 has been illustrated in
When used in a LAN networking environment, the computer 1210 is connected to the LAN 1271 through a network interface or adapter 1270. When used in a WAN networking environment, the computer 1210 typically includes a modem 1272 or other means for establishing communications over the WAN 1273, such as the Internet. The modem 1272, which may be internal or external, may be connected to the system bus 1221 via the user input interface 1260, or other appropriate mechanism. In a networked environment, program modules depicted relative to the computer 1210, or portions thereof, may be stored in the remote memory storage device. By way of example, and not limitation,
Having thus described several aspects of at least one embodiment of this invention, it is to be appreciated that various alterations, modifications, and improvements will readily occur to those skilled in the art. Such alterations, modifications, and improvements are intended to be part of this disclosure, and are intended to be within the spirit and scope of the invention. Accordingly, the foregoing description and drawings are by way of example only.
The above-described embodiments of the present invention can be implemented in any of numerous ways. For example, the embodiments may be implemented using hardware, software or a combination thereof. When implemented in software, the software code can be executed on any suitable processor or collection of processors, whether provided in a single computer or distributed among multiple computers. Such processors may be implemented as integrated circuits, with one or more processors in an integrated circuit component. Though, a processor may be implemented using circuitry in any suitable format.
Further, it should be appreciated that a computer may be embodied in any of a number of forms, such as a rack-mounted computer, a desktop computer, a laptop computer, or a tablet computer. Additionally, a computer may be embedded in a device not generally regarded as a computer but with suitable processing capabilities, including a Personal Digital Assistant (PDA), a smart phone or any other suitable portable or fixed electronic device.
Also, a computer may have one or more input and output devices. These devices can be used, among other things, to present a user interface. Examples of output devices that can be used to provide a user interface include printers or display screens for visual presentation of output and speakers or other sound generating devices for audible presentation of output. Examples of input devices that can be used for a user interface include keyboards, and pointing devices, such as mice, touch pads, and digitizing tablets. As another example, a computer may receive input information through speech recognition or in other audible format.
Such computers may be interconnected by one or more networks in any suitable form, including as a local area network or a wide area network, such as an enterprise network or the Internet. Such networks may be based on any suitable technology and may operate according to any suitable protocol and may include wireless networks, wired networks or fiber optic networks.
Also, the various methods or processes outlined herein may be coded as software that is executable on one or more processors that employ any one of a variety of operating systems or platforms. Additionally, such software may be written using any of a number of suitable programming languages and/or programming or scripting tools, and also may be compiled as executable machine language code or intermediate code that is executed on a framework or virtual machine.
In this respect, the invention may be embodied as a computer readable storage medium (or multiple computer readable media) (e.g., a computer memory, one or more floppy discs, compact discs (CD), optical discs, digital video disks (DVD), magnetic tapes, flash memories, circuit configurations in Field Programmable Gate Arrays or other semiconductor devices, or other tangible computer storage medium and/or storage device) encoded with one or more programs that, when executed on one or more computers or other processors, perform methods that implement the various embodiments of the invention discussed above. As is apparent from the foregoing examples, a computer readable storage medium may retain information for a sufficient time so as to be considered a non-transitory storage medium. Such a computer readable storage medium or media can be transportable, such that the program or programs stored thereon can be loaded onto one or more different computers or other processors to implement various aspects of the present invention as discussed above. As used herein, the term “computer-readable storage medium” encompasses only a computer-readable medium that can be considered an article of manufacture (e.g., a portable storage device) or a machine. Alternatively or additionally, the invention may be embodied as a computer readable medium other than a computer-readable storage medium, such as a propagating signal.
The terms “program” or “software” are used herein in a generic sense to refer to any type of computer code or set of computer-executable instructions that can be employed to program a computer or other processor to implement various aspects of the present invention as discussed above. Additionally, it should be appreciated that according to one aspect of this embodiment, one or more computer programs that when executed perform methods of the present invention need not reside on a single computer or processor, but may be distributed in a modular fashion amongst a number of different computers or processors to implement various aspects of the present invention.
Computer-executable instructions may be in many forms, such as program modules, executed by one or more computers or other devices. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types. Typically the functionality of the program modules may be combined or distributed as desired in various embodiments.
Data structures may be stored in computer-readable media in any suitable form. For simplicity of illustration, data structures may be shown to have fields that are related through location in the data structure. Such relationships may likewise be achieved by assigning storage for the fields with locations in a computer-readable medium that conveys relationship between the fields. However, any suitable mechanism may be used to establish a relationship between information in fields of a data structure, including through the use of pointers, tags or other mechanisms that establish relationship between data elements.
Various aspects of the present invention may be used alone, in combination, or in a variety of arrangements not specifically discussed in the embodiments described in the foregoing, and is therefore not limited in its application to the details and arrangement of components set forth in the foregoing description or illustrated in the drawings. For example, aspects of one embodiment may be combined in any manner with aspects of other embodiments.
Also, the invention may be embodied as a method, an example of which is described with reference to the drawings. The acts performed as part of the method may be ordered in any suitable way. Accordingly, embodiments may be constructed in which acts are performed in an order different than illustrated, which may include performing some acts simultaneously, even though shown as sequential acts in illustrative embodiments.
Use of ordinal terms such as “first,” “second,” “third,” etc., in the claims to modify a claim element does not by itself connote any priority, precedence, or order of one claim element over another or the temporal order in which acts of a method are performed, but are used merely as labels to distinguish one claim element having a certain name from another element having a same name (but for use of the ordinal term) to distinguish the claim elements.
Also, the phraseology and terminology used herein is for the purpose of description and should not be regarded as limiting. The use of “including,” “comprising,” or “having,” “containing,” “involving,” and variations thereof herein, is meant to encompass the items listed thereafter and equivalents thereof as well as additional items.
Sridharan, Murari, Dabagh, Alireza, Srinivasan, Ganesh, Nievelt, Joseph, Shahid, Muhammad Junaid
Patent | Priority | Assignee | Title |
Patent | Priority | Assignee | Title |
6249802, | Sep 19 1997 | Hewlett Packard Enterprise Development LP | Method, system, and computer program product for allocating physical memory in a distributed shared memory network |
6769017, | Mar 13 2000 | Hewlett Packard Enterprise Development LP | Apparatus for and method of memory-affinity process scheduling in CC-NUMA systems |
7627700, | Jun 05 2007 | Texas Instruments Incorporated | Expanded memory for communications controller |
7743383, | Nov 01 2004 | VALTRUS INNOVATIONS LIMITED | Adaptive cooperative scheduling |
20060168254, | |||
20080178192, | |||
20090313445, | |||
20100153974, | |||
20110153724, | |||
CN101794270, | |||
CN101910971, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Jun 26 2011 | SHAHID, MUHAMMAD JUNAID | Microsoft Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050448 | /0652 | |
Jun 27 2011 | ALIREZA, DABAGH | Microsoft Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050448 | /0652 | |
Jun 27 2011 | NIEVELT, JOSEPH | Microsoft Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050448 | /0652 | |
Jun 28 2011 | SRINIVASAN, GANESH | Microsoft Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050448 | /0652 | |
Jul 23 2011 | SRIDHARAN, MURARI | Microsoft Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050448 | /0652 | |
Jul 02 2015 | Microsoft Corporation | Microsoft Technology Licensing, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050448 | /0719 | |
Feb 16 2017 | Microsoft Technology Licensing LLC | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Apr 21 2023 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Date | Maintenance Schedule |
Nov 19 2022 | 4 years fee payment window open |
May 19 2023 | 6 months grace period start (w surcharge) |
Nov 19 2023 | patent expiry (for year 4) |
Nov 19 2025 | 2 years to revive unintentionally abandoned end. (for year 4) |
Nov 19 2026 | 8 years fee payment window open |
May 19 2027 | 6 months grace period start (w surcharge) |
Nov 19 2027 | patent expiry (for year 8) |
Nov 19 2029 | 2 years to revive unintentionally abandoned end. (for year 8) |
Nov 19 2030 | 12 years fee payment window open |
May 19 2031 | 6 months grace period start (w surcharge) |
Nov 19 2031 | patent expiry (for year 12) |
Nov 19 2033 | 2 years to revive unintentionally abandoned end. (for year 12) |