Embodiments of the present invention pertain to dynamically resizing a virtual machine container. According to one embodiment, an optimal utilization is determined based on a desired performance for a multi-tiered application and transaction mix information that describes a mix of transactions that result from executing the multi-tiered application. Automatic modification of an amount of a physical resource a virtual machine container is entitled to for executing a particular application component associated with the multi-tiered application is enabled where the amount is automatically modified based on the optimal utilization.
|
1. A method of dynamically resizing a virtual machine container, the method comprising:
determining a possible size of the virtual machine container within a resource-on-demand environment in accordance with a service level agreement;
determining a corrected optimal utilization based on a desired performance for a multi-tiered application and transaction mix information that describes a mix of transactions that result from executing the multi-tiered application, the determining of the corrected optimal utilization further comprising:
determining a correction term based on a difference between the desired performance and a measured performance; and
determining the corrected optimal utilization based on the correction term and an uncorrected optimal utilization, wherein the uncorrected optimal utilization is based at least on service time of transactions, the mix of transactions, and fixed delays;
enabling automatic modification of an amount of a physical resource a virtual machine container is entitled to execute a particular application component associated with the multi-tiered application, wherein the amount of the physical resource is determined based at least in part on a ratio of physical resource consumption and physical resource entitlement, wherein the amount is automatically modified based on the corrected optimal utilization and the possible size of the virtual machine container; and
using the corrected optimal utilization for all of the application components associated with the multi-tiered application.
9. A non-transitory computer readable storage medium having computer-readable program code embedded therein for causing a computer system to perform a method of dynamically resizing a virtual machine container, the method comprising:
using a queueing model to determine an uncorrected optimal utilization based on a desired performance for a multi-tiered application, transaction mix information that describes a mix of transactions that result from executing a plurality of application components associated with at least the multi-tiered application, service time of transactions, and fixed delays;
determining a possible size of the virtual machine container within a resource-on-demand environment in accordance with a service level agreement;
determining a corrected optimal utilization based on a desired application-level performance, a measured application-level performance and the uncorrected optimal utilization;
enabling automatic modification of an amount of a physical resource that a virtual machine container is entitled to execute a particular application component associated with the multi-tiered application, wherein the amount of the physical resource is determined based at least in part on a ratio of a previous amount of a physical resource that the virtual machine container was consuming over a previous amount of the physical resource that the virtual machine container was entitled to, wherein the amount is automatically modified based on the corrected optimal utilization and the possible size of the virtual machine container; and
using the corrected optimal utilization for all of the application components associated with the multi-tiered application.
4. A system for dynamically resizing a virtual machine container, the system comprising:
a circuit;
a virtual-machine-container-size-determiner configured to determine a possible size of the virtual machine container within a resource-on-demand environment in accordance with a service level agreement;
a resizing-virtual-machine-containers-for-a-multi-tiered-application-optimal-utilization-determiner configured for determining a corrected optimal utilization based on a desired performance for a multi-tiered application and transaction mix information that describes a mix of transactions that result from executing the multi-tiered application, wherein the resizing-virtual-machine-containers-for-a-multi-tiered-application-optimal utilization-determiner is a part of a feed-forward controller;
a feedback controller configured for determining the corrected optimal utilization based on the desired performance, a measured performance, and an uncorrected optimal utilization, wherein the uncorrected optimal utilization is based at least on service time of transactions, the mix of transactions, and fixed delays, ; and
a virtual-machine-container-for-a-multi-tiered-application-automatic-resouce-allocation-modification-enabler configured for enabling automatic modification of an amount of a physical resource a virtual machine container is entitled to for executing a particular application component associated with the multi-tiered application, wherein the amount is automatically modified based on the optimal utilization and the possible size of tile virtual machine container, and wherein the corrected optimal utilization is used for all of the application components associated with the multi-tiered application.
2. The method as recited by
determining different optimal utilizations for two or more application components associated with the multi-tiered application.
3. The method as recited by
minimizing a cost of resources to determine the different optimal utilizations for the two or more application components.
5. The system of
a multiple-application-component-optimal-utilization-determiner configured to determine different optimal utilizations for two or more application components associated with the multi-tiered application.
6. The system of
7. The system of
8. The system of
10. The non-transitory computer readable storage medium of
determining different optimal utilizations for two or more application components associated with the multi-tiered application.
11. The non-transitory computer readable storage medium of
12. The non-transitory computer readable storage medium of
|
The present application claims priority from the provisional application Ser. No.61/007,768, filed Dec. 13, 2007, the contents of which are incorporated herein by reference in their entirety.
Embodiments of the present invention relate to virtual machine container resource entitlement. More specifically, embodiments of the present invention relate to dynamically resizing a virtual machine container.
In many of today's data centers, servers are shared across multiple applications or application components. However, current tools for allocating servers or server partitions to applications typically rely on offline capacity planning and performing a static partitioning of system resources to support these co-hosted applications. For example, each application is allocated a maximum entitlement of system resources for execution over the application's lifetime. Many times, the amount of the maximum entitlement of system resources is either based on anticipated peak load or demand profiles computed from historic data and performance models. However, the entitlement of system resources is static. For example, the entitlement of system resources is determined and used for a long period of time before being re-evaluated. This static entitlement typically results in poor utilization of system resources, which in many instances is below 20%, and does not take full advantage of demands that vary over time due to changes in operating conditions and user demands.
One solution has been to control the number of requests for services provided by a server that are admitted (also commonly known as “admission control”) in order to improve the performance of requests that have already been admitted. However, this results in requests either not being admitted or in requests being dropped.
The accompanying drawings, which are incorporated in and form a part of this specification, illustrate embodiments of the invention and, together with the description, serve to explain the principles of the invention:
The drawings referred to in this description should not be understood as being drawn to scale except if specifically noted.
Reference will now be made in detail to various embodiments of the invention, examples of which are illustrated in the accompanying drawings.
While the invention will be described in conjunction with these embodiments, it will be understood that they are not intended to limit the invention to these embodiments. On the contrary, the invention is intended to cover alternatives, modifications and equivalents, which may be included within the spirit and scope of the invention as defined by the appended claims. Furthermore, in the following description of various embodiments of the present invention, numerous specific details are set forth in order to provide a thorough understanding of the present invention. In other instances, well-known methods, procedures, components, and circuits have not been described in detail as not to unnecessarily obscure aspects of the present invention.
Application component 2 is hosted on container 2 on host 2 and so on. A container is also referred to as a “virtual machine container.” Each application component associated with a multi-tiered application is considered to be on a “tier” of the multi-tiered application. For example, application component 1 is on one tier, application component 2 is on another tier, and application component 3 is on yet another tier.
One or more clients 110 interact with the multi-tiered application which results in a mix of transactions. For example, assume that one (application component 1) of the application components is a web server that enables a user of a client to browse items that are for sale and to order those items and another one (application component 3) of the application components is a database that stores information pertaining to those items. The transaction mix for a particular interval of time may reflect that 20 percent of the transactions pertain to ordering items, 60 percent of the transactions pertain to viewing items, and 20 percent of the transactions pertain to accessing information about the items in the database.
The “size” of a container is determined by the amount of physical resources, such as CPU, memory, storage, network bandwidth, input/output (I/O) bandwidth, that a container is entitled to, according to one embodiment. For example, a container that is entitled to 0.1 of a CPU is smaller than a container that is entitled to 0.2 of a CPU. The amount of a resource that a container is entitled to determines an upper bound on the amount of a resource that the application component associated with the container may use.
As will become more evident, various embodiments provide for determining a new entitlement that is used to resize a container for executing an application component in a multi-tiered application. Various pieces of information can be used as a part of determining a new entitlement. For example, these pieces of information may include, among other things, transaction mix information from client logs, a desired level of performance from a service level agreement, the measured performance from the perspective of a client, and the actual resources consumed by a virtual machine container.
Resource partitioning is a type of virtualization technology that enables multiple applications or application components to share system resources while maintaining performance isolation and differentiation among them. For example, CPUs or portions of a CPU, such as in terms of CPU cycles, may be allocated to different application components or different groups of application components. Partition sizes may be set to accommodate desired performance, such as those that are specified in service level agreements, for example. According to one embodiment, the partition sizes may vary depending on the demand for resources, as will become more evident.
In one embodiment, the resources of a resource-on-demand environment, such as a data center or a grid environment, are partitioned into containers. The containers as depicted in
A container can be represented as a process group that includes one or more application components. There may be a limit to the amount of a particular resource (also known as “entitlement”) that can be assigned to each process group. As depicted in
Different types of process groups, such as PSET and fair share scheduler (FSS), may be used. An integer number of CPUs can be allocated for PSET type process groups. A percentage of CPU cycles can be allocated to an FSS type process group. In one embodiment, a Hewlett-Packard™ HP-UX Process Resource Manager (PRM) is used as an FSS.
In another embodiment, the hosts are partitioned into virtual machines. For example, each container may include a virtual machine where a set of resources are allocated to each virtual machine. Virtual machines provide a technology to implement resource virtualization that allows multiple heterogeneous operating system instances to share a single physical workstation or server. Like a process group, each virtual machine receives an allocation of resources, such as a percentage of CPU capacity, memory capacity, network bandwidth, etc. In yet another embodiment, a container may include a server group.
Examples of a client include but are not limited to a Personal Computer (PC), a Personal Digital Assistant (PDA), a wireless terminal, a cellular telephone, or any other type of well known client device. The client may include servers which may be in different data centers. A host can be any computer, such as a personal computer, a server, etc. . . . that is capable of providing one or more virtual machine containers.
The transaction mix affects the amount and type of physical resources that an application component associated with a container would use. For example, a transaction mix that leans more toward computing orders may be more CPU intensive whereas a transaction mix that leans more toward accessing the database may require more I/O resources. According to one embodiment, the amount of physical resources allocated to a virtual machine container is based at least in part on transaction mix information that describes the mix of transactions that result from executing a multi-tiered application.
Frequently, clients store information that describes a mix of transactions (also referred to herein as “transaction mix information”) in logs. Transaction mix information, according to one embodiment, describes a mix of transactions in terms of intensity and types of transactions. Intensity may involve the number of users or transactions, for example, during a period of time. Examples of transaction types include but are not limited to accessing, storing, viewing, buying and so on.
Transaction mix information for a particular interval of time may reflect that there are 300 transactions executed in that interval of time, where 20 percent of the transactions pertain to ordering items, 60 percent of the transactions pertain to viewing items, and 20 percent of the transactions pertaining to accessing information about the items in the database.
Various performance metrics, such as mean response time, can be used as a part of specifying a desired level of performance (referred to herein as “desired performance”) and as a part of measuring the actual performance (referred to herein as “measured performance”). The desired performance may be a service level objective that is specified in a service level agreement for example between a customer and a service provider. One example of a performance metric is response time, for example, of a client that requests transactions from the multi-tiered application. The measured response time may be a mean response time. Other examples of performance metrics include but are not limited to percentile of response time samples, throughput, and loss rate. Many of the examples of embodiments discussed herein assume that performance is response time.
The service provider may own a data center or a grid environment that provides for example hosts, various resources, and application components using virtualization techniques. The difference between the desired performance and the measured performance provides an indication of how closely a service provider is meeting the desired performance. According to one embodiment, the difference between the desired performance and the measured performance is used as a part of resizing a container, as will become more evident. The difference, according to one embodiment, is expressed in terms of response time (ΔP).
Entitlement, according to one embodiment, is the amount of resources that a container is entitled to use. Consumption, according to one embodiment, is the amount of resources that a container actually uses. A container may not be allowed to consume more resources than it is entitled to, according to one embodiment.
According to one embodiment, utilization is a ratio of consumption over entitlement (U=C/E). For example, if a container is entitled to 1 CPU and the container is using 0.1 of that one CPU, the utilization is 10 percent (0.1/1.0). Assume for the sake of illustration that the optimal utilization is 50 percent. In this case, the entitlement of the container may be modified from 1 CPU to 0.2 CPU in order to cause the actual utilization (0.1/0.2) to match the optimal utilization. Assume for the sake of illustration that the consumption goes up to 0.2. In this case, the actual utilization will be 100 percent (0.2/0.2). The entitlement can be modified to 0.4 so that the actual utilization (0.2/0.4) again matches the optimal utilization. Assume that the consumption drops to 0.1 and the optimal utilization drops to 10 percent. The entitlement can be modified to 1 CPU so that the actual utilization (0.1/1.0) matches the optimal utilization. The discussion of various embodiments shall use the letters E, C, and U respectively to refer to entitlement, consumption and utilization.
According to one embodiment, a transaction-mix-based queuing model as represented by equation 2 is used for a multi-tiered application. For the sake of simplicity in illustrating equation 2, assume that each tier of the multi-tiered application hosts an application component in a single virtual machine container and each container receives a capped resource entitlement. The utilization of each container is defined as a ratio of average consumption and entitlement. Equation 2 can be solved for an uncorrected optimal utilization UUopt. Referring to
In real-time control, the transaction mix in the current interval may not be known before control is actuated. According to one embodiment, these values are estimated based on historical data. Different approaches may be used. For example, transaction mix information from the last interval may be used to estimate for the current interval.
In order to avoid multiple possible solutions for UUopt, according to one embodiment, equation 2 is solved for a single UUopt for all application components associated with a multi-tiered application. According to another embodiment, equation 2 is solved for a different UUopt for each application component associated with a multi-tiered application. For example, the performance utilization of the multiple tiers can be optimized such that the total cost of the resources is minimized, for example, using the equation f(E1, E2, E3).
Since models may not be totally correct all of the time, the results from equation 2 shall be referred to as an uncorrected optimal utilization UUopt. According to one embodiment, the correction term ΔU (refer to equation 1), which was determined based on the difference between the desired performance and the measured performance, is used to correct the UUopt resulting in a corrected optimal utilization UCopt, i.e., UCopt,=UUopt+ΔU.
Equation 1 depicted on
According to one embodiment GFB is 0.05. Other examples of GFB include but are not limited to 0.09 and 0.02. Typically, the larger that GFB is the more quickly adjustments will be made. According to one embodiment, if the utilization target is expressed as a percentage and therefore is in the range of [0, 1], the error term is normalized by the response time targets (assuming the performance is response time targets) so that a larger stability margin exists for a certain gain value.
Equation 3 depicted on
According to one embodiment, a new amount of the physical resource is determined based at least in part on a ratio of the previous amount of a physical resource that the virtual machine container was consuming over a previous amount of the physical resource that the virtual machine container was entitled to. For example, referring to equation 3, E1(k) is an example of a new amount and E1(k-1) is an example of a previous amount entitled to and C1 (k-1) is an example of a previous amount consumed. C1/E1 (k-1) is an example of a measured utilization for a container in a previous interval.
According to one embodiment, the determiner 220 and enabler 230 can be located on either a client or a server. Transaction mix information in logs associated with the client can be accessed by the determiner 220, for example, if the determiner resides on a client. Alternatively, information from the transaction mix information in the logs can be received by the determiner 220, for example, if the determiner 220 resides on a client or a server. The consumption of a container can be accessed by an enabler 230 that resides on a client or can be received by an enabler 230 that resides on a client or a server. According to one embodiment, system 200 may also include a correction determiner for determining ΔU (refer to equation 1 depicted on
The system 310 is depicted communicating with a client 320, for example, over a network. The network may be a private network such as an intranet, or a public network, such as the Internet, or a combination thereof. The system 300 may be a part of a resource-on-demand environment, such as a data center or a grid environment, which is partitioned into containers that the application components are associated with.
The determiner 220 (
Various pieces of information can be used as a part of determining a new entitlement (E1, E2, E3). For example, these pieces of information may include, among other things, transaction mix information from client logs, a desired level of performance (Pdes), for example, from a service level agreement, the measured performance (Pm), for example, from the perspective of a client, and the actual resources consumed (C1, C2, C3) by a virtual machine container. These various pieces of information can be used as a part of determining, for example at the server side, optimal utilization (UUopt or UCopt or a combination thereof) and in turn entitlement (E1, E2, E3).
All of, or a portion of, the embodiments described by flowchart 400 can be implemented using computer-readable and computer-executable instructions which reside, for example, in computer-usable media of a computer system or like device. As described above, certain processes and steps of the present invention are realized, in an embodiment, as a series of instructions (e.g., software program) that reside within computer readable memory of a computer system and are executed by the of the computer system. When executed, the instructions cause the computer system to implement the functionality of the present invention as described below.
The description of the flowchart shall refer to
At step 410, the process begins.
In step 420, an optimal utilization is determined based on a desired performance for a multi-tiered application and transaction mix information that describes a mix of transactions that result from executing the multi-tiered application. For example, the feed-forward controller 312, according to one embodiment, is a queuing model as represented by equation 2. The feed-forward controller 312 receives the desired performance (Pdes) from the client 320 and the transaction mix information from transaction logs associated with the client 320. The feed-forward controller 312 uses equation 2 to determine the uncorrected optimal utilization level (UUopt).
The feedback controller 314 receives a measured performance (Pm), such as a Measured Mean Response Time, from a client 320 and a desired performance (Pdes), such as a desired response time from, a service level agreement. The feedback controller 314 determines the correction term (ΔU) for the optimal utilization based on the difference between the measured performance (Pm) and the desired performance (Pdes). The feedback controller 314 uses equation 1 to determine correction term (ΔU) in terms of utilization.
At the + sign, a corrected optimal utilization level (UCopt) is determined based on the uncorrected optimal utilization level (UUopt) and the correction term (ΔU). As described herein, the corrected optimal utilization level may be the same for all of the utilization controllers 316, 317, 318 or may be different for 2 or more of the utilization controllers 316, 317, 318.
At step 430, an amount of a physical resource a virtual machine container is entitled to for executing a particular application component associated with the multi-tiered application is enabled. The amount of a physical resource is automatically modified based on the optimal utilization. For example, the utilization controllers 316-318 receive the corrected optimal utilization level(s) (UCopt). The utilization controllers 316-318 also receive resource consumption C1, C2, C3 from respective containers 1, 2, 3 (
The one or more new entitlements E1, E2, E3 are determined automatically and enable automatic modification of the amount of physical resources that a virtual machine container 1-3 is entitled to.
Therefore, various embodiments enable automatically and dynamically resizing a virtual machine container so that virtual machine containers are not required to be statically partitioned. Further, various embodiments enable automatically and dynamically resizing a virtual machine container without requiring admission control.
At 440, the process stops.
An experimental evaluation of various embodiments was performed. A configuration as depicted in
A modified version of the Rice University Bidding system (RUBiS) was used as the test application. It is an online auction benchmark with 22 transaction types providing services such as browsing for items, placing a bid, and viewing user information. In the test bed, the multi-tiered application runs on top of an Apache™ 2.0.55 Web server, a JBoss™ 4.0.2 application server and a MySQL™ database server.
A workload generator was developed to simulate a real world workload instead of using the default RUBiS generator. The workload generator bundled with RUBiS would be unsuitable because it produces only a stationary workload, in the sense that the relative frequencies of the different transaction types remain constant over time, whereas real world workloads are highly non-stationary in terms of transaction mix. The developed workload generator had two features that support much more realistic evaluations than are possible with typical benchmark-like generators. First, the developed generator allows for replaying renamed traces of transactions collected on real production systems; this means that the test workloads contain the same non-stationary behavior of transaction-mix found in production workloads. Second, the developed generator allows mimicking periodicity in real workloads (e.g., diumal cycles) by varying transaction arrival rates between specified maximum and minimum levels.
The developed workload was derived from a trace of transactions in the “VDR” application, a globally distributed business-critical enterprise application. The VDR transactions were renamed to RUBiS transactions in the following way: First, VDR and RUBiS transactions were separately ranked according to their popularity; for the later, the popularity of transactions in the workload generated by the default RUBiS generator was used. Then each transaction was replaced in the VDR trace with the RUBiS transaction with the same popularity rank to obtain a final renamed trace.
A customized workload generator, driven by a VDR-based trace was used. Client transactions were submitted to the RUBiS application as a Poisson process. For the experimental evaluation, the utilization controllers were located in Dom-0 for each of the three tiers, as it usually is in real products. The feed-forward and feedback controllers were both located at the client side, where the transaction mix and the response times were sampled from the logs of the client. The utilization targets were sent to the three containers through SSH calls. The sampling intervals for the utilization, feed-forward controllers, k, m, and i, were set respectively to 10, 30 and 90 seconds. The parameter P of the utilization controller was set to 1.0. GFB of the feedback controller was set to 0.5.
The results from the Autoparam system were compared with results from four other ways of controlling container size. Those four other ways are as follows:
(1) Fixed entitlement, or abbreviated as “Fixed Ent.” Each virtual container was entitled to 0.9 CPU in this case.
(2) Fixed utilization control, or “Fixed Util.” The optimal utilization was statically set to 75% in this case.
(3) Feed-forward plus utilization control, or “FF+Util.” No feedback compensation was applied for the response time control. The feed-forward control is expected to provide proactive and fast actions.
(4) Nested control, i.e., feedback performance control plus utilization control, or “FB+Util.” Compared with Autoparam, this design does not include the feed-forward part. For simplicity, a standard integrator as described by equation 1 was used for the feedback controller, except that the output of the controller is the utilization target itself instead of delta U.
Table 1 below shows the performance, i.e., the average of the per-interval mean response time across all the intervals in each experiment running for 2 hours. The average CPU entitlement across the three containers is also shown in number of CPUs.
TABLE 1
Experimental Results
Control Designs
Mean of MRT
Total entitlement
(targets)
(ms)
(# of CPUs)
Fixed Ent
204
2.7
Fixed Util (75%)
578
0.53
FF + Util (300 ms)
329
0.66
FB + Util (300 ms)
288
0.86
Autoparam (300 ms)
320
0.70
As can be seen Autoparam demonstrated a reasonable capacity allocation and achieved a reasonable mean response time.
Various embodiments have been described in various combinations. However, any two or more embodiments may be combined. Further, any embodiment may be used separately from other embodiments.
Embodiments of the present invention are thus described. While the present invention has been described in particular embodiments, it should be appreciated that the present invention should not be construed as limited by such embodiments, but rather construed according to the following claims.
Singhal, Sharad, Liu, Xue, Stewart, Christopher, Zhu, Xiaoyun, Wang, Zhikui, Kelly, Terence, Zhang, Alex
Patent | Priority | Assignee | Title |
10073974, | Jul 21 2016 | International Business Machines Corporation | Generating containers for applications utilizing reduced sets of libraries based on risk analysis |
10140162, | Jul 23 2015 | International Business Machines Corporation | Managing a shared pool of configurable computing resources using a set of scaling factors and a set of workload resource data |
10146586, | Jul 23 2015 | International Business Machines Corporation | Managing a shared pool of configurable computing resources using a set of scaling factors and a set of workload resource data |
10169086, | Sep 13 2015 | International Business Machines Corporation | Configuration management for a shared pool of configurable computing resources |
10176017, | Sep 13 2015 | International Business Machines Corporation | Configuration management for a shared pool of configurable computing resources |
10277708, | Jun 30 2016 | Amazon Technologies, Inc | On-demand network code execution with cross-account aliases |
10282229, | Jun 28 2016 | Amazon Technologies, Inc | Asynchronous task management in an on-demand network code execution environment |
10296363, | Sep 16 2016 | Oracle International Corporation | Tuning a virtual machine startup parameter |
10353678, | Feb 05 2018 | Amazon Technologies, Inc. | Detecting code characteristic alterations due to cross-service calls |
10353746, | Dec 05 2014 | Amazon Technologies, Inc. | Automatic determination of resource sizing |
10365985, | Dec 16 2015 | Amazon Technologies, Inc. | Predictive management of on-demand code execution |
10387177, | Feb 04 2015 | Amazon Technologies, Inc. | Stateful virtual compute system |
10402231, | Jun 29 2016 | Amazon Technologies, Inc. | Adjusting variable limit on concurrent code executions |
10430226, | May 03 2013 | VMware LLC | Dynamic virtual machine sizing |
10528390, | Sep 23 2016 | Amazon Technologies, Inc. | Idempotent task execution in on-demand network code execution systems |
10552193, | Feb 04 2015 | Amazon Technologies, Inc. | Security protocols for low latency execution of program code |
10564946, | Dec 13 2017 | Amazon Technologies, Inc. | Dependency handling in an on-demand network code execution system |
10592269, | Sep 30 2014 | Amazon Technologies, Inc. | Dynamic code deployment and versioning |
10623476, | Apr 08 2015 | Amazon Technologies, Inc. | Endpoint management system providing an application programming interface proxy service |
10691498, | Dec 21 2015 | Amazon Technologies, Inc. | Acquisition and maintenance of compute capacity |
10725752, | Feb 13 2018 | Amazon Technologies, Inc.; Amazon Technologies, Inc | Dependency handling in an on-demand network code execution system |
10733085, | Feb 05 2018 | Amazon Technologies, Inc. | Detecting impedance mismatches due to cross-service calls |
10754701, | Dec 16 2015 | Amazon Technologies, Inc | Executing user-defined code in response to determining that resources expected to be utilized comply with resource restrictions |
10776091, | Feb 26 2018 | Amazon Technologies, Inc. | Logging endpoint in an on-demand code execution system |
10776171, | Apr 08 2015 | Amazon Technologies, Inc. | Endpoint management system and virtual compute system |
10824484, | Sep 29 2015 | Amazon Technologies, Inc. | Event-driven computing |
10831898, | Feb 05 2018 | Amazon Technologies, Inc. | Detecting privilege escalations in code including cross-service calls |
10853112, | Feb 04 2015 | Amazon Technologies, Inc. | Stateful virtual compute system |
10884722, | Jun 26 2018 | Amazon Technologies, Inc. | Cross-environment application of tracing information for improved code execution |
10884787, | Sep 23 2016 | Amazon Technologies, Inc | Execution guarantees in an on-demand network code execution system |
10884802, | Sep 30 2014 | Amazon Technologies, Inc. | Message-based computation request scheduling |
10884812, | Dec 13 2018 | Amazon Technologies, Inc | Performance-based hardware emulation in an on-demand network code execution system |
10891145, | Mar 30 2016 | Amazon Technologies, Inc | Processing pre-existing data sets at an on demand code execution environment |
10908927, | Sep 27 2019 | Amazon Technologies, Inc | On-demand execution of object filter code in output path of object storage service |
10915371, | Sep 30 2014 | Amazon Technologies, Inc. | Automatic management of low latency computational capacity |
10942795, | Nov 27 2019 | Amazon Technologies, Inc | Serverless call distribution to utilize reserved capacity without inhibiting scaling |
10949237, | Jun 29 2018 | Amazon Technologies, Inc | Operating system customization in an on-demand network code execution system |
10956185, | Sep 30 2014 | Amazon Technologies, Inc. | Threading as a service |
10996961, | Sep 27 2019 | Amazon Technologies, Inc | On-demand indexing of data in input path of object storage service |
11003504, | Jun 28 2019 | COHESITY, INC | Scaling virtualization resource units of applications |
11010188, | Feb 05 2019 | Amazon Technologies, Inc. | Simulated data object storage using on-demand computation of data objects |
11016815, | Dec 21 2015 | Amazon Technologies, Inc. | Code execution request routing |
11023311, | Sep 27 2019 | Amazon Technologies, Inc | On-demand code execution in input path of data uploaded to storage service in multiple data portions |
11023416, | Sep 27 2019 | Amazon Technologies, Inc | Data access control system for object storage service based on owner-defined code |
11055112, | Sep 27 2019 | Amazon Technologies, Inc | Inserting executions of owner-specified code into input/output path of object storage service |
11099870, | Jul 25 2018 | Amazon Technologies, Inc. | Reducing execution times in an on-demand network code execution system using saved machine states |
11099917, | Sep 27 2018 | Amazon Technologies, Inc | Efficient state maintenance for execution environments in an on-demand code execution system |
11106477, | Sep 27 2019 | Amazon Technologies, Inc | Execution of owner-specified code during input/output path to object storage service |
11115404, | Jun 28 2019 | Amazon Technologies, Inc | Facilitating service connections in serverless code executions |
11119809, | Jun 20 2019 | Amazon Technologies, Inc; Amazon Technologies, Inc. | Virtualization-based transaction handling in an on-demand network code execution system |
11119813, | Sep 30 2016 | Amazon Technologies, Inc | Mapreduce implementation using an on-demand network code execution system |
11119826, | Nov 27 2019 | Amazon Technologies, Inc | Serverless call distribution to implement spillover while avoiding cold starts |
11126469, | Dec 05 2014 | Amazon Technologies, Inc. | Automatic determination of resource sizing |
11132213, | Mar 30 2016 | Amazon Technologies, Inc | Dependency-based process of pre-existing data sets at an on demand code execution environment |
11146569, | Jun 28 2018 | Amazon Technologies, Inc. | Escalation-resistant secure network services using request-scoped authentication information |
11159528, | Jun 28 2019 | Amazon Technologies, Inc | Authentication to network-services using hosted authentication information |
11188391, | Mar 11 2020 | Amazon Technologies, Inc.; Amazon Technologies, Inc | Allocating resources to on-demand code executions under scarcity conditions |
11190609, | Jun 28 2019 | Amazon Technologies, Inc | Connection pooling for scalable network services |
11243819, | Dec 21 2015 | Amazon Technologies, Inc. | Acquisition and maintenance of compute capacity |
11243953, | Sep 27 2018 | Amazon Technologies, Inc | Mapreduce implementation in an on-demand network code execution system and stream data processing system |
11250007, | Sep 27 2019 | Amazon Technologies, Inc | On-demand execution of object combination code in output path of object storage service |
11263034, | Sep 30 2014 | Amazon Technologies, Inc. | Low latency computational capacity provisioning |
11263220, | Sep 27 2019 | Amazon Technologies, Inc | On-demand execution of object transformation code in output path of object storage service |
11354169, | Jun 29 2016 | Amazon Technologies, Inc. | Adjusting variable limit on concurrent code executions |
11360793, | Feb 04 2015 | Amazon Technologies, Inc. | Stateful virtual compute system |
11360948, | Sep 27 2019 | Amazon Technologies, Inc | Inserting owner-specified data processing pipelines into input/output path of object storage service |
11386230, | Sep 27 2019 | Amazon Technologies, Inc | On-demand code obfuscation of data in input path of object storage service |
11388210, | Jun 30 2021 | Amazon Technologies, Inc.; Amazon Technologies, Inc | Streaming analytics using a serverless compute system |
11394761, | Sep 27 2019 | Amazon Technologies, Inc | Execution of user-submitted code on a stream of data |
11416628, | Sep 27 2019 | Amazon Technologies, Inc | User-specific data manipulation system for object storage service based on user-submitted code |
11461124, | Feb 04 2015 | Amazon Technologies, Inc. | Security protocols for low latency execution of program code |
11467890, | Sep 30 2014 | Amazon Technologies, Inc. | Processing event messages for user requests to execute program code |
11550713, | Nov 25 2020 | Amazon Technologies, Inc | Garbage collection in distributed systems using life cycled storage roots |
11550944, | Sep 27 2019 | Amazon Technologies, Inc | Code execution environment customization system for object storage service |
11561811, | Sep 30 2014 | Amazon Technologies, Inc. | Threading as a service |
11593270, | Nov 25 2020 | Amazon Technologies, Inc; INC , AMAZON TECHNOLOGIES | Fast distributed caching using erasure coded object parts |
11656892, | Sep 27 2019 | Amazon Technologies, Inc | Sequential execution of user-submitted code and native functions |
11714675, | Jun 20 2019 | Amazon Technologies, Inc. | Virtualization-based transaction handling in an on-demand network code execution system |
11714682, | Mar 03 2020 | Amazon Technologies, Inc. | Reclaiming computing resources in an on-demand code execution system |
11755384, | Jun 28 2019 | Cohesity, Inc. | Scaling virtualization resource units of applications |
11775640, | Mar 30 2020 | Amazon Technologies, Inc | Resource utilization-based malicious task detection in an on-demand code execution system |
11797327, | May 03 2013 | VMware LLC | Dynamic virtual machine sizing |
11836516, | Jul 25 2018 | Amazon Technologies, Inc. | Reducing execution times in an on-demand network code execution system using saved machine states |
11860879, | Sep 27 2019 | Amazon Technologies, Inc. | On-demand execution of object transformation code in output path of object storage service |
11861386, | Mar 22 2019 | Amazon Technologies, Inc.; Amazon Technologies, Inc | Application gateways in an on-demand network code execution system |
11875173, | Jun 25 2018 | Amazon Technologies, Inc. | Execution of auxiliary functions in an on-demand network code execution system |
11943093, | Nov 20 2018 | Amazon Technologies, Inc | Network connection recovery after virtual machine transition in an on-demand network code execution system |
11968280, | Nov 24 2021 | Amazon Technologies, Inc | Controlling ingestion of streaming data to serverless function executions |
12164970, | Jun 28 2019 | Cohesity, Inc. | Scaling virtualization resource units of applications |
9785460, | May 03 2013 | VMware LLC | Dynamic virtual machine sizing |
9785474, | Jul 23 2015 | International Business Machines Corporation | Managing a shared pool of configurable computing resources using a set of scaling factors and a set of workload resource data |
9785475, | Jul 23 2015 | International Business Machines Corporation | Managing a shared pool of configurable computing resources using a set of scaling factors and a set of workload resource data |
ER3610, | |||
ER5949, |
Patent | Priority | Assignee | Title |
5442730, | Oct 08 1993 | International Business Machines Corporation | Adaptive job scheduling using neural network priority functions |
6662221, | Apr 12 1999 | Lucent Technologies, INC | Integrated network and service management with automated flow through configuration and provisioning of virtual private networks |
7676578, | Jul 25 2006 | Hewlett Packard Enterprise Development LP | Resource entitlement control system controlling resource entitlement based on automatic determination of a target utilization and controller gain |
7680916, | Apr 24 2007 | LTI EQUIPMENT FINANCE CORP | System for improving the performance of a computer software application in a server network |
7720955, | Oct 10 2006 | Hewlett Packard Enterprise Development LP | Determining performance of an application based on transactions |
20030014524, | |||
20030028624, | |||
20040143664, | |||
20050091310, | |||
20050240558, | |||
20070043860, | |||
20070074191, | |||
20070110077, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Feb 11 2008 | ZHU, XIAOYUN | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 022993 | /0274 | |
Feb 11 2008 | LIU, XUE | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 022993 | /0274 | |
Feb 12 2008 | WANG, ZHIKUI | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 022993 | /0274 | |
Feb 12 2008 | Zhang, Xin | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 022993 | /0274 | |
Feb 12 2008 | STEWART, CHRISTOPHER CHARLES | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 022993 | /0274 | |
Feb 12 2008 | KELLY, TERENCE P | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 022993 | /0274 | |
Feb 12 2008 | SINGHAL, SHARAD | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 022993 | /0274 | |
Oct 31 2008 | Hewlett-Packard Development Company, L.P. | (assignment on the face of the patent) | / | |||
Oct 27 2015 | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | Hewlett Packard Enterprise Development LP | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 037079 | /0001 |
Date | Maintenance Fee Events |
Mar 21 2017 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Mar 23 2021 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Jan 02 2025 | M1553: Payment of Maintenance Fee, 12th Year, Large Entity. |
Date | Maintenance Schedule |
Oct 22 2016 | 4 years fee payment window open |
Apr 22 2017 | 6 months grace period start (w surcharge) |
Oct 22 2017 | patent expiry (for year 4) |
Oct 22 2019 | 2 years to revive unintentionally abandoned end. (for year 4) |
Oct 22 2020 | 8 years fee payment window open |
Apr 22 2021 | 6 months grace period start (w surcharge) |
Oct 22 2021 | patent expiry (for year 8) |
Oct 22 2023 | 2 years to revive unintentionally abandoned end. (for year 8) |
Oct 22 2024 | 12 years fee payment window open |
Apr 22 2025 | 6 months grace period start (w surcharge) |
Oct 22 2025 | patent expiry (for year 12) |
Oct 22 2027 | 2 years to revive unintentionally abandoned end. (for year 12) |