A computing system implemented method, in one embodiment, can include a cloud control module receiving a constraint for cloud architecture. In addition, the method can include the cloud control module receiving a plurality of cloud service provider capabilities. Furthermore, the method can include the cloud control module filtering the plurality of cloud service provider capabilities to identify a cloud service provider capable of satisfying the constraint. Moreover, the method can include the cloud control module outputting an instruction for a resource from the cloud service provider.
|
1. A method comprising:
in a cloud computing configuration system, with cloud control circuitry:
determining to implement a cloud architecture by deploying instances of a workload component;
determining constraints applicable to the workload component;
determining capabilities of available cloud computing platforms hosted by different providers;
selecting deployment cloud computing platforms from among the available cloud computing platforms responsive to the constraints applicable to the workload component and the capabilities of the cloud computing platforms; and
with instantiation circuitry:
for each of the deployment cloud computing platforms that were selected, perform a respective sizing and sourcing determination based on information received from the respective deployment cloud computing platform of the deployment cloud computing platforms that were selected, the sizing and sourcing determination including an assignment of roles for the instances of the workload component;
based on the respective sizing and sourcing determinations, provisioning and configuring a workload distribution script, in accord with the assignment of roles, for deployment of the instances of the workload component, the workload distribution script configured to install a pre-defined image of the workload component; and
initiating the deployment of the instances of the workload component to the deployment cloud computing platforms, where multiple different providers host the deployment cloud computing platforms.
8. A system comprising:
control circuitry configured to:
determine to scale an existing cloud computing platform via deployment of instances of a workload component of the existing cloud computing platform;
determine constraints applicable to the workload component;
determine capabilities of available cloud computing platforms hosted by different providers;
analyze the constraints and the capabilities, and responsively select deployment cloud computing platforms from among the available cloud computing platforms; and
transformation circuitry configured to:
after selection of the deployment cloud computing platforms, obtain a transformed instance by transforming a first instance of the instances of the workload component responsive to a computing environment for a first deployment cloud computing platform of the deployment cloud computing platforms; and
instantiation circuitry configured to:
for each of the deployment cloud computing platforms that were selected, perform a respective sizing and sourcing determination based on information received from the respective deployment cloud computing platform of the deployment cloud computing platforms that were selected, the sizing and sourcing determination including an assignment of roles for the instances of the workload component;
based on the respective sizing and sourcing determinations, provision and configure a workload distribution script, in accord with the assignment of roles, for the deployment of the instances of the workload component, the workload distribution script configured to install a pre-defined image of the workload component; and
initiate the deployment of the instances of the workload component to the deployment cloud computing platforms by deploying the transformed instance to the first deployment cloud computing platform, the deployment cloud computing platforms characterized in that multiple different vendors provide the deployment cloud computing platforms.
17. A system comprising:
control circuitry configured to:
obtain a performance constraint for an existing cloud computing platform;
determine whether the performance constraint meets a defined threshold;
when the performance constraint does not meet the defined threshold:
identify a computing component as a cause that the performance constraint does not meet the defined threshold;
determine to increase capacity for the existing cloud computing platform via deployment of instances of the computing component;
determine a constraint applicable to the computing component;
determine a capability of available cloud computing platforms hosted by different providers; and
analyze the constraint and the capability and responsively select deployment cloud computing platforms from among the available cloud computing platforms; and
transformation circuitry configured to:
after selection of the deployment cloud computing platforms, obtain a transformed instance by transforming a first instance of the instances of the computing component responsive to a computing environment for a first deployment cloud computing platform of the deployment cloud computing platforms; and
instantiation circuitry configured to:
receive a workload assignment from the control circuitry with regard to the deployment cloud computing platforms that were selected; and
for each of the deployment cloud computing platforms that were selected, perform a respective sizing and sourcing determination based on information received from the respective deployment cloud computing platform of the deployment cloud computing platforms that were selected, the sizing and sourcing determination including an assignment of roles for the instances of the computing component;
based on the respective sizing and sourcing determinations, provision and configure a computing distribution script, in accord with the assignment of roles, for the deployment of the instances of the computing component, the computing distribution script configured to install a pre-defined image of the computing component; and
initiate the deployment of the instances of the computing component to the deployment cloud computing platforms provided by multiple different vendors by deploying the transformed instance to the first deployment cloud computing platform.
2. The method of
the workload component comprises a software instance, and the deployment cloud computing platforms will execute the software instances.
3. The method of
providing configuration information to the deployment cloud computing platforms for executing the workload component.
4. The method of
determining to implement the cloud architecture comprises:
determining to scale an existing cloud computing platform that comprises already-deployed instances of the workload component.
5. The method of
determining to scale comprises:
obtaining a service level requirement from a configuration database; and
determining that the service level requirement does not meet a defined threshold with respect to the existing cloud computing platform.
6. The method of
identifying the workload component as a cause that the service level requirement does not meet the defined threshold.
7. The method of
determining that the workload component is horizontally scalable by performing common functionality in parallel across the deployment cloud computing platforms, responsive to a probability that the deployment cloud computing platforms are unable to meet a demand.
9. The system of
the instantiation circuitry is further configured to receive a workload assignment from the control circuitry that specifies the deployment cloud computing platforms.
10. The system of
the control circuitry is further configured to filter the available cloud computing platforms to select the deployment cloud computing platforms.
11. The system of
the control circuitry is configured to filter the available cloud computing platforming using a multiple parameter requirements filter.
12. The system of
the multiple parameter requirements filter is responsive to:
compliance requirements for the available cloud computing platforms.
13. The system of
the multiple parameter requirements filter is responsive to:
capability requirements for the available cloud computing platforms and the capabilities of the available cloud computing platforms.
14. The system of
the control circuitry is further configured to:
obtain a service level requirement from a configuration database; and
determine that the service level requirement does not meet a defined threshold with respect to the existing cloud computing platform.
15. The system of
the control circuitry is further configured to:
identify the workload component as a cause that the service level requirement does not meet the defined threshold.
16. The system of
the control circuitry is further configured to determine that the workload component is horizontally scalable by performing common functionality in parallel across the deployment cloud computing platforms, responsive to a probability that the deployment cloud computing platforms are unable to meet a demand.
19. The system of
the performance constraint comprises completion of items to be processed within a batch job within a specified time allotment.
20. The system of
the control circuitry is further configured to:
determine how many deployment cloud computing platforms to use for the deployment in order to bound session blocking probability to less than a specified probability.
|
This application is a continuation of U.S. application Ser. No. 14/510,408 filed Oct. 9, 2014, which is a continuation of U.S. application Ser. No. 13/071,442 filed Mar. 24, 2011 (now U.S. Pat. No. 8,886,806), which claims priority to provisional application Ser. No. 61/321,761, filed Apr. 7, 2010, all of which are entirely incorporated by reference.
Cloud computing is appealing to most enterprises because of its promise of commodity and on-demand utility computing. However, cloud computing faces key barriers to adoption. For example, key policy issue barriers can include legal, indemnity, and compliance. In addition, key technology issue barriers can involve security, reliability, and performance. Furthermore, what makes navigating these issues difficult is that cloud-based application architecture is confusing. First, the term “cloud” is overloaded and used by many vendors each with vastly differing offerings. Second, the on-demand and scalability features of cloud computing are new and architects are still working to understand how to leverage these capabilities.
A computing system implemented method, in one embodiment, can include a cloud control module receiving a constraint for cloud architecture. In addition, the method can include the cloud control module receiving a plurality of cloud service provider capabilities. Furthermore, the method can include the cloud control module filtering the plurality of cloud service provider capabilities to identify a cloud service provider capable of satisfying the constraint. Moreover, the method can include the cloud control module outputting an instruction for a resource from the cloud service provider. Note that in various embodiments, the instruction for a resource can be implemented as, but is not limited to, a number of appliances to provision from the cloud service provider, a selection of a vendor provided platform from the cloud service provider, and/or a selection of a vendor provided software application from the cloud service provider.
While particular embodiments in accordance with the disclosure have been specifically described within this Summary, it is noted that the disclosure and the claimed subject matter are not limited in any way by these embodiments.
The discussion below makes reference to [ ]. Within the accompanying drawings, various embodiments in accordance with the disclosure are illustrated by way of example and not by way of limitation. It is noted that like reference numerals denote similar elements throughout the drawings.
Reference will now be made in detail to various embodiments in accordance with the disclosure, examples of which are illustrated in the accompanying drawings. While the disclosure will be described in conjunction with various embodiments, it will be understood that these various embodiments are not intended to limit the disclosure. On the contrary, the disclosure is intended to cover alternatives, modifications and equivalents, which may be included within the scope of the disclosure. Furthermore, in the following detailed description of various embodiments in accordance with the disclosure, numerous specific details are set forth in order to provide a thorough understanding of the disclosure. However, it will be evident to one of ordinary skill in the art that the disclosure may be practiced without these specific details. In other instances, well known methods, procedures, components, and circuits have not been described in detail as not to unnecessarily obscure aspects of the disclosure.
Some portions of the detailed descriptions that follow are presented in terms of procedures, logic blocks, processing, and other symbolic representations of operations on data bits within a computer memory. These descriptions and representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. In the present application, a procedure, logic block, process, or the like, is conceived to be a self-consistent sequence of operations or steps or instructions leading to a desired result. The operations or steps are those utilizing physical manipulations of physical quantities. Usually, although not necessarily, these quantities can take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated in a computer system or computing device. It has proven convenient at times, principally for reasons of common usage, to refer to these signals as transactions, bits, values, elements, symbols, characters, samples, pixels, or the like.
It is noted that in one embodiment in accordance with the disclosure, a general description of “cloud” is a collection of network-hosted services accessible from almost anywhere with the following attributes: elastically scalable, illusion of infinite capacity, available on-demand, and consumption-based charges with no upfront commitment, but is not limited to such. This description can apply to all levels of the cloud (e.g., process, software application, compute platform, or virtual machine hosting).
In various embodiments, one or more of the operations of the cloud management system 100 can be performed by software, by firmware, by hardware, or by any combination thereof, but is not limited to such. The cloud management system 100 can include processes of embodiments of the disclosure which can be carried out by a processor(s) and electrical components under the control of computer or computing device readable and executable instructions (or code). The computer or computing device readable and executable instructions (or code) may reside, for example, in data storage features such as computer or computing device usable volatile memory, computer or computing device usable non-volatile memory, and/or computer or computing device usable mass data storage. However, the computer or computing device readable and executable instructions (or code) may reside in any type of computer or computing device readable medium.
Within
Within
The cloud Instantiation Layer 106 in one embodiment can execute the specification output by the cloud control layer 104 (e.g., automation of scripts provision, maintain, and tear-down the prescribed deployment of virtual appliances). It is noted that as part of provisioning, the scripts can add the assigned content, software, data, configurations, and connectivity to form the application platform. For example in an embodiment, the cloud Instantiation Layer 106 can apply the Internet Protocol (IP) address and update the Domain Name System (DNS). Working with the Control Layer 104, the Instantiation Layer 106 can pass monitored statistics and then implement the updates when sourcing arrangements are recomputed. The Instantiation Layer 106 can also govern how provisioning and configuration occurs specific to the cloud-based service provider (e.g., data center, EC2, or GoGrid).
Within
Within an embodiment, the scaling algorithms or procedures of the cloud Control Layer 104 can be generic and independent of application or platform. In networking, the Transmission Control Protocol (TCP) can provide reliable end-to-end delivery of a packet stream regardless of the content being delivered or the network implementation. In a similar manner, the cloud Control Layer 104 algorithms or procedures can source and possibly scale the compute platform without knowledge of the content (e.g., the application program and data), or the details of the instantiation or implementation of the virtual machine appliances or vendor packaged platform. The sourcing algorithm or procedure works to prioritize preferred providers and filter out non-compliant providers that do not fulfill application requirements; then the scaling algorithm or procedure works to meet availability and performance objectives. The instantiation of the outputs prescribed by the Control Layer 104 are left to other orchestration mechanisms (e.g., provision and configure the virtual machine appliances).
In an embodiment, logic or functionality of the Cloud Control Layer 204 of the cloud management system 200 can consider various desired attributes like, but not limited to, performance, availability, desired cost, and compliance; and chooses the best set and number of resources based on the offered capability of providers or vendors with defined terms like, but not limited to, capability, average capacity, maximum capacity, start-up time, availability, published cost, and/or location. It is pointed out that a provider's values for these terms may be based on published metrics or values, or estimated from monitored and measured metrics collected over time, or collected measurements of the provider, or any combination thereof. Note that a provider's vector of values for performing a given capability is enough to summarize the application constraints for the Control Layer 204. Knowledge of the specific implementation within the virtual machine appliance or vendor packaged platform is not needed. In addition, note that when the capability is summarized, the algorithms (or functions or procedures) of the cloud Control Layer 204 do not deal directly with virtual machine appliance implementation.
Within
When determined to build a platform, the logic of the cloud Control Layer 204 can stipulate the scaling to meet performance and availability. For example in an embodiment, the cloud Control Layer 204 chooses a number of primary active nodes needed to meet performance constraints. Then, a number of additional redundant nodes needed to guarantee availability is determined by the cloud Control Layer 204. The Cloud Control Layer 204 can take these numbers along with a cost calculation that multiplies the number of resources and their cost thereby resulting in a system of equations or procedures for determining the required resources.
Or instead of or in addition to building a platform, the logic of the cloud Control Layer 204 can stipulate a vendor packaged platform to meet performance and availability. For example in an embodiment, the cloud Control Layer 204 compares the price of a vendor packaged cloud-based software application service that satisfies SLA constraints with that of a self-sourced version. Then the cloud Control Layer 204 can determine the vendor packaged platform, self-sourced platform, or any combination that minimizes costs needed to guarantee performance and availability.
Within
Based on the determination of the Cloud Control Layer 204, the cloud Instantiation Layer 206 automates the appropriate scripts to provision the appliances or packaged platform, connect the resulting appliances or platforms, and tear-down appliances or platforms when finished. It is noted that in one embodiment, the cloud Instantiation Layer 206 can use a script that installs a pre-defined image of the appliances, but is not limited to such. Implementation of logic at the Cloud Control Layer 204 can determine the best possible or appropriate sourcing in terms of the quantity and placement of the appliances or packaged platforms and allows realization of custom attributes.
Within
Within the cloud management system 200, in one embodiment, security presents different concerns at each of the Cloud Transformation Layer 202, the Cloud Control Layer 204, and the Cloud Instantiation Layer 206. For example, the Cloud Instantiation Layer 206 can apply network port configurations to minimize intrusion, can constantly monitor against compromise or network flooding, and can signal the Cloud Control Layer 204 for updates when adverse conditions are detected. In addition, the Cloud Transformation Layer 202 can provide encryption. As such, each layer adds a different level of protection.
Within
At operation 210 of the cloud management system 200, the cloud control layer 204 can receive business service level agreements (SLAs). It is noted that operation 210 can be implemented in a wide variety of ways. For example in one embodiment, the business service level agreements can be received by the cloud control layer 204 at operation 210 from a user interface of a computing device or system. In an embodiment, the business service level agreements can be read in at operation 210 from a file from the configuration management database. Note that operation 210 can be implemented in any manner similar to that described herein, but is not limited to such.
At operation 230 of
At operation 212, the cloud control layer 204 can determine whether each of the monitored statistics of the service level agreements is satisfying their corresponding defined threshold. If each threshold is satisfied, no action is taken at operation 212 by the cloud control layer 204. However, if it is determined at operation 212 that one or more thresholds are not satisfied, the cloud control layer 204 can proceed to operation 214. It is pointed out that operation 212 can be implemented in a wide variety of ways. For example, operation 212 can be implemented in any manner similar to that described herein, but is not limited to such.
At operation 214 of
At operation 216, the cloud control layer 204 can determine whether scaling the compute platform capacity will potentially eliminate the one or more bottlenecks determined at operation 214. If so, the cloud control layer 204 can proceed to operation 218. However, if it is determined at operation 216 that scaling the compute platform capacity will not eliminate the one or more bottlenecks, the cloud control layer 204 can proceed to operation 228. Note that operation 216 can be implemented in a wide variety of ways. For example, operation 216 can be implemented in any manner similar to that described herein, but is not limited to such.
At operation 208 of
At operation 228 of
At operation 232, after receiving the workload assignment information from the cloud control layer 204, the cloud instantiation layer 206 can distribute the workload across the prescribed deployment based on the workload assignment. Furthermore, as part of the workload distribution at operation 232, the cloud instantiation layer 206 can provision and configure script 238 in order to distribute the workload across the prescribed deployment based on the workload assignment. It is pointed out that operation 232 can be implemented in a wide variety of ways. For example, operation 232 can be implemented in any manner similar to that described herein, but is not limited to such.
At operation 222 of
At operation 224, the cloud control layer 204 can receive one or more compliance requirements for service providers or vendors. Note that operation 224 can be implemented in a wide variety of ways. For example in an embodiment, the one or more compliance requirements for service providers can be received by the cloud control layer 204 at operation 224 from a user interface of a computing device or system. It is pointed out that operation 224 can be implemented in any manner similar to that described herein, but is not limited to such.
At operation 226 of
At operation 236, the cloud instantiation layer 206 can acquire and transmit capability information for one or more service providers to the cloud control layer 204. It is pointed out that operation 236 can be implemented in a wide variety of ways. For example, operation 236 can be implemented in any manner similar to that described herein, but is not limited to such.
At operation 220 of
At operation 218, in order to scale the compute platform capacity to potentially eliminate the one or more bottlenecks determined at operation 214, the cloud control layer 204 can utilize the received one or more cost constraints and the allowable service providers to determine sizing and sourcing from each service provider or vendor. Additionally, the cloud control layer 204 can transmit or output at operation 218 sizing and sourcing information to the cloud instantiation layer 206. It is pointed out that operation 218 can be implemented in a wide variety of ways. For example in one embodiment, the sizing and sourcing determination at operation 218 by the cloud control layer 204 can include the assignment of roles (e.g., primary and redundant, or worker and aggregator), but is not limited to such. In addition, the output of the cloud control layer 204 at operation 218 can be a number and type of virtual machine appliances from each vendor for the cloud instantiation layer 206 to provision, but is not limited to such. Furthermore, the output of the cloud control layer 204 at operation 218 may prescribe a vendor provided platform that serves the basic functionality, or the details of building the platform, or a number of virtual machine appliances that augment or replicate this basic functionality. Operation 218 can be implemented in any manner similar to that described herein, but is not limited to such.
At operation 234 of
Below is described the formulation by the Cloud Control Layer 204 for scaling virtual appliances to meet performance and availability. For example in an embodiment, first a number is chosen of primary active nodes needed to meet performance constraints. Then a determination is made of the number of additional redundant nodes needed to guarantee availability. Taking these numbers along with a cost calculation that multiplies the number of resources and their cost; the result is a system of equations for the Cloud Control Layer 204 to determine the required resources.
In one embodiment, a determination is first made by the Cloud Control Layer 204 of the number of primary nodes needed to meet performance constraints. At the highest level, the Cloud Control Layer 204 can in an embodiment consider a virtual machine appliance or when horizontally scalable a group of virtual machine appliances each fulfilling different roles (e.g., load-balancer and worker node). Note that the functionality of each role is not interchangeable so the composite performance is based on the performance of each role. For example, the bottleneck can be the minimum capacity between the roles. Note that if there is a vendor packaged platform already in existence, it can be considered on its own by the Cloud Control Layer 204. For example, the Cloud Control Layer 204 can consider the maximum capacity it can receive from that vendor packaged platform. It is noted that the vendor capacity can then be augmented by other vendors or from a sourcing of virtual appliances.
In an embodiment, the focus can be on the determination of each role separately. In the case where the appliance does not horizontally scale, then the number of nodes can be fixed by the Cloud Control Layer 204. For example, if the load balancer appliance does not scale, there can be one active node.
In one embodiment, one case is horizontally scalable where appliances perform the same functionality in parallel. Generally this case is additive in capacity by increasing the number of primary nodes N, the capacity increases linearly with N. The performance constraint varies based on the type of workload: events, sessions, or batch items (which are considered below). The number of active primary worker nodes N can be scaled by the Cloud Control Layer 204 to meet performance constraints. For example, if a maximum or average capacity is associated per node, a scheme (or technique) scales N so that the sum of the maximum or average capacity of the N nodes is bounded by some number. Another scheme (or technique) by the Cloud Control Layer 204 is described below in which the probability that the capacity of N nodes not being able to fulfill demand.
Within
In one embodiment the cloud control layer 204 can let T be the time to process one event. Guarantees around T determine the performance of event-based systems. For example, find the smallest N so that the probability of exceeding delay threshold Tthreshold is bounded by P, in notation that is:
Pr(T≥Tthreshold)≤P.
This system can be modeled by the cloud control layer 204 as an M/M/N queue with arrival rate λ, where arrivals experience exponential service time μ, and are served by N similar and independent servers. The cloud Instantiation Layer 206 can provide estimates of these parameters by monitoring the average inter-arrival time of events (1/λ) and the average time for a unit to be processed (1/μ).
In addition within
Pr(T≥Tthreshold)=1−W(Tthreshold).
In one embodiment, the cloud control layer 204 can select N=min{n:W(Tthreshold)≥1−P}, the smallest non-negative integer N that meets the constraint.
It is noted that in an embodiment, session-based appliances load-balance worker nodes by sessions. Each appliance has a known capacity C number of sessions. In one embodiment, the cloud control layer 204 can provision N so that at any time the available system capacity NC is greater than the number of desired active sessions, N is chosen to bound the probability of session blocking to be less than P.
Within
Note that in an embodiment the cloud control layer 204 can suppose N is periodically chosen with period T>Tstart the start-up time of a new machine. At the start of the period time 0, X0=n. Then N can be updated so that for the hitting time:
T(NC)=min{t>0: Xt=NC}, Pr(T(NC)<T|X0=n)≤P.
In an embodiment, the cloud control layer 204 can assume in batch jobs that the items to be processed are already collected as part of a separate process. In addition, the worker nodes can process their items in parallel and present a summarized result for aggregation.
In one embodiment of
The cloud control layer 204 can assume in an embodiment the work is evenly partitioned so that with N worker nodes and a total workload W items that each node processes about k=W/N jobs. It is noted that for a given worker, the time to process k events is X=X1+ . . . +Xk for the cloud control layer 204 where the sequence Xi is the time to process items i is independent and identically (or similarly) distributed with mean d and variance σ2. Then for one worker the probability that it finishes in T is approximately:
where Q(.) is the cumulative distribution function (CDF) of the standard normal (Central Limit Theorem). Since each worker node has independent operations, the cloud control layer 204 can choose N={min k≥0: ρk≥P} so that the probability the batch job completes on time is large.
In one embodiment, given the primary nodes N, the number and sourcing of redundant nodes can be determined by the cloud control layer 204 that are needed to meet availability guarantees. Note that in an embodiment the cloud control layer 204 can be given the appliance and provider availability. As such, the cloud control layer 204 can let qi denote the availability of an appliance of type i. Furthermore, qi can be independent across providers. The cloud control layer 204 can let Pj denote the availability of provider j. Note that Pj can account for the availability of the network and of the provider so that appliances located in the same data center region are dependent and subject to the same availability (e.g., EC2 regions). It is noted that resources across regions can be independent. It is pointed out that like performance described above, the availability of a vendor package platform can be considered on its own by the cloud control layer 204. Or the vendor package platform can be considered by the cloud control layer 204 as a companion with other vendors or from the platform derived from a self-sourced collection of virtual appliances.
Within an embodiment, two cases can be considered by the cloud control layer 204. The first case deals with appliances serving the same function. Then the desired redundancy can be obtained by the cloud control layer 204 by limiting the maximum number of failures from a set of indistinct components. Otherwise, the second case deals with appliances that serve distinct functions and the composite availability is the product of the availability of each of the components. A nested formulation can be generated of these compositions by the cloud control layer 204 by conditioning on the availability of the providers.
More specifically, the application A 302 can be deployed with three appliances (e.g., 310, 312, and 314) on cloud 306 and two appliances (e.g., 316 and 318) in data center 308, and the application B 304 has two appliances (e.g., 320 and 322) on the cloud 306 and one appliance (e.g., 324) in the data center 308. It is noted that each of appliances 310, 312, 314, 316, and 318 is an appliance of type 1 that is active (which has an availability of q1). Moreover, each of appliances 320 and 324 is an appliance of type 2 that is stand-by (which has an availability of q2) while appliance 322 is an appliance of type 2 that is active (which has an availability of q2).
Within
F5,A(3)PCPD+F3,A(1)PC(1−PD)+F2,A(0)(1−PC)PD.
Similarly the availability of the application B 304 can be determined by the cloud control layer 204 using:
F3,B(2)PCPD+F2,B(2)PC(1−PD)+F1,B(0)(1−PC)PD.
Since the applications A 302 and B 304 are distinct, the availability of the Composite Application A and B can be determined by the cloud control layer 204 using:
F5,A(3)F3,B(2)PC PD+F3,A(1)F2,B(2)PC(1−PD)F2,A(0)F1,B(0)(1−PC)PD.
It is noted that the automated algorithms of the cloud control layer 204 can leverage cloud's scalability to create a platform that meets user-defined reliability objectives. Note that the separation of concerns allows the cloud control layer 204 to remain generic. Scaling algorithms may be reused by the cloud control layer 204 for applications and platforms based on, but not limited to, specified application requirements, published or estimated provider values, and broad categorizations along with handling of events, sessions, or batch.
In an embodiment, cloud can be on-demand, elastic, and pay-per-use. Note that in one embodiment, the cloud control layer 204 sources a combination of basic external and internal computing satisfying this cloud definition to meet custom service level agreements (SLAs). Therefore, for one-time provisioning and for real-time use, the cloud control layer 204 can leverage cloud's dynamics to address reliability concerns.
Specifically,
It is noted that the present embodiment illustrates that the cloud control layer 402 can be application generic since it operates in a similar manner while handling different cloud applications. For example, the storefront portal application 404, the order system application 406, and the analytics application 408 can each involve different performance criteria or requirements from the cloud control layer 402. For example in one embodiment, the storefront portal application 404 can include, but is not limited to, an access availability of 99.99%, a maximum latency of 300 milliseconds (ms), and an average latency of 150 ms, which can all be dependent on an order system. Furthermore, the order system application 406 can include, but is not limited to, an access availability of 99.999%, a maximum latency of 300 ms, and an average latency of 150 ms, herein a constraint is that things remain internal. In addition, the analytics application 408 can include, but is not limited to, a turn-around of less than 24 hours (hr) and a cost within $800, wherein a constraint is that things remain internal. Given the differences of the storefront portal application 404, the order system application 406, and the analytics application 408, it is pointed out that the cloud control layer 402 can operate in a similar manner while handling these different cloud applications.
Within
More specifically, within the present embodiment the one or more completion metrics 504 can include, but are not limited to, a turnaround time in 4 hours and a less than 1% error with 95% confidence. In addition, the availability metric 506 can include, but is not limited to, a 99.999% uptime while the performance metric 508 can include, but is not limited to, user requests responded to in less than 350 ms. Furthermore, the one or more constraints 510 can include, but is not limited to, records remain on premise and costs within $800 per day.
Within
In one embodiment, it is pointed out that the mechanisms or functions 514 of the cloud control layer 512 can be based on the received cloud variables 516 and the received demand variables 518. As such, the received cloud variables 516 and the received demand variables 518 can be taken into consideration by the mechanisms or functions 514 of the cloud control layer 512. The one or more cloud variables 516 can include, but are not limited to, machine and provider elements or variables such as, but not limited to, failure, capacity, throughput, time to start, cost, and capability. Additionally, the one or more demand variables 518 can include, but are not limited to, active sessions, current processing time, and job sizes.
Within
More specifically, within the present embodiment the what to buy 522 can include, but are not limited to, buying five Amazon Web Servers plus two data centers. Additionally, the when to buy 524 can include, but is not limited to, once the threshold is at 80%, start two additional Amazon Web Servers. Moreover, the what to do with it 526 can include, but is not limited to, bad balance across five primary nodes and duplicate two. It is understood that the what to buy 522, the when to buy 524, and the what to do with it 526 can be implemented with any number and/or values of elements and are not limited in any way to those shown in the present embodiment.
In one embodiment, the user interface 600 can be implemented to include multiple tabs. For example within the present embodiment, the user interface 600 can include a “Specify Traffic” tab 602, a “Specify Deployment” tab 604, a “View Traffic SLA Status” tab 606, a “View Deployment SLA Status” tab 608, and an “Update Settings” tab 610, but is not limited such. Note that the “Specify Traffic” tab 602 is shown selected within the user interface 600. In addition, the example contents of the “Specify Traffic” tab 602 can include, but is not limited to, a “TrafficSLAs” sub-tab 612 and a “Traffic-Application” sub-tab 624. It is noted that the “TrafficSLAs” sub-tab 612 is shown selected and its example contents can include, but is not limited to, a “New TrafficSLA” activation button 614, an “Update TrafficSLA” activation button 616, and a “Delete TrafficSLA” activation button 618.
Within
Note that the user interface 600 may not include all of the elements illustrated by
Within the present embodiment, the example contents of the Traffic-Application sub-tab 624 can include, but is not limited to, a role identifier drop-down menu selector 702, a “Current Application List” display area 704, an “Available Application List” display area 706, an “Add” activation button 708, and a “Remove” activation button 710. It is pointed out that the role identifier drop-down menu selector 702 can be utilized by a user to assign or associate one or more selected roles to a particular application. For instance within the present embodiment, the Web application listed within the Current Application List display area 704 currently has a web role as indicated by the role identifier drop-down menu selector 702.
Within
Within the present embodiment, the example contents of the Specify Deployment tab 604 can include, but is not limited to, a “Get Optimal Deployment” activation button 806, an “Update Deployment” activation button 808, an “Optimize All Deployment” activation button 810, a “Primary Deployment” information fields 802, a “Backup Deployment” information fields 804, and a “Choose Deployment to update” selection area 812. Note that the “Get Optimal Deployment” activation button 806 can also be referred to as the “Get Best Possible Deployment” activation button 806 or the “Get Appropriate Deployment” activation button 806, but is not limited to such. Furthermore, the “Optimize All Deployment” activation button 810 can also be referred to as the “Enhance Effectiveness of All Deployment” activation button 810 or “Get Appropriate Deployment for All” activation button 810, but is not limited to such. It is pointed out that when the Update Deployment activation button 808 is selected by a user, the user interface 600 displays the “Choose Deployment to update” selection area 812 along with the Primary Deployment information fields 802 and Backup Deployment information fields 804 that correspond to the highlighted application within the “Choose Deployment to update” selection area 812.
Within
It is pointed out that if the Optimize All Deployment activation button 810 is selected by a user, the user interface 600 applies the optimum (e.g., minimum cost) or best possible or appropriate deployment that satisfies the service level agreements. Additionally, if the Get Optimal Deployment activation button 806 is selected by a user, the user interface 600 determines and highlights the application listed within the “Choose Deployment to update” selection area 812 that currently has the optimal (e.g., minimum cost) or best possible or appropriate deployment.
Within the present embodiment, the example contents of the “View Traffic SLA Status” tab 606 can include, but is not limited to, an application identifier drop-down menu selector 902, a Cost graph 904, a Capacity graph 906, a Latency graph 908, and an Availability graph 910. It is noted that the application identifier drop-down menu selector 902 can be utilized by a user to select which composite application service level agreements to view. For example within the present embodiment, the “Web” application is currently selected within the application identifier drop-down menu selector 902. As such, the Cost graph 904, Capacity graph 906, Latency graph 908, and Availability graph 910 display the composite service level agreements for the Web application.
Specifically within
Within the present embodiment, the example contents of the “View Deployment SLA Status” tab 608 can include, but is not limited to, an application identifier drop-down menu selector 1002, a service level agreement attribute drop-down menu selector 1004, a first role graph 1006, and a second role graph 1008. Note that the application identifier drop-down menu selector 1002 can be utilized by a user to select which application service level agreements to view. In addition, the service level agreement attribute drop-down menu selector 1004 can be utilized by the user to select which service level agreement attribute to view of the selected application.
Within the present embodiment of
Specifically within the present embodiment, the contents of the “Update Settings” tab 610 includes, but is not limited to, a “Providers” sub-tab 1102, an “Application” sub-tab 1112, and a “Reset Simulation” sub-tab 1114. It is noted that the “Providers” sub-tab 1102 is shown selected thereby enabling a user to update and view one or more settings associated with one or more cloud providers. The example contents of the “Providers” sub-tab 1102 can include, but is not limited to, a “Clear Provider” activation button 1104, an “Update Provider” activation button 1106, a “Choose Provider to update” selection area 1110, and Cloud Provider information fields 1108.
Within
Within the present embodiment, the example contents of the “Application” sub-tab 1112 can include, but is not limited to, a “New Application” activation button 1202, an “Update Application” activation button 1206, and a “Delete Application” activation button 1208. Note that when a user selects the “Update Application” activation button 1206, a “Choose Application to update” selection area 1210 can be displayed thereby enabling the user to select the application settings he or she desires to update and/or view. For example, within the present embodiment of the “Choose Application to update” selection area 1210, the user can to update the settings of the “Web” application, the “DB” application, a “Batch” application, or a “LoadBalance” application. Once an application is selected within the “Choose Application to update” selection area 1210, one or more application provider capacity fields 1204 can be displayed thereby enabling the user to view and/or update one or more of them. R is noted that the capacity of a provider may be updated utilizing either estimated values from actual measurements of the provider or from published (or known) values, but is not limited to such.
Specifically within the present embodiment of
Note that the user interface 600 may not include all of the elements illustrated by
Computer system 1300 can include an address/data bus 1310 for communicating information, one or more central processors 1302 coupled with bus 1310 for processing information and instructions. Central processor unit(s) 1302 may be a microprocessor or any other type of processor. The computer 1300 can also include data storage features such as computer usable volatile memory 1304, e.g., random access memory (RAM), static RAM, dynamic RAM, etc., coupled with bus 1310 for storing information and instructions for central processor(s) 1302, computer usable non-volatile memory 1306, e.g., read only memory (ROM), programmable ROM, flash memory, erasable programmable read only memory (EPROM), electrically erasable programmable read only memory (EEPROM), etc., coupled with bus 1310 for storing static information and instructions for processors) 1302.
System 1300 of
Optionally, computer system 1300 can include an alphanumeric input device 1314 including alphanumeric and function keys coupled to the bus 1310 for communicating information and command selections to the central processor(s) 1302. The computer 1300 can also include an optional cursor control or cursor directing device 1316 coupled to the bus 1310 for communicating user input information and command selections to the processor(s) 1302. The cursor directing device 1316 can be implemented using a number of well-known devices such as, but not limited to, a mouse, a track ball, a track pad, an optical tracking device, a touch screen, etc. Alternatively, it is appreciated that a cursor can be directed and/or activated via input from the alphanumeric input device 1314 using special keys and key sequence commands. The present embodiment is also well suited to directing a cursor by other means such as, for example, voice commands.
Within
Note that the volatile memory 1304 may store a cloud management system or module 1320 in accordance with various embodiments of the disclosure. The cloud management system or module 1320 may include instructions to cause the system 1300 to operate or function in any manner similar to that described herein, but not limited to such. It is pointed out that in various embodiments, the cloud management system or module 1320 (or one or more of its components) may be stored by the volatile memory 1304, or the non-volatile memory 1306, or the mass data storage device 1318, or any combination thereof. In various embodiments, the cloud management system or module 1320 can be an implementation of the cloud management system or module 100 of
Within
Note that the system 1300 may not include all of the elements illustrated by
The foregoing descriptions of various specific embodiments in accordance with the disclosure have been presented for purposes of illustration and description. They are not intended to be exhaustive or to limit the disclosure to the precise forms disclosed, and many modifications and variations are possible in light of the above teaching.
Tung, Teresa S., Tobolski, Joseph F., Swaminathan, Kishore S.
Patent | Priority | Assignee | Title |
11012374, | May 20 2019 | Citrix Systems, Inc. | Systems and methods for virtual session connection using component-based connection leases |
11386265, | Dec 15 2020 | International Business Machines Corporation | Facilitating information technology solution templates |
11483255, | May 20 2019 | Citrix Systems, Inc. | Systems and methods for virtual session connection using component-based connection leases |
11645595, | Dec 15 2020 | International Business Machines Corporation | Predictive capacity optimizer |
11809907, | Aug 11 2016 | RESCALE, INC. | Integrated multi-provider compute platform |
Patent | Priority | Assignee | Title |
6725378, | Apr 15 1998 | Purdue Research Foundation | Network protection for denial of service attacks |
6925642, | Apr 29 1999 | Hewlett Packard Enterprise Development LP | Distributed computer network which spawns inter-node parallel processes based on resource availability |
6963828, | Mar 21 2001 | Unisys Corporation | Metafarm sizer configuration optimization method for thin client sizing tool |
7249176, | Apr 30 2001 | Oracle America, Inc | Managing user access of distributed resources on application servers |
7376693, | Feb 08 2002 | JPMORGAN CHASE BANK, N A | System architecture for distributed computing and method of using the system |
7668703, | Jun 07 2005 | Hewlett Packard Enterprise Development LP | Determining required capacity for a resource |
7747750, | Jun 27 2007 | EMC IP HOLDING COMPANY LLC | Method for reserving resources in a storage area network with selective capabilities |
8341270, | Jan 24 2006 | Citrix Systems, Inc | Methods and systems for providing access to a computing environment |
8484355, | May 20 2008 | ATLASSIAN US, INC | System and method for customer provisioning in a utility computing platform |
8886806, | Apr 07 2010 | Accenture Global Services Limited | Generic control layer in a cloud environment |
9215190, | Apr 07 2010 | Accenture Global Services Limited | Generic control layer in a cloud environment |
20030200300, | |||
20040006589, | |||
20040111506, | |||
20050044220, | |||
20050080838, | |||
20060153090, | |||
20070043860, | |||
20070300297, | |||
20080091806, | |||
20080114624, | |||
20080163194, | |||
20080244233, | |||
20080244579, | |||
20080250267, | |||
20080320482, | |||
20090083390, | |||
20090300210, | |||
20090327495, | |||
20100042720, | |||
20100217864, | |||
20100306377, | |||
20100332617, | |||
20110016214, | |||
20110087776, | |||
20110087783, | |||
20110119381, | |||
20110131306, | |||
20110131329, | |||
20110166952, | |||
20110191838, | |||
20110202925, | |||
20110213875, | |||
20110296021, | |||
20120030356, | |||
20120124211, | |||
20140172782, | |||
WO2010035281, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Apr 14 2011 | TUNG, TERESA S | Accenture Global Services Limited | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 037309 | /0237 | |
Apr 14 2011 | TOBOLSKI, JOSEPH F | Accenture Global Services Limited | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 037309 | /0237 | |
Apr 16 2011 | SWAMINATHAN, KISHORE S | Accenture Global Services Limited | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 037309 | /0237 | |
Nov 11 2015 | Accenture Global Services Limited | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Feb 16 2022 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Date | Maintenance Schedule |
Sep 04 2021 | 4 years fee payment window open |
Mar 04 2022 | 6 months grace period start (w surcharge) |
Sep 04 2022 | patent expiry (for year 4) |
Sep 04 2024 | 2 years to revive unintentionally abandoned end. (for year 4) |
Sep 04 2025 | 8 years fee payment window open |
Mar 04 2026 | 6 months grace period start (w surcharge) |
Sep 04 2026 | patent expiry (for year 8) |
Sep 04 2028 | 2 years to revive unintentionally abandoned end. (for year 8) |
Sep 04 2029 | 12 years fee payment window open |
Mar 04 2030 | 6 months grace period start (w surcharge) |
Sep 04 2030 | patent expiry (for year 12) |
Sep 04 2032 | 2 years to revive unintentionally abandoned end. (for year 12) |