In certain embodiments, a computer-implemented method includes receiving a request for a reliability estimate associated with an ad hoc application. In response to the request, one or more components associated with the ad hoc application and upon which the ad hoc application relies are identified. The method also includes generating a directed graph. The directed graph identifies one or more dependency relationships among the identified components. The method also includes calculating, based at least in part on the directed graph, a reliability estimate for the ad hoc application.
|
11. A computer-implemented method, comprising:
identifying one or more components associated with an ad hoc application and upon which the ad hoc application relies, wherein identifying one or more components comprises:
obtaining an application definition associated with the ad hoc application, the application definition received from a user and comprising a first group of components, the first group of components including the ad hoc application and one or more components relied upon by the ad hoc application;
obtaining tag information, the tag information indicating one or aspects of the first group of components; and
obtaining, based at least in part on the tag information and the application data, infrastructure data from an infrastructure repository, the infrastructure data identifying a second group of components, the second group of components comprising one or more computing resources of a distributed architecture associated with the ad hoc application;
generating a directed graph, the directed graph comprising at least a subset of components of the first group of components and the second group of components and indicating one or more dependency relationships among the subset of components;
generating a plurality of conditional probability tables, based at least in part on the subset of components in the directed graph, wherein the plurality of conditional probability tables are based at least in part on one or more of the dependency relationships identified in the directed graph and indicates availability of a respective component based at least in part on availability of at least one relied upon component; and
calculating, based at least in part on the directed graph, the reliability estimate for the ad hoc application, wherein calculating the reliability estimate for the ad hoc application comprises performing a plurality of trials, wherein performing the plurality of trials comprises, assigning a status of either available or not available to at least a portion of the components in the directed graph, the assigned status based at least in part on a particular conditional probability table associated with a particular component of the portion of the components in the directed graph and the assigned status of one or more directly relied upon components.
1. A system comprising:
one or more memory units with executable instructions; and
one or more processing units that, when executing the instructions in the one or more memory units, are operable to:
receive an application definition associated with an ad hoc application provisioned from one or more computing resources delivered over a network, the application definition identifying a first group of components, the first group of components comprising the ad hoc application and one or more computing resources relied on by the ad hoc application;
receive tag information from a user, the tag information indicating one or more aspects of the first group of components;
access infrastructure data from an infrastructure repository, the infrastructure data identifying a second group of components, the second group of components comprising one or more computing resources of a distributed architecture that are associated with at least a subset of one or more components in the first group of components;
generate a plurality of conditional probability tables, one conditional probability table for at least a first subset of the components in the first group of components and at least a second subset of the second group of components, the plurality of conditional probability tables identifying at least an availability of a respective component of at least the first subset of the first group of components or at least the second subset of the second group of components based at least in part on a second availability of one or more relied upon components of the first group of components or the second group of component, where the one or more relied upon components are components utilized, at least in part, during operation of the respective component; and
based at least in part on the plurality of conditional probability tables, calculate a reliability estimate for the ad hoc application by at least performing a plurality of trials, wherein performing the plurality of trials comprises assigning a status of either available or not available to at least a portion of the components in a directed graph, the status based at least in part on a particular conditional probability table associated with a particular component and the status of one or more directly relied upon components.
19. A non-transitory computer-readable medium comprising logic, the logic when executed by one or more processing units operable to perform operations comprising:
receiving, from a user, a request for a reliability estimate associated with an ad hoc application;
in response to the request, identifying one or more components associated with the ad hoc application and upon which the ad hoc application relies;
accessing infrastructure data from an infrastructure repository, the infrastructure data identifying a second group of components, the second group of components comprising one or more computing resources of a distributed architecture associated with the ad hoc application;
generating a directed graph, the directed graph comprising one or more identified components and indicating one or more dependency relationships among the one or more identified components;
accessing reliability metrics data, the reliability metrics data comprising at least one of user reliability data, instance reliability data, and class reliability data, the reliability metrics data associated with one or more identified components in the directed graph;
generating a plurality of conditional probability tables, at least one conditional probability table for at least a subset of the one or more identified components in the directed graph, wherein at least a portion of the plurality of conditional probability tables are based at least in part on the one or more of the dependency relationships identified in the directed graph and indicating availability of a respective component based at least in part on availability of at least one relied upon component;
calculating, based at least in part on the directed graph and the reliability metrics data, the reliability estimate for the ad hoc application, wherein calculating the reliability estimate for the ad hoc application comprises performing a plurality of trials, wherein performing the plurality of trials comprises, assigning a status of either available or not available to at least a subset of the one or more identified components in the directed graph, the assigned status being based at least in part on a particular conditional probability table of the plurality of conditional probability tables associated with a particular component and the assigned status of one or more directly relied upon components; and
transmitting the reliability estimate to the user.
2. The system of
based at least in part on the application definition, the tag information, and the infrastructure data, generate the directed graph, the directed graph comprising the components from the first group of components and second group of components and indicating one or more dependency relationships among the components; and
wherein at least a portion of the plurality of conditional probability tables table is are associated with at least one of the one or more components in the directed graph, and the one or more processing units are further operable to calculate the reliability estimate based at least in part on the plurality of conditional probability tables and the directed graph.
3. The system of
user reliability data, the user reliability data comprising historical availability data of the ad hoc application determined by one or more users of the ad hoc application;
instance reliability data, the instance reliability data comprising historical availability of a particular component associated with the ad hoc application;
class reliability data, the class reliability data comprising historical availability data associated with a plurality of types of components associated with the ad hoc application; and
wherein the processing units are operable to generate the conditional probability table for at least the first subset of the first group of components and at least the second subset of the second group of components based at least in part on the reliability metrics data.
4. The system of
for at least a third subset of components in the first group of components and the second group of components, determining the one or more relied upon components;
generating one or more rows in the conditional probability table, the one or more rows comprising a subset of permutations, the subset of permutations indicating availability of at least a portion of the one or more relied upon components by the third subset of components; and
for the one or more rows, determining the availability of the respective component based at least in part on the subset of permutations.
5. The system of
after performing the plurality of trials, calculating a first number of times the ad hoc application is assigned a status of available;
after performing the plurality of trials, calculating a second number of times the ad hoc application is assigned a status of not available; and
comparing the first number of times the ad hoc application is assigned a status of available to the second number of times the ad hoc application is assigned a status of not available.
6. The system of
7. The system of
sampling a plurality of availability configurations from a set of all availability configurations, the sampled availability configurations based at least in part on the directed graph and indicating a status of available or not available to at least a portion of the components in the directed graph; and
for at least a subset of the sampled availability configuration, determining the probability of the availability configuration based at least in part on a particular conditional probability table associated with at least a subset of the components in the directed graph.
8. The system of
9. The system of
10. The system of
12. The method of
Generating generating the directed graph based at least in part on the application definition, the tag information, the infrastructure data and the conditional probability tables.
13. The method of
obtaining reliability metrics data associated with one or more components in the directed graph, the reliability metrics data comprising at least one of:
user reliability data, the user reliability data comprising historical availability data of the ad hoc application determined by one or more users of the ad hoc application;
instance reliability data, the instance reliability data comprising historical availability of components associated with the ad hoc application; and
class reliability data, the class reliability data comprising historical availability data associated with one or more types of components associated with the ad hoc application.
14. The method of
determining whether a particular component of a portion of components directly relied upon by a respective component is assigned available or not available status; and
if the assigned status of the particular component is available, assigning the status to the respective component based at least in part on the reliability metrics data.
15. The method of
after performing the plurality of trials, calculating a first number of times the ad hoc application is assigned the status of available;
after performing the plurality of trials, calculating a second number of times the ad hoc application is assigned the status of not available; and
comparing the first number of times the ad hoc application is assigned the status of available to the second number of times the ad hoc application is assigned the status of not available.
16. The method of
sampling a plurality of availability configurations from a set of availability configurations, at least a subset of the availability configurations based at least in part on the directed graph and indicating the status of available or not available for at least the portion of the components in the directed graph; and
for at least a subset of the sampled availability configurations, determining the probability of a particular availability configuration based at least in part on the conditional probability table associated with a portion of the components in the directed graph.
17. The method of
18. The method of
20. The non-transitory computer-readable medium of
accessing an application definition associated with the ad hoc application, the application definition received from the user and comprising a first group of components, the first group of components including the ad hoc application and one or more components relied upon by the ad hoc application;
accessing tag information, the tag information indicating one or more aspects of the first group of components; and
accessing, based at least in part on the tag information and application data, infrastructure data from an infrastructure repository, the infrastructure data identifying the second group of components, the second group further comprising one or more components of the ad hoc application.
21. The non-transitory computer-readable medium of
generating the directed graph based at least in part on the application definition, the tag information, the infrastructure data and the conditional probability tables.
22. The non-transitory computer-readable medium of
the user reliability data comprises historical availability data of the ad hoc application determined by one or more other users of the ad hoc application;
the instance reliability data comprises historical availability of a particular component associated with the ad hoc application; and
the class reliability data comprises historical availability data associated with one or more types of components associated with the ad hoc application.
23. The non-transitory computer-readable medium of
determining whether the subset of the one or more identified components directly relied upon by the respective component is assigned available or not available status; and
if the status of the subset of the one or more identified components is available, assigning the status to the respective component based at least in part on the reliability metrics data.
24. The non-transitory computer-readable medium of
after performing the plurality of trials, calculating a first number of times the ad hoc application is assigned the status of available;
after performing the plurality of trials, calculating a second number of times the ad hoc application is assigned the status of not available; and
comparing the first number of times the ad hoc application is assigned the status of available to the second number of times the ad hoc application is assigned the status of not available.
25. The non-transitory computer-readable medium of
26. The non-transitory computer-readable medium of
sampling a plurality of availability configurations from a set of availability configurations, the set of the availability configurations based at least in part on the directed graph and indicating the status of available or the status of not available for at least a subset of the one or more identified components in the directed graph; and
for at least a portion of the sampled availability configuration, determining a probability of the availability configuration based at least in part on a particular conditional probability table associated with for at least a subset of the one or more identified components in the directed graph.
27. The non-transitory computer-readable medium of
28. The non-transitory computer-readable medium of
29. The non-transitory computer-readable medium of
0. 30. The system of claim 2, wherein the tag information furthers indicates at least one dependency relationship between at least one of the components from the first group of components and at least one of the components from the second group of components.
0. 31. The system of claim 2, wherein the tag information furthers indicates at least one component from the first or the second group of components to exclude from the directed graph.
0. 32. The method of claim 12, wherein the tag information furthers indicates at least one dependency relationship between at least one of the components from the first group of components and at least one of the components from the second group of components.
0. 33. The method of claim 12, wherein the tag information furthers indicates at least one component from the first or the second group of components to exclude from the directed graph.
|
This application, U.S. patent application Ser. No. 15/466,626, along with U.S. patent application Ser. No. 15/449,814 filed on Mar. 3, 2017, now U.S. Pat. No. RE47,593, are both reissue applications of U.S. patent application Ser. No. 13/223,972, filed Sep. 1, 2011, now U.S. Pat. No. 8,972,564, entitled “RELIABILITY ESTIMATOR FOR AD HOC APPLICATIONS.”
Reliability is an important business property. Reliability, however, can be difficult to measure in a distributed system comprising many disparate components with differing levels of availability and redundancy. This is particularly true when portions of the service infrastructure are purchased from another company, which may not reveal details of its internal infrastructure. Formal models, end-to-end system descriptions, and simple, uncorrelated modes of failure may be inadequate in more complicated systems in which internal components are obscured from a user.
For a more complete understanding of the present disclosure and its advantages, reference is made to the following descriptions, taken in conjunction with the accompanying drawings, in which:
Cloud providers deliver a set of services that can be used to construct applications in a reliable, scalable, and inexpensive manner. These benefits, however, should be obtained by using the services in a careful manner. While some properties such as cost are relatively easy to measure, other properties such as reliability are not. Past solutions to measure reliability of interconnected components, systems, and/or applications have included manual efforts to calculate application reliability from fault trees, reliability block diagrams, and other modeling approaches. These calculations often require access to exact network schematics and aggregate reliability data, which may be highly confidential and proprietary business information. Alternatively, efforts to calculate application reliability have treated these factors as black boxes, which limits the estimation of reliability to coarse-grain measures. These methods often incorrectly assume that all failures are independent and that the rate at which failures occurs is constant.
Particular embodiments of the present disclosure address these and other limitations of previous systems by incorporating user input of an application definition and relationships between computing resources to determine an infrastructure and application configuration. Based on historical availability of the infrastructure and application resources, conditional probability tables are generated that indicate the availability of infrastructure and application components under various circumstances (such, as, e.g., whether directly relied upon components are available or not available). A reliability estimate is generated by running a large number of successive trials in which the availability or non availability of an infrastructure or application component is determined in accordance with the statistical probabilities indicated in the generated conditional probability tables. Thus, the reliability estimate may then be estimated based on the aggregate number of times an ad hoc application is determined to be available or not available in the total number of trials. The reliability estimate may be transmitted to a user of an ad hoc application.
In general, portions of system 100 provide an environment in which one or more computing resources (e.g., computing resources 110) is made available over a communication network (e.g., network 104) to one or more remote computer systems, such as user system 102. In certain embodiments, server system 106, storage module 108, and computing resources 110 may be communicatively coupled together over a high speed communication network and collectively may comprise a computing infrastructure, which may be referred to as a provisioned computing resources environment 112. User system 102 and/or network 104 may be external to provisioned computing resources environment 112 and may be referred to as an external computing environment 114.
In certain embodiments, provisioned computing resources environment 112 (including, for example, one or more of server system 106, storage module 108, and computing resources 110) may provide a collection of remote computing services offered over a network (which may or may not be network 104). Those computing services may include, for example, storage, computer processing, networking, applications, or any other suitable computing resources that may be made available over a network. In some embodiments, computing resources may be referred to as ad hoc applications, which may be provisioned or de-provisioned according to the requirements and/or configuration of external computing environment 114. In certain embodiments, entities accessing those computing services may gain access to a suite of elastic information technology (IT) infrastructure services (e.g., computing resources 110) as the entity requests those services. Provisioned computing resources environment 112 may provide a scalable, reliable, and secure distributed computing infrastructure.
In association with making those computing resources 110 available over the network (e.g., provisioning the computing resources 110), a variety of reliability parameters may be generated. These reliability parameters may indicate or represent the availability or non-availability of a particular provisioned ad hoc application (or its underlying infrastructure or application components) to user system 102 or external computing environment 114. Reliability parameters may be referred to as reliability metrics data. Server 106 uses reliability metrics data to determine a reliability estimate for one or more ad hoc applications. Reliability metrics data may be associated with a particular component, system, software, application, interface, and/or network included in provisioned computing resources environment 112. Particular examples of reliability metrics data may include user reliability data 124, instance reliability data 126, and class reliability data 128, discussed further below.
Portions of system 100 may determine reliability metrics data associated with components of system 100 (e.g., computing resources 110). It may be appropriate to communicate a portion or all of this reliability metrics data over a network (e.g., network 104) to a server so that the server (e.g., server system 106) may use the communicated reliability metrics data. For example, reliability metrics data may be communicated over a network (e.g., network 104) to a server (e.g., server system 106), so that server system 106 may calculate reliability estimate 134 for one or more ad hoc applications. A particular reliability estimate 134 may be communicated over network 104 to user system 102 in response to a query for reliability data associated with a particular ad hoc application.
User system 102 may include one or more computer systems at one or more locations. Each computer system may include any appropriate input devices, output devices, mass storage media, processors, memory, or other suitable components for receiving, processing, storing, and communicating data. For example, each computer system may include a personal computer, workstation, network computer, kiosk, wireless data port, personal data assistant (PDA), one or more Internet Protocol (IP) telephones, smart phones, table computers, one or more servers, a server pool, one or more processors within these or other devices, or any other suitable processing device. User system 102 may be a stand-alone computer or may be a part of a larger network of computers associated with an entity.
User system 102 may include processing unit 116 and memory unit 118. Processing unit 116 may include one or more microprocessors, controllers, or any other suitable computing devices or resources. Processing unit 116 may work, either alone or with other components of system 100, to provide a portion or all of the functionality of system 100 described herein. Memory unit 118 may take the form of volatile or non-volatile memory including, without limitation, magnetic media, optical media, RAM, ROM, removable media, or any other suitable memory component.
In general, user system 102 communicates tag information 138 and application definition 140 to server system 106 to facilitate reliability estimation for an ad hoc application. First, user system 102 may interact with component tagging module 142 to apply one or more metadata tags (e.g., tag information 138) to computing resources 110. A metadata tag may be a short, textual string that describes one or more aspects of the relevant computing resource 110. For example, if user system 102 is provisioned with an ad hoc application (e.g., an accounting software package) that runs on two processing computing resources 110 and one database computing resource 110, the user may tag each of the computing resources 110 with the string ‘accounting’ to associate the computing resources with the provisioned ad hoc application. Tag information 138 may also describe configuration relations. For example, tag information 138 may link resources with resource addresses, access control policies, firewall rules, or connection strings. In general, tag information 138 includes metadata information that associates a particular computing resource 110 with an ad hoc application provided to user system 102.
Second, user system 102 may interact with application definition module 144 to create an application definition (e.g., application definition 140) of a provisioned ad hoc application. Application definition 140 includes at least a primary computing resource 110 for which reliability estimate 134 is to be calculated. Application definition 140 may include one or more secondary computing resources 110 that are supportive of the primary computing resource 110. For example, the primary computing resource 110 may be software service while a secondary computing resource 110 may be a web service accessed by the software service. In some embodiments, application definition 140 may not define all secondary computing resources 110 used by a particular ad hoc application. Graph inference module 146 may expand the user-provided ad hoc application definition 140 into a more comprehensive application definition. In some embodiments, application definition module 144 defines the starting seeds for graph inference module 146. Graph inference module 146 is discussed in greater detail below with respect to
A user of user system 102 may include, for example, a person capable of requesting and receiving a reliability estimate for an ad hoc application. As a more particular example, a user of system 102 may be associated with an entity using the computing resources (e.g., computing resources 110) made available over a network.
Network 104 facilitates wireless or wireline communication. Network 104 may communicate, for example, IP packets, Frame Relay frames, Asynchronous Transfer Mode (ATM) cells, voice, video, data, and other suitable information between network addresses. Network 104 may include one or more local area networks (LANs), radio access networks (RANs), metropolitan area networks (MANs), wide area networks (WANs), mobile networks (e.g., using WiMax (802.16), WiFi (802.11), 3G, or any other suitable wireless technologies in any suitable combination), all or a portion of the global computer network known as the Internet, and/or any other communication system or systems at one or more locations, any of which may be any suitable combination of wireless and wireline.
Server system 106 may include one or more computer systems at one or more locations. Each computer system may include any appropriate input devices, output devices, mass storage media, processors, memory, or other suitable components for receiving, processing, storing, and communicating data. For example, each computer system may include a personal computer, workstation, network computer, kiosk, wireless data port, PDA, one or more IP telephones, one or more servers, a server pool, one or more processors within these or other devices, or any other suitable processing device. Server system 106 may be a stand-alone computer or may be a part of a larger network of computers associated with an entity.
Server system 106 may include processing unit 122 and memory unit 124. Processing unit 122 may include one or more microprocessors, controllers, or any other suitable computing devices or resources. Processing unit 122 may work, either alone or with other components of system 100, to provide a portion or all of the functionality of system 100 described herein. Memory unit 124 may take the form of volatile or non-volatile memory including, without limitation, magnetic media, optical media, RAM, ROM, removable media, or any other suitable memory component.
Server system 106 may calculate reliability estimate 134 for one or more ad hoc applications. In particular, server system 106 calculates reliability estimate 134 based on data received from or determined in conjunction with other components of system 100. In particular, server system 106 may calculate reliability estimate 134 based on one or more of reliability data 124, instance reliability data 126, class reliability 128, infrastructure repository 136, user tag information 138, and application definition 140. As described further below, server system 106 may process this received data using one or more of component tagging module 142, application definition module 144, graph inference module 146, reliability estimator module 148, application probability module 150, and infrastructure probability module 152 to calculate reliability estimate 134.
User reliability data 124 represents historical availability or non-availability of a computing resource 110 (e.g., an ad hoc application) as determined by direct observation of user system 102. For example, user system 102 may periodically perform a health check on an ad-hoc application to determine whether the ad hoc application is operational. User system 102 may communicate the results of the health check to server system 106, which may store the results as user reliability data 124.
Instance reliability data 126 represents the historical availability or non-availability of a particular computing resource 110 (e.g., a server, disk drive, network interface, power supply, etc.). For example, one or more components of system 100 (e.g., server system 106) may periodically perform a health check of infrastructure components to determine their respective availability or non-availability to user system 102. Server system 106 stores the results of the health check as instance reliability data 126.
Class reliability data 128 represents the historical availability or non-availability of a particular class of computing resources 110. For example, one or more components of system 100 (e.g., server system 106) may periodically perform a health check of one or more similar computing resources 110 to determine the availability or non-availability as a class. Class reliability data 128 may bias towards components with measures of similarity; such as hardware revision, order date, time in service, installation location, or maintenance record. In some embodiments, class reliability data 128 may be used as a proxy for instance reliability data 126 if or when instance reliability data 126 is unavailable for a particular computing resource 110.
Infrastructure repository 136 stores information related to computing resources 110. For example, server system 106 may store a hardware type, hardware parameters (e.g., processor speed, storage space, etc.), hardware revision, order date, time in service, installation location, or maintenance record for each computing resource 110 in infrastructure repository 136. Infrastructure repository 136 may additionally store details regarding the connections between computing resources 110, such as network links, network speeds, network availability, and/or connection type. In some embodiments, server system 106 may store information related to computing resources 110 in a database on storage module 108.
Component tagging module 142 receives tag information 138 from user system 102 and stores tag information 138 in storage module 108. As discussed above, tag information 138 indicates relationships between computing resources 110 and an ad hoc application.
Application definition module 144 receives application definition 140 from user system 102 and stores application definition 140 in storage module 108. Application definition 140 identifies one or more component computing resources 110 for an ad hoc application.
Graph inference module 146 constructs a directed graph (e.g., directed graph 500 illustrated in
Application probability calculator 148 constructs conditional probability tables 600 for the directed graph based on user reliability data 124 and instance reliability data 126 for components of an ad hoc application. Example conditional probability tables 600 generated by application probability calculator 148 are discussed further below with respect to
Infrastructure probability calculator 150 constructs conditional probability tables 600 for the directed causality graph based on instance reliability data 126 and class reliability data 128 for infrastructure components relied on by an ad hoc application. For an infrastructure component, the infrastructure probability calculator 150 may access databases for instance reliability data and class reliability data to construct a historical availability record for the component. Example conditional probability tables 600 generated by infrastructure probability calculator 150 are discussed further below with respect to
Reliability estimator module 152 calculates reliability estimate 134 based on conditional probability tables 600 and directed graph 500. Reliability estimator module 152 may evaluate the inferred directed graph and constructed conditional probability tables 600 as a Bayesian network to produce reliability estimate 134. Exact computation of the reliability of the primary resource may be possible for simple directed graphs, such as graphs with only a single path to any component. In many cases, however, the inferred directed graph may not have a direct solution. In some embodiments, reliability estimator module 152 supports stochastic simulation of the inferred directed graph to compute the reliability of the primary resource. For example, reliability estimator module 152 may run a number of trials sampling different availability configurations according to the conditional probabilities for availability of each component in the directed graph.
Computing resources 110 may include any suitable computing resources that may be made available over a network (which may or may not be network 104). Computing resources 110 may include any suitable combination of hardware, firmware, and software. As just a few examples, computing resources 110 may include any suitable combination of applications, power, processors, storage, and any other suitable computing resources that may be made available over a network. Computing resources 110 may each be substantially similar to one another or may be heterogeneous. As described above, entities accessing computing services provided by the provisioned computing resources environment may gain access to a suite of elastic IT infrastructure services (e.g., computing resources 110) as the entity requests those services. Provisioned computing resources environment 112 may provide a scalable, reliable, and secure distributed computing infrastructure.
In the illustrated example, each computing resource 110 comprises processing unit 130 and memory unit 132. Processing unit 130 may include one or more microprocessors, controllers, or any other suitable computing devices or resources. Processing unit 130 may work, either alone or with other components of system 100, to provide a portion or all of the functionality of system 100 described herein. Memory unit 132 may take the form of volatile or non-volatile memory including, without limitation, magnetic media, optical media, RAM, ROM, removable media, or any other suitable memory component. In certain embodiments, a portion of all of memory unit 132 may include a database, such as one or more structured query language (SQL) servers or relational databases. Although
Server system 106 may be coupled or otherwise associated with a storage module 108. Storage module 108 may take the form of volatile or non-volatile memory including, without limitation, magnetic media, optical media, RAM, ROM, removable media, or any other suitable memory component. In certain embodiments, a portion of all of storage module 108 may include a database, such as one or more SQL servers or relational databases. Storage module 108 may be a part of or distinct from memory unit 122 of server system 106.
Storage module 108 may store a variety of information and applications that may be used by server system 106 or other suitable components of system 100. In the illustrated example, storage module 108 may store user reliability data 124, instance reliability data 126, class reliability 128, and infrastructure repository 136. Although storage module 108 is described as including particular information and applications, storage module 108 may store any other suitable information and applications. Furthermore, although these particular information and applications are described as being stored in storage module 108, the present description contemplates storing this particular information and applications in any suitable location, according to particular needs.
System 100 provides just one example of an environment in which the reliability estimation for ad hoc applications technique of the present disclosure may be used. The present disclosure contemplates use of the data transfer technique in any suitable computing environment. Additionally, although functionality is described as being performed by certain components of system 100, the present disclosure contemplates other components performing that functionality. As just one example, functionality described with reference to server system 106 may be performed by one or more components of computing resources 110 and/or user system 102. Furthermore, although certain components are illustrated as being combined or separate, the present disclosure contemplates separating and/or combining components of system 100 in any suitable manner. As just one example, server system 106 and one or more of computing resources 110 may be combined in a suitable manner.
Certain embodiments of the present disclosure may provide some, none, or all of the following technical advantages. For example, certain embodiments provide a reliability estimate for computing resources based on a user identification of key components and associations. Receiving user indication of certain component relationships may allow providers to generate a reliability estimate for ad hoc applications without having to disclose infrastructure, network, and computing resource details to a user of the ad hoc application. As a result, particular embodiments of the present disclosure may provide a reliability estimate in a distributed system comprising many disparate components with differing levels of availability and redundancy. Thus, providers of ad hoc applications may satisfy a user demand for reliability estimates without having to reveal the details of the provisioned system. Accordingly, having a quantifiable measure of reliability for an application increases trust and lessens the risk of using a cloud provider or ad hoc applications.
In some embodiments, multiple users using one or more user systems 102 communicate tag information 138 that include the same string. Component tagging module 142 may disambiguate usage by placing metadata tags into a namespace associated with the user that applied the tag. For example, component tagging module 142 may record the string ‘accounting’ for a first user as ‘user1:accounting’ and the string ‘account’ for a second users as ‘user2:accounting’. Although component tagging module 142 may record each string using a namespace, when displaying tag information 138, component tagging module 142 may hide the namespace from the user. An example of tag information 138 applied to ad hoc application components is shown in
Additionally or alternatively, a user at user system 102 interacts with application definition module 144 to create an application definition (e.g., application definition 140) of a provisioned ad hoc application, as represented by arrow 202. Application definition 140 includes at least a primary computing resource 110 for which reliability estimate 134 is to be calculated. In some embodiments, application definition 140 may include one or more secondary computing resources 110 that are supportive of the primary computing resource 110. For example, the primary computing resource 110 may be an software service while a secondary computing resource 110 may be a web service accessed by the software service. An example application definition 140 is shown in
In some embodiments, application definition 140 may not define all secondary computing resources 110 used by a particular ad hoc application. Thus, graph inference module 146 may use application definition 140, tag information 138, and infrastructure repository 136 to construct the dependencies and interrelationships among the various computing resources 110 utilized by a particular ad hoc application for which reliability estimate 134 is sought, represented by arrow 203. In some embodiments, application definition module 144 defines the starting seeds for the graph inference module 146, and graph inference module 146 may expand the user-provided ad hoc application definition 140 into a more comprehensive application definition.
An example directed graph is shown in
Returning to
Graph inference module 146 may request infrastructure components that are allocated to the application components from the infrastructure repository module 136. Infrastructure components are part of the infrastructure provider's implementation of a resource and generally kept secret, for example, the arrangement of physical racks, network switches, power supplies, air conditioners, fire suppression units, telecommunication links, and buildings.
Graph inference module 146 may request application components associated with tags from component tagging module 142. For example, graph inference module 146 may locate application components tagged with an identifier associated with a resource.
Graph inference module 146 may display the identified application components to a user at user system 102 for validation. Although the inference of application components may be beneficial to the user by reducing time spent defining the application or tagging, an incorrect inference may unnecessarily expand the directed graph. In some embodiments, the graph inference module 146 may support a mechanism for excluding specific application components shown in a particular directed graph, for example, by having the user a apply a tag excluding the undesired component (e.g., a “does not require” tag) to override the standard inference algorithm used by graph inference module 146.
Once directed graph 500 is generated, application probability calculator 148 analyzes dependency relationships among application components in directed graph 500 to construct a conditional probability table 600. For example, in directed graph 500 shown in
Additionally, once directed graph 500 is generated, infrastructure probability calculator 150 analyzes dependency relationships among infrastructure components in directed graph 500 to construct a conditional probability table 600. For example, in directed graph 500 shown in
In some embodiments, infrastructure probability calculator 150 may introduce a noise term into conditional probability table 600 to obscure the exact configuration of infrastructure components. The use of noisy probability may improve accuracy by permitting the infrastructure provider to include infrastructure details in the model that might be revealed through inspection of reliability estimates.
In some embodiments, infrastructure probability calculator 150 may factor either instance reliability data 126 or class reliability data 128 more heavily in its calculation. For example, direct observation of infrastructure components availability (e.g., instance reliability data 126) may be preferred for component availability. If no direct observation exists, the component availability may be estimated based on fleet statistics for the component (e.g., class reliability data 128).
Once conditional probability tables 600 are calculated for each component in directed graph 500, reliability estimator module 152 evaluates the directed graph 500 and conditional probability tables 600 as a Bayesian network to produce reliability estimate 134. Exact computing of reliability of the primary resource (such as, e.g., Application in directed graph 500), may be possible for simple directed graphs, such as graphs with only a single path to any component. However, in many case, directed graph may not have a direct solution. In some embodiments reliability estimator module 152 may support stochastic simulation of the inferred directed graph 500 to compute the reliability of the primary resource. For example, reliability estimator module 152 may run a number of trials sampling different availability configurations according to the conditional probabilities (as shown, e.g., in conditional probability tables 600) for each component in directed graph 500. The availability of the primary resource may then be estimated by counting the number of failures of the primary resource according to the inferred directed graph 500 over a large number of trials. Numerous trials may be run in order to obtain reliability estimate 134 of the primary resource.
A table displaying the results of an example series of trials is disclosed in Table 7 of
Next, reliability estimator performs a second trial (“Trial 2”), the results of which are shown in table 700 in
Successive trials are performed (e.g., Trial 3 through Trial 1,000,000 shown in Table 700), and the number of times Application is assigned a 1 is compared to the number of times Application is assigned 0 in the aggregate number of trials. For example, reliability estimator module 152 may determine that in 99.89% of trials, Application is assigned a 1. Thus, reliability estimate 134 is calculated to be 99.89%. Once calculated, reliability estimate 134 may be stored in storage unit 108 and/or transmitted to user system 102 to be displayed to a user.
In some embodiments, a series of a trials may represent sampling from among all possible combinations of the availability status of each component in a directed graph. For example, reliability estimator module 152 may perform availability sampling to determine reliability estimate 134 for a particular primary computing resource 110 (such as, e.g., Application shown in
Reliability estimator module 152 may calculate the probability of a particular availability configuration based on conditional probability tables 600. As discussed further below, conditional probability tables 600 give a probability for each component to exist in a particular availability configuration, given the availability status of relied upon components. Since an availability configuration gives an availability status for each component in a directed graph simultaneously, the probability of the availability configuration occurring in practice is then the product of each of the component probabilities as indicated in the conditional probability table 600 associated with each component. The set of all availability configurations can be enumerated in a table in which each row is a particular availability configuration and each column is a component in a directed graph (such as, e.g., directed graph 500).
Since each component in a directed graph is assigned a 0 or 1, the total number of configurations (rows in the table) is 2 to the power of the number of components present in the directed graph (i.e., 2N). Even a small number of components makes examining every row (i.e., the probability associated with each availability configuration) infeasible. For example, a directed graph with 50 components would have a table with over one quadrillion rows. Therefore, in some embodiments, selected availability configurations are sampled in order to calculate reliability estimate 134. Sampling may be performed according to one or more methods. For example, in some embodiments, reliability estimator module 152 may divide the availability configurations into groups of relatively equal probability and may select particular samples from each group. The sampling performed may be an orthogonal sampling method, such as orthogonal Lain hypercube sampling.
In some embodiments, reliability estimator module 152 performs sampling by working backwards from the availability of the primary resource (such as, e.g., Application in directed graph 500 shown in
After sampling is performed, the sample probabilities are summed according to whether the primary resource is available or unavailable, producing two probabilities: an available probability (“A”) and an unavailable probability (“U”). The sum of A and U is greater than or equal to 0 but less than or equal to 1. In some embodiments, reliability estimator module 152 scales the available probability to 1 to calculate reliability estimate 134 (i.e., by calculating A/(A+U)).
Reliability estimate 134 is most accurate when A+U is close to 1 and becomes increasingly inaccurate as A+U approaches 0 since scaling the measurements is an approximation for the configurations that are not sampled. Reliability estimator module 152 may go back and perform additional sampling if A+U is too small to improve the accuracy of reliability estimate 134.
Conditional probability tables 600 for application components may be calculated by application probably calculator 148 and infrastructure components may be calculated by infrastructure probability calculator 150. For example, conditional probably table 600a illustrates conditional probabilities for Server 1 (box 505 in
Conditional probability table 600b illustrates conditional probabilities for the Database component illustrated in
Conditional probability table 600c illustrates conditional probabilities for the Switch component illustrated in
Computer system 800 may have one or more input devices 802 (which may include a keypad, keyboard, mouse, stylus, or other input devices), one or more output devices 804 (which may include one or more displays, one or more speakers, one or more printers, or other output devices), one or more storage devices 806, and one or more storage media 808. An input device 802 may be external or internal to computer system 800. An output device 804 may be external or internal to computer system 800. A storage device 806 may be external or internal to computer system 800. A storage medium 808 may be external or internal to computer system 800.
System bus 810 couples subsystems of computer system 800 to each other. Herein, reference to a bus encompasses one or more digital signal lines serving a common function. The present disclosure contemplates any suitable system bus 810 including any suitable bus structures (such as one or more memory buses, one or more peripheral buses, one or more a local buses, or a combination of the foregoing) having any suitable bus architectures. Example bus architectures include, but are not limited to, Industry Standard Architecture (ISA) bus, Enhanced ISA (EISA) bus, Micro Channel Architecture (MCA) bus, Video Electronics Standards Association local (VLB) bus, Peripheral Component Interconnect (PCI) bus, PCI-Express bus (PCI-X), and Accelerated Graphics Port (AGP) bus.
Computer system 800 includes one or more processors 812 (or central processing units (CPUs)). A processor 812 may contain a cache 814 for temporary local storage of instructions, data, or computer addresses. Processors 812 are coupled to one or more storage devices, including memory 816. Memory 816 may include RAM 818 and ROM 820. Data and instructions may transfer bi-directionally between processors 812 and RAM 818. Data and instructions may transfer uni-directionally to processors 812 from ROM 820. RAM 818 and ROM 820 may include any suitable computer-readable storage media.
Computer system 800 includes fixed storage 822 coupled bi-directionally to processors 812. Fixed storage 822 may be coupled to processors 812 via storage control unit 807. Fixed storage 822 may provide additional data storage capacity and may include any suitable computer-readable storage media. Fixed storage 822 may store an operating system (OS) 824, one or more executables (EXECs) 826, one or more applications or programs 828, data 830 and the like. Fixed storage 822 is typically a secondary storage medium (such as a hard disk) that is slower than primary storage. In appropriate cases, the information stored by fixed storage 822 may be incorporated as virtual memory into memory 816. In certain embodiments, fixed storage 822 may include network resources, such as one or more storage area networks (SAN) or network-attached storage (NAS).
Processors 812 may be coupled to a variety of interfaces, such as, for example, graphics control 832, video interface 834, input interface 836, output interface 837, and storage interface 838, which in turn may be respectively coupled to appropriate devices. Example input or output devices include, but are not limited to, video displays, track balls, mice, keyboards, microphones, touch-sensitive displays, transducer card readers, magnetic or paper tape readers, tablets, styli, voice or handwriting recognizers, biometrics readers, or computer systems. Network interface 840 may couple processors 812 to another computer system or to network 842. Network interface 840 may include wired, wireless, or any combination of wired and wireless components. Such components may include wired network cards, wireless network cards, radios, antennas, cables, or any other appropriate components. With network interface 840, processors 812 may receive or send information from or to network 842 in the course of performing steps of certain embodiments. Certain embodiments may execute solely on processors 812. Certain embodiments may execute on processors 812 and on one or more remote processors operating together.
In a network environment, where computer system 800 is connected to network 842, computer system 800 may communicate with other devices connected to network 842. Computer system 800 may communicate with network 842 via network interface 840. For example, computer system 800 may receive information (such as a request or a response from another device) from network 842 in the form of one or more incoming packets at network interface 840 and memory 816 may store the incoming packets for subsequent processing. Computer system 800 may send information (such as a request or a response to another device) to network 842 in the form of one or more outgoing packets from network interface 840, which memory 816 may store prior to being sent. Processors 812 may access an incoming or outgoing packet in memory 816 to process it, according to particular needs.
Certain embodiments involve one or more computer-storage products that include one or more tangible, computer-readable storage media that embody software for performing one or more steps of one or more processes described or illustrated herein. In certain embodiments, one or more portions of the media, the software, or both may be designed and manufactured specifically to perform one or more steps of one or more processes described or illustrated herein. Additionally or alternatively, one or more portions of the media, the software, or both may be generally available without design or manufacture specific to processes described or illustrated herein. Example computer-readable storage media include, but are not limited to, CDs (such as CD-ROMs), FPGAs, floppy disks, optical disks, hard disks, holographic storage devices, ICs (such as ASICs), magnetic tape, caches, PLDs, RAM devices, ROM devices, semiconductor memory devices, and other suitable computer-readable storage media. In certain embodiments, software may be machine code which a compiler may generate or one or more files containing higher-level code which a computer may execute using an interpreter.
As an example and not by way of limitation, memory 816 may include one or more tangible, computer-readable storage media embodying software and computer system 800 may provide particular functionality described or illustrated herein as a result of processors 812 executing the software. Memory 816 may store and processors 812 may execute the software. Memory 816 may read the software from the computer-readable storage media in mass storage device 816 embodying the software or from one or more other sources via network interface 840. When executing the software, processors 812 may perform one or more steps of one or more processes described or illustrated herein, which may include defining one or more data structures for storage in memory 816 and modifying one or more of the data structures as directed by one or more portions the software, according to particular needs.
In certain embodiments, the described processing and memory elements (such as processors 812 and memory 816) may be distributed across multiple devices such that the operations performed utilizing these elements may also be distributed across multiple devices. For example, software operated utilizing these elements may be run across multiple computers that contain these processing and memory elements. Other variations aside from the stated example are contemplated involving the use of distributed computing.
In addition or as an alternative, computer system 800 may provide particular functionality described or illustrated herein as a result of logic hardwired or otherwise embodied in a circuit, which may operate in place of or together with software to perform one or more steps of one or more processes described or illustrated herein. The present disclosure encompasses any suitable combination of hardware and software, according to particular needs.
Although the present disclosure describes or illustrates particular operations as occurring in a particular order, the present disclosure contemplates any suitable operations occurring in any suitable order. Moreover, the present disclosure contemplates any suitable operations being repeated one or more times in any suitable order. Although the present disclosure describes or illustrates particular operations as occurring in sequence, the present disclosure contemplates any suitable operations occurring at substantially the same time, where appropriate. Any suitable operation or sequence of operations described or illustrated herein may be interrupted, suspended, or otherwise controlled by another process, such as an operating system or kernel, where appropriate. The acts can operate in an operating system environment or as stand-alone routines occupying all or a substantial part of the system processing.
Moreover, data transfer techniques consistent with the present disclosure may be used to communicate any suitable type of data over any suitable type of network. For example, although the present disclosure has been described primarily with reference to reliability metrics data, the present disclosure contemplates processing any suitable type of data for communication of a communication network (e.g., network 104).
Patent | Priority | Assignee | Title |
Patent | Priority | Assignee | Title |
6606658, | Oct 17 1997 | Fujitsu Limited | Apparatus and method for server resource usage display by comparison of resource benchmarks to determine available performance |
7703072, | Sep 29 2004 | Hitachi, LTD | Component-based application constructing method |
7823009, | Feb 16 2001 | Virtuozzo International GmbH | Fault tolerant distributed storage for cloud computing |
20020138353, | |||
20070294290, | |||
20100042720, | |||
20100076856, | |||
20100299366, | |||
20100332262, | |||
20100332629, | |||
20110030065, | |||
20110061041, | |||
20110239039, | |||
20110271270, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Sep 01 2011 | ALLEN, NICHOLAS | Amazon Technologies, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 048094 | /0689 | |
Mar 22 2017 | Amazon Technologies, Inc. | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Sep 06 2022 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Date | Maintenance Schedule |
Apr 07 2023 | 4 years fee payment window open |
Oct 07 2023 | 6 months grace period start (w surcharge) |
Apr 07 2024 | patent expiry (for year 4) |
Apr 07 2026 | 2 years to revive unintentionally abandoned end. (for year 4) |
Apr 07 2027 | 8 years fee payment window open |
Oct 07 2027 | 6 months grace period start (w surcharge) |
Apr 07 2028 | patent expiry (for year 8) |
Apr 07 2030 | 2 years to revive unintentionally abandoned end. (for year 8) |
Apr 07 2031 | 12 years fee payment window open |
Oct 07 2031 | 6 months grace period start (w surcharge) |
Apr 07 2032 | patent expiry (for year 12) |
Apr 07 2034 | 2 years to revive unintentionally abandoned end. (for year 12) |