user-to-software-application-instance-pairings are created. Each of the pairings is a unique relationship between one of the users and one of the instances of the software applications. identifiers for the user-to-software-application-instance-pairings are received. There is a separate identifier for each of the user-to-software-application-instance-pairings. One of the log creation facilities is associated with each of the user-to-software-application-instance-pairings. log files are created at corresponding ones of the log creation facilities in response to detecting errors during execution of the instances of the software applications. The log files are categorized based on error categories. A request for a post error analysis report is received. The request specifies one of the error categories. A subset of the log files is determined based on the specified error category specified in the request. The subset of the log files is displayed. One of the identifiers is displayed for each error described in the post error analysis report.
|
11. A method of implementing a post error analysis system that includes log creation facilities associated with instances of software applications that users interact with, wherein the method comprises:
creating user-to-software-application-instance-pairings, wherein each of the pairings is a unique relationship between one of the users and one of the instances of the software applications;
receiving identifiers for the user-to-software-application-instance-pairings, wherein there is a separate identifier for each of the user-to-software-application-instance-pairings;
associating one of the log creation facilities with each of the user-to-software-application-instance-pairings;
creating log files at corresponding ones of the log creation facilities in response to detecting errors during execution of the instances of the software applications;
categorizing the log files based on error categories, wherein each of the log files includes a corresponding one of the error categories;
storing the log files in a log file repository of the post error analysis system;
receiving a request for a post error analysis report, wherein the request specifies one of the error categories;
determining a subset of the log files based on the specified error category specified in the request; and
displaying the subset of the log files in a post error analysis report, wherein the post error analysis report includes one of the identifiers for each error described in the post error analysis report whereby post error analysis is performed on a per user basis.
20. An apparatus for implementing a post error analysis system that includes log creation facilities associated with instances of software applications that users interact with comprising:
one or more processors; and
a non-transitory processor-readable storage device including instructions for:
creating user-to-software-application-instance-pairings, wherein each of the pairings is a unique relationship between one of the users and one of the instances of the software applications;
receiving identifiers for the user-to-software-application-instance-pairings, wherein there is a separate identifier for each of the user-to-software-application-instance-pairings;
associating one of the log creation facilities with each of the user-to-software-application-instance-pairings;
creating log files at corresponding ones of the log creation facilities in response to detecting errors during execution of the instances of the software applications;
categorizing the log files based on error categories, wherein each of the log files includes a corresponding one of the error categories;
storing the log files in a log file repository of the post error analysis system;
receiving a request for a post error analysis report, wherein the request specifies one of the error categories;
determining a subset of the log files based on the specified error category specified in the request; and
displaying the subset of the log files in the post error analysis report, wherein the post error analysis report includes one of the identifiers for each error described in the post error analysis report whereby post error analysis is performed on a per user basis.
1. A non-transitory processor-readable storage device including instructions for a method of implementing a post error analysis system that includes log creation facilities associated with instances of software applications that users interact with, wherein the non-transitory processor-readable storage device includes instructions executable by one or more processors for:
creating user-to-software-application-instance-pairings, wherein each of the pairings is a unique relationship between one of the users and one of the instances of the software applications;
receiving identifiers for the user-to-software-application-instance-pairings, wherein there is a separate identifier for each of the user-to-software-application-instance-pairings;
associating one of the log creation facilities with each of the user-to-software-application-instance-pairings;
creating log files at corresponding ones of the log creation facilities in response to detecting errors during execution of the instances of the software applications;
categorizing the log files based on error categories, wherein each of the log files includes a corresponding one of the error categories;
storing the log files in a log file repository of the post error analysis system;
receiving a request for a post error analysis report, wherein the request specifies one of the error categories;
determining a subset of the log files based on the specified error category specified in the request; and
displaying the subset of the log files in the post error analysis report, wherein the post error analysis report includes one of the identifiers for each error described in the post error analysis report whereby post error analysis is performed on a per user basis.
2. The non-transitory processor-readable storage device of
3. The non-transitory processor-readable storage device of
4. The non-transitory processor-readable storage device of
interspersing log files for different software applications and different hosts with each other in the log file repository.
5. The non-transitory processor-readable storage device of
6. The non-transitory processor-readable storage device of
associating each of the user-to-software-application-pairs and each of the associated facilities with a cloud service environment selected from a plurality of cloud service environments.
7. The non-transitory processor-readable storage device of
8. The non-transitory processor-readable storage device of
9. The non-transitory processor-readable storage device of
determining the subset of the log files based on the specified error category and one of the identifiers.
10. The non-transitory processor-readable storage device of
comparing log files for a list of product packages;
determining exception lines for each of the log files based on the comparing; and
categorizing the log files based, at least in part, on respective ones of the exception lines.
12. The method of
14. The method of
interspersing log files for different software applications and different hosts with each other in the log file repository.
15. The method of
16. The method of
17. The method of
18. The method of
19. The method of
determining the subset of the log files based on the specified error category and one of the identifiers.
|
This application is a continuation of the following application, U.S. patent application Ser. No. 15/664,956, entitled SYSTEM AND METHOD OF PROVIDING POST ERROR ANALYSIS FOR INSTANCES OF APPLICATIONS IN CLOUD SERVICE ENVIRONMENTS ON A PER USER BASIS, filed on Jul. 31, 2017, which is hereby incorporated by reference as if set forth in full in this application for all purposes.
A conventional software error prediction system analyzes volumes of information gathered from across an entire system for patterns in sequences of events documented in the gathered information. The conventional software error prediction systems uses the patterns to predict potential future problems. The pattern analysis can involve machine learning techniques.
Conventional security breach alerting systems also analyze volumes of information gathered from across an entire system. The analysis involves determining security breaches in one part of the system that the perpetrators may launch against another part of the system or even against a different system.
With both types of systems, information technology specialists are informed about the problems. Both types of systems rely on gathering extensive information across the entire system in order to detect errors in one part of the system and to determine whether those errors apply to other parts of the system. Both types of systems also rely on real time notification of the determined problems. Both types of systems rely on taking action in one part of a system based on a problem that was determined for another part of the system. Various types of error systems also rely on modifying applications in the system to provide information describing errors.
Various embodiments provide for performing post error analysis for instances of applications in cloud service environments on a per user basis. One embodiment provides for the instances of the applications executing in the cloud service environments that users interact with. Each of the cloud service environments includes a respective set of instances of the applications. Each of the cloud service environments and the respective set of instances is associated with a different one of the users. Errors are detected during the execution of the instances of the applications. Sets of log file information describing the errors are created. Each of the sets of log file information describes one of the errors. Log files are created. Each of the log files include one of the sets of log file information and an identification of a cloud service environment where an associated error occurred. The log files are categorized based on identifications of the cloud service environments. A post error analysis report including information from the categorized log files is provided for a particular cloud service environment whereby the post error analysis is performed on a per user basis.
An embodiment provides for tangible processor-readable storage device including instructions for a method of performing post error analysis for instances of applications in cloud service environments on a per user basis, wherein the tangible processor-readable storage device includes instructions executable by one or more processors for: executing the instances of the applications in the cloud service environments that users interact with, wherein each of the cloud service environments includes a respective set of instances of the applications and wherein each of the cloud service environments and the respective set of instances is associated with a different one of the users; detecting errors during the execution of the instances of the applications; creating sets of log file information describing the errors, wherein each of the sets of log file information describes one of the errors; creating log files that each include one of the sets of log file information and an identification of a cloud service environment where an associated error occurred; categorizing the log files based on identifications of the cloud service environments; and providing a post error analysis report including information from the categorized log files for a particular cloud service environment whereby the post error analysis is performed on a per user basis.
A further understanding of the nature and the advantages of particular embodiments disclosed herein may be realized by reference of the remaining portions of the specification and the attached drawings.
QA engineers test systems to locate errors in the system. Log files may have been created for the errors. However, when the system adequately recovers from errors, the QA engineer may not have seen function issues that arise from those errors. Further, in conventional systems, the volume of created log files can overwhelm the QA engineers.
Therefore, according to various embodiments, different parts of the system are executed on behalf of different users. Errors for parts of a system that respective users interact with are analyzed and post error analysis reports are provided, for example, to a QA engineer on a per user basis. Therefore, error information from many different sources does not need to be gathered. Instead, the QA engineer is provided with a single source of error information in the form of a report for a particular user of the system.
Any action that is taken based on a post error analysis report would apply to the part of the system that the report documented errors for. For example, if the report indicates that there is a software bug in application A, then the action taken would be to report and fix the bug in application A. The errors can be addressed in the near future or sometime later in the future. Further, the reports are provided without modifying the logic of the applications. Embodiments are well suited for production environments as well as test environments.
The system 100 includes a bug data query engine 140, a file system of previously debugged errors 142, a log file analysis engine 150, a QA user interface 160, a QA engineer 170, cloud service environments (CSEs) 110 and 120 with respective names CSE 1 and CSE 2, and a log file search facility 130. The bug data query engine 140 communicates with the file system 142. CSE 110 includes hosts 111 and 112. CSE 120 includes hosts 121 and 122. The hosts can be either hardware computer systems or virtual machines (VMs) installed on one or more hardware computer systems. Application instance 111 and log creation facility (LCF) 111-2 execute on host 111. Application instance 112-1 and LCF 112-2 execute on host 112. Application instance 121-1 and LCF 121-2 execute on host 121. Application instance 122-1 and LCF 122-2 execute on host 122. An application instance is an instance of an application. Two or more application instances may be instances of the same application. For example, application instance 121-1 and 111-1 may be instances of the same application for adding a consumer item to a cart. In another example, application instances 112-1 and 122-1 may be instances of the same application for billing the customer for the consumer item in the cart. User 181 interacts with CSE 110 and user 182 interacts with CSE 120.
According to one embodiment, the log file analysis engine 150 is implemented as a node.JS® application; the file system 142 is a Hadoop® Distributed File System (HDFS); the QA user interface 160 is implemented as a JQuery™/HyperText Markup Language (HTML) based application; the log creation facilities 111-2, 112-2, 121-2, 122-2 are implemented with Elastic™'s logstash from the Elasticsearch-Logstash-Kibana™ (ELK stack); and the log file search facility is implemented with Elastic™'s elasticsearch from the Elasticsearch-Logstash-Kibana™ (ELK stack);
The application instances 111-1, 112-1, 121-1, 122-1 of each of the CSEs 110, 120 execute in response to the requests of the respective users 181, 182. When an application instance encounters an exception, an exception handler for that type of exception gains control. Examples of exceptions include a null pointer was found, authorization failed, a class was not found, an instantiation exception, the method does not exist, an arithmetic exception, a class cast exception, and so on. The exception handler creates exception information describing that exception. A log creation facility 111-2, 112-2, 121-2, and 122-2 is associated with each of the application instances 111-1, 112-1, 121-1, 122-1. The log creation facility creates a log file for an exception from the application instance it is associated with and augments the exception information with error augmentation information.
Since each of the log creation facilities is associated with one application, one host, and one CSE, the log creation facilities know what host name, application instance name and cloud service environment name to augment the log files with. The log files with their respective augmentations are communicated to the log file search facility 130 where they are stored in the log files repository 131. According to one embodiment, the log creation facilities are configured to poll service logs and store the logs in the log file repository 131 of the log file search facility 130. According to one embodiment, the log file search facility 130 is a central ElasticSearch™ server. When requested, the log file search facility 130 provides a list of errors from the repository 131 to the log file analysis engine 150, as will become more evident.
Bug data query engine 140 provides GetBugsForError Representational State Transfer (REST) Application Program Interface (API) 241. Log file analysis engine 150 includes a bug correlation module 210, an error categorization module 220 and a log mining module 230. The bug correlation module 210 provides a GetAllIssuesForErrors REST API 211. The error categorization module 220 provides a GetAllCategorizedErrors REST API 221.
The QA UI 160 receives a request from the QA engineer requesting error information for a specified CSE. The QA UI 160 invokes the GetAllCategorizedErrors Rest API 221 requesting 274 error information for the specified CSE. In response, the log mining module 230 requests 281 log file information for that specified CSE from the log file search facility 130. The log file search facility 130 obtains the requested log files from the log file repository 131 and passes back the log files in the list of errors 282. Module 230 provides the list of errors 282 to module 220. The error categorization module 220 creates a RestErrorObjectArray 273 by processing the list of errors 282. The processing of the list of errors 282 includes correlating and categorizing error information in the list of errors 282 to create the RestErrorObjectArray 273. Similar or identical log files are correlated with each other to determine categorizations. A count of the similar or identical log files for an error category is determined. Error information along with the count is provided for each of the error categories. Each of the error categories are objects in the RestErrorObjectArray 273. The error information in the RestErrorObjectArray 273 is displayed to the QA engineer 170. The QA engineer 170 can select one or more of the displayed error categories. The GetAllKnownIssuesForErrors REST API 211 receives the one or more selected error categories 272. The error message 262 and exception line 263 associated with each of the selected error categories 272 is communicated from the bug correlation module 210 to the GetBugsForError REST API 241 of the engine 140. The bug data query engine 140 obtains a list of requested known bugs from the file system 142 and communicates the list of requested known bugs 261 back to module 210 in the form of a tuple<error object, list of known bugs>. A list of known bugs 271 is then communicated from the REST API 211 to the user interface 160 where it is displayed to the QA engineer.
According to one embodiment, a CSE 110, 120 (
When an application instance encounters an exception, such as a null pointer or authorization failed, an exception handler, according to one embodiment, for that type of exception gains control. The exception handler creates exception information 600 describing that exception. The exception information 600 includes an exception message 601, a main exception object 602, and an error trace stack 620. The exception message is text describing the type of exception error that occurred. As depicted in exception information 600, the type of exception error was “a null object found.” Another type of exception error is that authorization failed. The main exception object 602 specifies the object that implements the exception handler. The error trace stack 620 includes code lines 611, 612, 613. Each of the code lines 611-613 specify the method, the class the method is in, the package the class is in and a line of code that either invoked the next method up in the stack or in the case of the uppermost code line 611, the line of code where the exception occurred. Code line 611 specifies that mymethod1 is in myclass1.java in com.example.mypackage1, and so on with code lines 612 and 613. The error trace stack 620 also describes the order in which the methods communicated with each other. For example, at line 47 mymethod3 invoked mymethod 2; at line 72 mymethod2 invoked mymethod 1. Then the “null object found” exception occurred at line 56 of mymethod 1. According to one embodiment, an exception information 600 is a text description of the error.
Although many embodiments are illustrated in the context of exceptions and exception handlers, embodiments are well suited to errors that are not handled by exception handlers. For example, the errors may be for bugs in application instances that include logic that creates error information similar to the exception information 600. More specifically in the case of null pointer type errors, the application instance logic could include instructions that if a pointer equals null, then create information that is the same or similar to exception information 600.
The log creation facilities 111-2, 112-2, 121-2, and 122-2 create log files for the exceptions from the application instances that they are associated with and augment the exception information 600 with error augmentation information 700. The error augmentation information 700 includes an EntryType 701, a date/time field 702, a hostname field 703, an application name field 704 and a cloud service environment name field 705. The EntryType field 701 is set to the value “ERROR.” The date/time field 702 specifies the date and time that the error occurred. The hostname field 703 specifies the hostname of the system where the error originated. Examples of hostnames are host1, host2, host3, and host4 for respective hosts 111, 112, 121, and 122. The application name field 704 specifies the name of the application instance that generated the error. Examples of application names are app1, app2, app3, and app4 for respective application instances 111-1, 112-1, 121-1, and 122-1. The cloud service environment field 705 specifies the name of the cloud service environment. Examples of cloud service environment names are CSE1 and CSE2. Since each of the log creation facilities is associated with one application, one host, and one CSE, the log creation facilities know what host name 703, application instance name 704 and cloud service environment name 705 to augment the log files with.
According to one embodiment, the hosts are application servers that are running and configured with Logstash, providing the log creation facilities, as discussed herein. The LogStash configuration, according to one embodiment, is written as a grok parser that tells logstash how to parse the logs and send them to an elasticsearch based log file search facility 130 (
Each of the log files include information created by an exception handler along with the error augmentation information. For example, log file 810 includes error augmentation information 811-815 and exception information 816-820; log file 830 includes error augmentation information 831-835 and exception information 836-840; log file 850 includes exception information 851-855 and error augmentation information 856-860; log file 870 includes exception information 871-875 and error augmentation information 876-880.
For the sake of simplicity, it is assumed that the log files 810, 830, 850 and 870 appear consecutively in the log file repository 800. However, typically, these log files may be interspersed with log files from other application instances on other hosts for other CSEs. For example, the log files for application instances 111-1, 112-1, 121-1, and 122-1 (
Referring to
According to one embodiment, the repository 131 (
As depicted in
The list 900 is used in determining the relevant exception line for log files in a log file repository 800. The relevant exception line is the highest code line in an error trace stack that is also in the list of product packages 900. For example, referring to the error trace stack that includes lines 836-840 for log file 830 in
Typically, quality assurance personnel would want the first code line in an error stack trace to be the relevant exception line. The relevant exception line is the code line that is of the most interest to the quality assurance personnel. It may be a code line for a type of error that they have not handled well in the past.
The relevant exception lines are used as part of categorizing the exception errors. For example, since the log files 810, 850, 870 (
Still referring to
Error categorization module 220 (
At 1010, the method starts.
At 1020, execute the instances of the applications in the cloud service environments that users interact with.
For example, referring to
At 1030, detect errors during the execution of the instances of the applications.
For example, when an application instance 111-1, 112-1, 121-1, and 122-1 (
At 1040, create sets of log file information describing the errors.
For example, referring to
At 1050, create log files that each include one of the sets of log file information and an identification of a cloud service environment that an associated error occurred in.
For example, one of the log creation facilities 111-2, 112-2, 121-2, and 122-2 (
According to one embodiment, each of the log files are augmented with an identifier that uniquely identifies only one of the users. That identifier could be any one or more of an application instance name, a host name, and a CSE name. Each CSE and its associated application instances and hosts are associated with only one user. Therefore, any one or more of a CSE and corresponding application instance name(s) or host name(s) to uniquely identify one of the users, and, thus to augment log files to provide reports on a per user basis. For similar reasons, a CSE name, or a name of an application instance or a host that resides in that CSE can be used to identify a CSE and in turn the user of that CSE.
At 1060, categorize the log files based on the identifications of the cloud service environments.
For example, the QA UI 160 (
The log file search facility 130 (
The log mining module 230 (
Referring to
At 1070, provide a post error analysis report that includes information from categorized log files for a particular cloud service environment whereby the post error analysis is performed on a per user basis.
For example, the RestErrorObjectArray 273 (
At 1080, the method ends.
The QA engineer 170 (
An embodiment provides for augmenting each of the sets of log file information with additional information for the associated error, wherein the additional information includes an entry type, a date that the associated error occurred, hostname that the associated error occurred on, a name of an application the associated error occurred in, and the identification of the cloud service environment that the associated error occurred in. An example of a set of log file information is the exception information 600 (
An embodiment provides for wherein the creating of the sets of log file information further comprises: creating error related information for a particular error, wherein the error related information includes an exception message, a relevant exception line and a stack trace. An example of error related information is exception information 600 (
An embodiment provides for wherein the categorizing of the log files further includes: categorizing errors with at least similar error related information. For example, as discussed herein, log files 810, 850, and 870 are at least similar.
An embodiment provides for wherein the method further comprises: counting number of errors with the at least similar error related information in the log files; and providing in the post error analysis report, information about the categorizing of the errors with at least similar error related information and the number of errors with the at least similar error related information. For example, the table 410 (
An embodiment provides for determining if any of the errors have been previously debugged; and indicating which of the errors have been previously debugged in the post error analysis report. For example, the file system of previously debugged errors 142 can provide a list of requested known bugs 261. The bug details 522, 532, 542 can indicate which of the errors have been previous debugged. For example, the bug details can indicate that the error is in the process of being debugged or whether the patch for fixing the bug is available.
An embodiment provides for executing the instances on hosts, wherein each of the instances is executed on a different one of the hosts. For example, each of the application instances 111-1, 112-1, 121-1, and 122-1 execute on only one of the hosts 111, 112, 121, and 122.
An embodiment provides for wherein the method is performed without modifying the applications. For example, if the instances 111-1 and 121-1 are instances of an application to add a consumer item to a cart and instances 112-1 and 122-1 are instances of an application to bill the customer for the consumer item in the cart, the shopping cart application and the bill customer application have not been modified or are not required to be modified in order to perform the method of flowchart 1000 or any other embodiments described herein.
Some conventional systems require that the applications are modified, for example, by embedding a plugin called a “notifier” into their application. For example, if the application is a java application, a java based notifier may be embedded in the application. The java based notifier may rely on using a java script library in the application. Further, code may need to be added to the application to utilize the notifier and to communicate data to the conventional system. Conventional types of analysis tracking tools frequently require developers to embed at least certain custom code into their applications so that an end system can detect the type of event or issue that is of interest.
An embodiment provides for associating log creation facilities with the instances, wherein each of the instances is associated with a different one of the log creation facilities. For example, LCF 111-2 is associated with application instance 111-1; LCF 112-2 is associated with application instance 112-1; LCF 121-2 is associated with application instance 121-1; and LCF 122-2 is associated with application instance 122-1.
An embodiment provides for wherein the creating of the log files further comprises: augmenting, performed by log creation facilities, the sets of log file information with the identifications of the cloud service environments, wherein there is one identification with each of the cloud service environments; and creating, performed by the log creation facilities, the log files that each include one of the sets of log file information and the identification of the cloud service environment where the associated error occurred. For example, the log creation facilities 111-2, 112-2, 121-2, and 122-2 (
Unless otherwise specified, any one or more of the embodiments described herein can be implemented using processor readable instructions which reside, for example, in tangible processor-readable storage device of a computer system or like device. The tangible processor-readable storage device can be any kind of physical memory that instructions can be stored on. Examples of the tangible processor-readable storage device include but are not limited to a disk, a compact disk (CD), a digital versatile device (DVD), read only memory (ROM), flash, and so on. As described above, certain processes and operations of various embodiments of the present invention are realized, in one embodiment, as a series of processor readable instructions (e.g., software program) that reside within tangible processor-readable storage device of a computer system and are executed by one or more processors of the computer system. When executed, the instructions cause a computer system to implement the functionality of various embodiments of the present invention. For example, the instructions can be executed by a processor. The processor is a hardware processor, such as a central processing unit, associated with the computer system. The tangible processor-readable storage device is hardware memory and the one or more processors are hardware processors.
An embodiment provides for tangible processor-readable storage device including instructions for a method of performing post error analysis for instances of applications in cloud service environments on a per user basis, wherein the tangible processor-readable storage device includes instructions executable by one or more processors for: executing the instances of the applications in the cloud service environments that users interact with, wherein each of the cloud service environments includes a respective set of instances of the applications and wherein each of the cloud service environments and the respective set of instances is associated with a different one of the users; detecting errors during the execution of the instances of the applications; creating sets of log file information describing the errors, wherein each of the sets of log file information describes one of the errors; creating log files that each include one of the sets of log file information and an identification of a cloud service environment where an associated error occurred; categorizing the log files based on identifications of the cloud service environments; and providing a post error analysis report for a particular cloud service environment whereby the post error analysis is performed on a per user basis.
An embodiment provides for an apparatus for performing post error analysis for instances of applications in cloud service environments on a per user basis comprising: one or more processors; and a tangible processor-readable storage device including instructions for: executing the instances of the applications in the cloud service environments that users interact with, wherein each of the cloud service environments includes a respective set of instances of the applications and wherein each of the cloud service environments and the respective set of instances is associated with a different one of the users; detecting errors during the execution of the instances of the applications; creating sets of log file information describing the errors, wherein each of the sets of log file information describes one of the errors; creating log files that each include one of the sets of log file information and an identification of a cloud service environment where an associated error occurred; categorizing the log files based on identifications of the cloud service environments; and providing a post error analysis report for a particular cloud service environment whereby the post error analysis is performed on a per user basis.
The general system 1100 includes user devices 1160-1190, including desktop computers 1160, notebook computers 1170, smartphones 1180, mobile phones 1185, and tablets 1190. The general system 1100 can interface with any type of user device, such as a thin-client computer, Internet-enabled mobile telephone, mobile Internet access device, tablet, electronic book, or personal digital assistant, capable of displaying and navigating web pages or other types of electronic documents and UIs, and/or executing applications. Although the system 1100 is shown with five user devices, any number of user devices can be supported.
A web server 1110 is used to process requests from web browsers and standalone applications for web pages, electronic documents, enterprise data or other content, and other data from the user computers. The web server 1110 may also provide push data or syndicated content, such as RSS feeds, of data related to enterprise operations.
An application server 1120 operates one or more applications. The applications can be implemented as one or more scripts or programs written in any programming language, such as Java, C, C++, C#, or any scripting language, such as JavaScript or ECMAScript (European Computer Manufacturers Association Script), Perl, PHP (Hypertext Preprocessor), Python, Ruby, or TCL (Tool Command Language). Applications can be built using libraries or application frameworks, such as Rails, Enterprise JavaBeans, or .NET. Web content can created using HTML (HyperText Markup Language), CSS (Cascading Style Sheets), and other web technology, including templating languages and parsers.
The data applications running on the application server 1120 are adapted to process input data and user computer requests and can store or retrieve data from data storage device or database 1130. Database 1130 stores data created and used by the data applications. In an embodiment, the database 1130 includes a relational database that is adapted to store, update, and retrieve data in response to SQL format commands or other database query languages. Other embodiments may use unstructured data storage architectures and NoSQL (Not Only SQL) databases.
In an embodiment, the application server 1120 includes one or more general-purpose computers capable of executing programs or scripts. In an embodiment, web server 1110 is implemented as an application running on the one or more general-purpose computers. The web server 1110 and application server 1120 may be combined and executed on the same computers.
An electronic communication network 1140-1150 enables communication between user computers 1160-1190, web server 1110, application server 1120, and database 1130. In an embodiment, networks 1140-1150 may further include any form of electrical or optical communication devices, including wired network 1140 and wireless network 1150. Networks 1140-1150 may also incorporate one or more local-area networks, such as an Ethernet network, wide-area networks, such as the Internet; cellular carrier data networks; and virtual networks, such as a virtual private network.
The system 1100 is one example for executing applications according to an embodiment of the invention. In another embodiment, web server 1110, application server 1120, and optionally database 1130 can be combined into a single server computer application and system. In a further embodiment, virtualization and virtual machine applications may be used to implement one or more of the web server 1110, application server 1120, and database 1130.
In still further embodiments, all or a portion of the web and application serving functions may be integrated into an application running on each of the user computers. For example, a JavaScript application on the user computer may be used to retrieve or analyze data and display portions of the applications.
With reference to
In a particular example embodiment, browsers of the desktop computer 1160, notebook computer 1170, smartphone 1180, mobile phone 1185, tablet 1190 of
Computing device 1200 also includes a software application 1210, which may be stored on memory 1206 or on any other suitable storage location or computer-readable medium. Software application 1210 provides instructions that enable processor 1202 to perform the functions described herein and other functions. The components of computing system 1200 may be implemented by one or more processors or any combination of hardware devices, as well as any combination of hardware, software, firmware, etc.
For ease of illustration,
The hosts 111, 112, 121, 122 (
Although the description has been described with respect to particular embodiments thereof, these particular embodiments are merely illustrative, and not restrictive. For example, the features and operations could be arranged differently.
Although many examples provided herein are for analysis of weblogic server logs of java/javascript/html errors, which are the typical ones seen in a J2EE server logs, the various embodiments are not limited to J2EE systems. Servers that are built on other technologies, such as Perl, Python, Jython, CGI, Php, among others, typically generate errors/exceptions that typically include the same attributes i.e., error message and stack trace elements. Similarly, any system that can query an “issue/bug database” and identify potential known issues would work for any of the technologies and is not restricted to J2EE technologies. In essence, various embodiments are extendable to optimization of QA engineer log exception analysis activities on any kind of system built on any technology.
Although many embodiments are illustrated in the context of exceptions and exception handlers, embodiments are well suited to errors that are not handled by exception handlers. For example, the errors may be for bugs in application instances that include logic that create error information similar to the exception information 600. More specifically in the case of null pointer type errors, the application instance logic could include instructions that if a pointer equals null, then create information that is the same or similar to exception information 600.
Any suitable programming language can be used to implement the routines of particular embodiments including C, C++, Java, assembly language, etc. Different programming techniques can be employed such as procedural or object oriented. The routines can execute on a single processing device or multiple processors. Although the steps, operations, or computations may be presented in a specific order, this order may be changed in different particular embodiments. In some particular embodiments, multiple steps shown as sequential in this specification can be performed at the same time.
Particular embodiments may be implemented in a computer-readable storage medium for use by or in connection with the instruction execution system, apparatus, system, or device. Particular embodiments can be implemented in the form of control logic in software or hardware or a combination of both. The control logic, when executed by one or more processors, may be operable to perform that which is described in particular embodiments.
Particular embodiments may be implemented by using a programmed general purpose digital computer, by using application specific integrated circuits, programmable logic devices, field programmable gate arrays, optical, chemical, biological, quantum or nanoengineered systems, components and mechanisms may be used. In general, the functions of particular embodiments can be achieved by any means as is known in the art. Distributed, networked systems, components, and/or circuits can be used. Communication, or transfer, of data may be wired, wireless, or by any other means.
It will also be appreciated that one or more of the elements depicted in the drawings/figures can also be implemented in a more separated or integrated manner, or even removed or rendered as inoperable in certain cases, as is useful in accordance with a particular application. It is also within the spirit and scope to implement a program or code that can be stored in a machine-readable medium to permit a computer to perform any of the methods described above.
A “processor” includes any suitable hardware and/or software system, mechanism or component that processes data, signals or other information. A processor can include a system with a general-purpose central processing unit, multiple processing units, dedicated circuitry for achieving functionality, or other systems. Processing need not be limited to a geographic location, or have temporal limitations. For example, a processor can perform its functions in “real time,” “offline,” in a “batch mode,” etc. Portions of processing can be performed at different times and at different locations, by different (or the same) processing systems. Examples of processing systems can include servers, clients, end user devices, routers, switches, networked storage, etc. A computer may be any processor in communication with a memory. The memory may be any suitable processor-readable storage medium, such as random-access memory (RAM), read-only memory (ROM), magnetic or optical disk, or other tangible media suitable for storing instructions for execution by the processor.
As used in the description herein and throughout the claims that follow, “a”, “an”, and “the” includes plural references unless the context clearly dictates otherwise. Also, as used in the description herein and throughout the claims that follow, the meaning of “in” includes “in” and “on” unless the context clearly dictates otherwise.
Thus, while particular embodiments have been described herein, latitudes of modification, various changes, and substitutions are intended in the foregoing disclosures, and it will be appreciated that in some instances some features of particular embodiments will be employed without a corresponding use of other features without departing from the scope and spirit as set forth. Therefore, many modifications may be made to adapt a particular situation or material to the essential scope and spirit.
Vedurumudi, Prasad V V, Morusupalli, Praveen
Patent | Priority | Assignee | Title |
Patent | Priority | Assignee | Title |
10061680, | Mar 17 2014 | SPLUNK Inc. | Mobile application error monitoring system |
10110419, | Dec 17 2015 | CA, Inc. | Alarm to event tracing |
10185613, | Apr 29 2016 | VMware LLC | Error determination from logs |
10223367, | Aug 27 2014 | SALESFORCE, INC | Distributed sorting of event log files |
10241848, | Sep 30 2016 | Microsoft Technology Licensing, LLC | Personalized diagnostics, troubleshooting, recovery, and notification based on application state |
10476768, | Oct 03 2016 | Microsoft Technology Licensing, LLC | Diagnostic and recovery signals for disconnected applications in hosted service environment |
7047100, | Jul 05 2001 | SCREEN HOLDINGS CO , LTD | Substrate processing system managing apparatus information of substrate processing apparatus |
7490268, | Jun 01 2004 | The Trustees of Columbia University in the City of New York | Methods and systems for repairing applications |
8032866, | Mar 27 2003 | BMC Software Israel Ltd | System and method for troubleshooting runtime software problems using application learning |
8255902, | Mar 17 2008 | GEN DIGITAL INC | Systems and methods for determining and quantifying the impact of an application on the health of a system |
8365019, | Jun 16 2009 | MAPLEBEAR INC | System and method for incident management enhanced with problem classification for technical support services |
8418000, | Mar 13 2012 | BFACTOR LLC | System and methods for automated testing of functionally complex systems |
8656226, | Jan 31 2011 | CLOUD BYTE LLC | System and method for statistical application-agnostic fault detection |
9086960, | Aug 21 2012 | International Business Machines Corporation | Ticket consolidation for multi-tiered applications |
9262253, | Jun 28 2012 | Microsoft Technology Licensing, LLC | Middlebox reliability |
9465685, | Feb 02 2015 | International Business Machines Corporation | Identifying solutions to application execution problems in distributed computing environments |
9652316, | Mar 31 2015 | CA, Inc.; CA, INC | Preventing and servicing system errors with event pattern correlation |
9881470, | Oct 07 2013 | GOOGLE LLC | Automated crowdsourced power outage identification and staggering of HVAC system restarts |
9940223, | Apr 05 2013 | FEV NORTH AMERICA, INC | Human-machine interface test system |
20020022969, | |||
20020032763, | |||
20050144526, | |||
20060070037, | |||
20090055684, | |||
20120005542, | |||
20120166869, | |||
20130219050, | |||
20130227115, | |||
20140136690, | |||
20140149576, | |||
20140229606, | |||
20150081918, | |||
20150221109, | |||
20150227598, | |||
20150271008, | |||
20160259677, | |||
20170054610, | |||
20170099191, | |||
20170102989, | |||
20170279928, | |||
20170286199, | |||
20180285234, | |||
20180287856, | |||
20190005018, | |||
20190081871, | |||
20190097969, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Jul 01 2019 | Oracle International Corporation | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Jul 01 2019 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Nov 13 2024 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Date | Maintenance Schedule |
Jun 01 2024 | 4 years fee payment window open |
Dec 01 2024 | 6 months grace period start (w surcharge) |
Jun 01 2025 | patent expiry (for year 4) |
Jun 01 2027 | 2 years to revive unintentionally abandoned end. (for year 4) |
Jun 01 2028 | 8 years fee payment window open |
Dec 01 2028 | 6 months grace period start (w surcharge) |
Jun 01 2029 | patent expiry (for year 8) |
Jun 01 2031 | 2 years to revive unintentionally abandoned end. (for year 8) |
Jun 01 2032 | 12 years fee payment window open |
Dec 01 2032 | 6 months grace period start (w surcharge) |
Jun 01 2033 | patent expiry (for year 12) |
Jun 01 2035 | 2 years to revive unintentionally abandoned end. (for year 12) |