An approach is provided for selecting one or more trust factors from trust factors included in a trust index repository. Thresholds are identified corresponding to one or more of the selected trust factors. Actions are identified to perform when the selected trust factors reach the corresponding threshold values. The identified thresholds, identified actions, and selected trust factors are stored in a data store. The selected trust factors are monitored by comparing one or more trust metadata scores with the stored identified thresholds. The stored identified actions that correspond to the selected trust factors are performed when one or more of the trust metadata scores reach the identified thresholds. At least one of the actions includes an event notification that is provided to a trust data consumer.
|
1. A computer-implemented method comprising:
selecting one or more trust factors from a plurality of trust factors included in a trust index repository;
receiving one or more priority values that correspond to one or more of the trust factors;
generating, by a processor, utilizing the one or more priority values, one or more trust metadata scores that correspond to the one or more trust factors;
identifying one or more thresholds corresponding to one or more of the selected trust factors;
determining that one or more of the trust metadata scores reach one or more of the identified thresholds; and
performing one or more actions in response to determining that one or more of the trust metadata scores reached one or more of the identified thresholds, wherein at least one of the actions includes sending an event notification to a trust data consumer that includes one or more of the trust metadata scores that reached one or more of the identified thresholds.
12. A computer program product stored in a non-transitory computer readable medium, comprising functional descriptive material that, when executed by an information handling system, causes the information handling system to perform actions that include:
selecting one or more trust factors from a plurality of trust factors included in a trust index repository;
receiving one or more priority values that correspond to one or more of the trust factors:
generating, utilizing the one or more priority values, one or more trust metadata scores that correspond to the one or more trust factors;
identifying one or more thresholds corresponding to one or more of the selected trust factors;
determining that one or more of the trust metadata scores reach one or more of the identified thresholds; and
performing one or more actions in response to determining that one or more of the trust metadata scores reached one or more of the identified thresholds, wherein at least one of the actions includes sending an event notification to a trust data consumer that includes one or more of the trust metadata scores that reached one or more of the identified thresholds.
7. An information handling system comprising:
one or more processors;
a memory accessible by at least one of the processors;
one or more nonvolatile storage devices accessible by at least one of the processors;
a set of instructions which are loaded into memory and executed by at least one of the processors in order to perform actions of:
selecting one or more trust factors from a plurality of trust factors included in a trust index repository stored on at least one of the nonvolatile storage devices;
receiving one or more priority values that correspond to one or more of the trust factors;
generating, utilizing the one or more priority values, one or more trust metadata scores that correspond to the one or more trust factors;
identifying one or more thresholds corresponding to one or more of the selected trust factors;
determining that one or more of the trust metadata scores reach one or more of the identified thresholds; and
performing one or more actions in response to determining that one or more of the trust metadata scores reached one or more of the identified thresholds, wherein at least one of the actions includes sending an event notification to a trust data consumer that includes one or more of the trust metadata scores that reached one or more of the identified thresholds.
2. The method of
3. The method of
computing the atomic metadata scores using one or more facts retrieved from one or more data sources as input to the corresponding atomic trust factors; and
computing the composite metadata scores using one or more of the atomic metadata scores as input to the composite trust factors.
4. The method of
5. The method of
6. The method of
8. The information handling system of
9. The information handling system of
computing the atomic metadata scores using one or more facts retrieved from one or more data sources as input to the corresponding atomic trust factors; and
computing the composite metadata scores using one or more of the atomic metadata scores as input to the composite trust factors.
10. The information handling system of
11. The information handling system of
13. The computer program product of
14. The computer program product of
computing the atomic metadata scores using one or more facts retrieved from one or more data sources as input to the corresponding atomic trust factors; and
computing the composite metadata scores using one or more of the atomic metadata scores as input to the composite trust factors.
15. The computer program product of
16. The computer program product of
17. The computer program product of
|
1. Technical Field
The present invention relates to an approach for assessing the trust of data information, and related metadata. More particularly, the present invention relates to an approach for assessing trust of data, information and related metadata used by an business or organization.
2. Description of the Related Art
Many consumers of information—organizations and individuals—struggle with the uncertainty of trust in the information that they need to consume. Based on this struggle as well as the struggle of others, they know that the information that they are consuming can sometimes be trusted, and at other times not trusted. A challenge in the current state of the art is that information consumers do not know when and why they can trust information and which information they can trust. Often, the only reasonable approach for information consumers is to assume—with some level of doubt—that the information can be trusted until problems occur. Information providers have a similar challenge. Not all of the information is managed in the same way and may therefore have different levels of trust. Today, the information provider is challenged in distinguishing between the varying degrees of trust when delivering the information to the information consumer. Problems with possibly a small set of their information may reduce significantly the value of the overall provider system as the consumer does not know which information can be trusted.
It has been discovered that the aforementioned challenges are resolved by selecting one or more trust factors from trust factors included in a trust index repository. Thresholds are identified corresponding to one or more of the selected trust factors. Actions are identified to perform when the selected trust factors reach the corresponding threshold values. The identified thresholds, identified actions, and selected trust factors are stored in a data store. The selected trust factors are monitored by comparing one or more trust metadata scores with the stored identified thresholds. The stored identified actions that correspond to the selected trust factors are performed when one or more of the trust metadata scores reach the identified thresholds. At least one of the actions includes an event notification that is provided to a trust data consumer.
The foregoing is a summary and thus contains, by necessity, simplifications, generalizations, and omissions of detail; consequently, those skilled in the art will appreciate that the summary is illustrative only and is not intended to be in any way limiting. Other aspects, inventive features, and advantages of the present invention, as defined solely by the claims, will become apparent in the non-limiting detailed description set forth below.
The present invention may be better understood, and its numerous objects, features, and advantages made apparent to those skilled in the art by referencing the accompanying drawings, wherein:
The following is intended to provide a detailed description of an example of the invention and should not be taken to be limiting of the invention itself. Rather, any number of variations may fall within the scope of the invention, which is defined in the claims following the description.
Certain specific details are set forth in the following description and figures to provide a thorough understanding of various embodiments of the invention. Certain well-known details often associated with computing and software technology are not set forth in the following disclosure, however, to avoid unnecessarily obscuring the various embodiments of the invention. Further, those of ordinary skill in the relevant art will understand that they can practice other embodiments of the invention without one or more of the details described below. Finally, while various methods are described with reference to steps and sequences in the following disclosure, the description as such is for providing a clear implementation of embodiments of the invention, and the steps and sequences of steps should not be taken as required to practice this invention. Instead, the following is intended to provide a detailed description of an example of the invention and should not be taken to be limiting of the invention itself. Rather, any number of variations may fall within the scope of the invention, which is defined by the claims that follow the description.
The following detailed description will generally follow the summary of the invention, as set forth above, further explaining and expanding the definitions of the various aspects and embodiments of the invention as necessary. To this end, this detailed description first sets forth a computing environment in
Northbridge 115 and Southbridge 135 are connected to each other using bus 119. In one embodiment, the bus is a Direct Media Interface (DMI) bus that transfers data at high speeds in each direction between Northbridge 115 and Southbridge 135. In another embodiment, a Peripheral Component Interconnect (PCI) bus is used to connect the Northbridge and the Southbridge. Southbridge 135, also known as the I/O Controller Hub (ICH) is a chip that generally implements capabilities that operate at slower speeds than the capabilities provided by the Northbridge. Southbridge 135 typically provides various busses used to connect various components. These busses can include PCI and PCI Express busses, an ISA bus, a System Management Bus (SMBus or SMB), a Low Pin Count (LPC) bus. The LPC bus is often used to connect low-bandwidth devices, such as boot ROM 196 and “legacy” I/O devices (using a “super I/O” chip). The “legacy” I/O devices (198) can include serial and parallel ports, keyboard, mouse, floppy disk controller. The LPC bus is also used to connect Southbridge 135 to Trusted Platform Module (TPM) 195. Other components often included in Southbridge 135 include a Direct Memory Access (DMA) controller, a Programmable Interrupt Controller (PIC), a storage device controller, which connects Southbridge 135 to nonvolatile storage device 300 such as a hybrid hard disk drive, using bus 184.
ExpressCard 155 is a slot used to connect hot-pluggable devices to the information handling system. ExpressCard 155 supports both PCI Express and USB connectivity as it is connected to Southbridge 135 using both the Universal Serial Bus (USB) the PCI Express bus. Southbridge 135 includes USB Controller 140 that provides USB connectivity to devices that connect to the USB. These devices include webcam (camera) 150, infrared (IR) receiver 148, Bluetooth device 146 which provides for wireless personal area networks (PANs), keyboard and trackpad 144, and other miscellaneous USB connected devices 142, such as a mouse, removable nonvolatile storage device 145, modems, network cards, ISDN connectors, fax, printers, USB hubs, and many other types of USB connected devices. While removable nonvolatile storage device 145 is shown as a USB-connected device, removable nonvolatile storage device 145 could be connected using a different interface, such as a Firewire interface, etc.
Wireless Local Area Network (LAN) device 175 is connected to Southbridge 135 via the PCI or PCI Express bus 172. LAN device 175 typically implements one of the IEEE 802.11 standards of over-the-air modulation techniques that all use the same protocol to wireless communicate between information handling system 100 and another computer system or device. Optical storage device 190 is connected to Southbridge 135 using Serial ATA (SATA) bus 188. Serial ATA adapters and devices communicate over a high-speed serial link. The Serial ATA bus is also used to connect Southbridge 135 to other forms of storage devices, such as hard disk drives. Audio circuitry 160, such as a sound card, is connected to Southbridge 135 via bus 158. Audio circuitry 160 is used to provide functionality such as audio line-in and optical digital audio in port 162, optical digital output and headphone jack 164, internal speakers 166, and internal microphone 168. Ethernet controller 170 is connected to Southbridge 135 using a bus, such as the PCI or PCI Express bus. Ethernet controller 170 is used to connect information handling system 100 with a computer network, such as a Local Area Network (LAN), the Internet, and other public and private computer networks.
While
The Trusted Platform Module (TPM 195) shown in
Administration and configuration process 340 is used to set up and administer the Trust Index Functionality. For example, the configuration is used to identify which data items (facts) are being analyzed for the organization. Other configuration operations include identifying the algorithms and other metadata items, such as priority values, that are used to generate trust factor scores. Administration is used to identify the users that are authorized to access various trust data stored in the trust index repository. For example, Information Consumer 360 may be an executive in the organization, such as a Chief Information Officer (CIO) or Chief Executive Officer (CEO). Executives likely have different data needs than a sales or marketing manager. The sales/marketing manager may have a need to access sales and marketing related data as well as its associated trust data maintained in Trust Index Repository 320, but does not have a need to access strategic business data. Likewise, the CIO or CEO likely has a need for the strategic business data. Consequently, process 340 identifies the roles that individuals have and their corresponding data access needs. Events management 345 monitors facts 300 and trust data that is generated and stored in Trust Index Repository 320. For example, an Information Consumer may wish to be alerted when a particular trust factor reaches a particular threshold. In one embodiment, a message is sent to a trust data consumer that includes trust metadata scores that reached the identified threshold. As shown, Information Consumer 360 can be an individual, such as an employee, or a separate process, such as a sales and marketing computer system, that interacts with Trust Index Functionality 305.
Security 350 is used to ensure that individuals and processes access data that is applicable to the particular Information Consumer as established by Administration process 340. In addition, security maintains audit trails to establish which user or process updated any particular facts, algorithms, or trust factors.
User Interface process 355 is, as the name implies, used to provide a user interface to Information Consumer 360 in order to retrieve facts and trust metadata from Trust Index Repository 320. The user interface includes a capability to “drill down” into a trust factor so that the information consumer can better understand the underlying data and trust metadata to, for example, generate a particular composite trust factor score.
As shown in
Trust Factor 440 represents the measurement of trust in the information that is identified in the scope. In one embodiment, the level of trust is measured by a score in a scale from 0 (for the lowest level of trust) to 100 (for the highest level of trust). Threshold Value 460 defines configurable boundaries to specify for each trust factor which part of the scale is related to a more coarse-grained scale such as the typical green-yellow-red. In one embodiment, to classify a score as green means that the level of trust in the associated information meets the requirements of the consumer. The threshold is determined by the consumer, i.e. the consumer specifies the boundaries of the thresholds for each score. For example, if the availability of data is 99.97% and the requirement was >99.9% availability, then the trust score for information availability would be green. If the score is classified as yellow, then information should be consumed with care. Even though it doesn't meet the defined requirements, it may be still “good enough” to be consumed. For example, the completeness of information may not be 99% as required due to data load exceptions but higher than 90%. A classification of a score as red means that the information cannot be trusted.
As described above, trust factors can be aggregated. This is represented in the model above in the two subtypes Atomic Trust Factor 420 and Composite Trust Factor 450. The score of Composite Trust Factor 450 is calculated based on its related set of trust factors (Trust Factor 440). The score of Atomic Trust Factor 460 is calculated from facts 410.
Score Calculation object 430 provides algorithms to calculate scores from underlying facts or other factors. The set of algorithms can be modified and extended depending on the requirements of specific projects.
Fact object 410 represents the input data to calculate the score of atomic trust factors (Atomic Trust Factor 420). It is populated from the databases of information providers and other (meta-)data repositories. The definition of trust factors determines the related input data, which are the facts for atomic trust factors or other trust factors for composite trust factors. The definition of the facts then determines the sources—such as a metadata repository—and the algorithm for collecting the fact data from the source and possibly transforming it so that it can be populated into the fact object.
Atomic trust scores can also be directly used by the information consumer or can be further processed by aggregation hierarchy process 550 to generate composite trust scores. The various atomic trust factors (factors 510 to 540) are fed into aggregation hierarchy process 550 which generates composite scores using one or more of the atomic trust factors as inputs (e.g., term related factors, data profiling factors, data lineage factors, and security factors, etc.). In addition, weighting factors 560 can be applied to the composite factor and/or one or more of the underlying atomic trust factors. For example, when an organization is initially running aggregation process 550, a weighting factor might be applied so that if a particular atomic factor is not resolved or is resolved poorly, the factor is still used in the aggregation hierarchy. One example could be multiplying a particular atomic trust factor by a weighting factor so that the lack of reliability (trust) in the atomic trust factor does not overly reduce the resulting composite score 570. However, when the data is more settled, this weighting factor can be changed in order to highlight the issue regarding the particular atomic trust factor so that the underlying trustworthiness of the data is addressed.
One embodiment for using weighting factors 560 is using a weight default configuration. Coarse- or fine-grained weighting can be applied. An example of course-grained weighting, would be “high” (H), “medium” (M), “low” (L), and not applicable (N/A) with corresponding fine-grained weighting being 7 to 9 for “high,” 4 to 6 for “medium,” 1 to 3 for “low,” and 0 for “not applicable.” A high weighting for a trust factor would imply that the associated trust factor (and its score) are of highest importance in the project. In one embodiment, the default mapping of the coarse grained weight ‘high’ maps to the fine-grained weight ‘8’. A medium weighting for a trust factor would imply that the associated trust factor (and its score) are of medium importance in the project, while a low weighting for a trust factor would imply that the associated trust factor (and its score) are of least importance in the project. A weighing of “not applicable” (N/A or ‘0’) implies that the trust factor (and the associated score) do not apply to the specific project.
The result of aggregation hierarchy process 550 is one or more composite and/or atomic trust scores 570. As shown, one or more composite trust scores can be used as inputs to aggregation hierarchy process 550 to create additional levels of composite trust scores. A composite score is calculated from a set of elementary scores. The elementary scores can either be atomic, i.e. calculated from facts, or they can be composite scores themselves. Weights can be specified for scores which can be used by algorithms to calculate composite scores. In one embodiment, defining composite scores in a multi level hierarchy follows a logical grouping of trust factors. For example, some trust factors might be related to a column of a table in a database. Those trust factors could be aggregated into a single column score for a particular column. All the column scores for a particular table could then be aggregated into a table score for that table. And then all the table scores of a database could be aggregated into the score for the database.
One type of an aggregate score is a “weighted average” score:
Another type aggregate score is a Minimum Score for High Priority Elementary Factors where:
Name of Term (600)
Stewardship Definition for Term (610)
Short Description of Term (620)
Long Description of Term (630)
Usage Definition of Term (640)
Example Definition for Term (650)
Status of Term (660)
Association to Physical Data Model (670)
Industry Model Association of Term (675)
Currency of a Term (680)
Other Factors (690)—other term related trust factors that can be defined as needed by the organization.
Date Types Validity (711):
Length Validity (712):
Precision Validity (713):
Scale Validity (714):
Nullability Validity (715):
Cardinality/Uniqueness Validity (716):
Cardinality/Constant Values Validity (717):
Also within atomic data profiling trust factors are domain analysis trust factors 720. Domain analysis trust factors 720 are shown with two examples (721 and 722). Domain analysis trust factors examples are shown below along with various descriptions, constraints, facts, algorithms, and priorities:
Value Completeness (721):
Value Validity (722):
One of the column-related atomic trust factors that falls within atomic data profiling trust factors are format analysis trust factors 730.
Column-based trust factors are atomic and related to a physical column in a data store, such as a database. Format analysis trust factors 730 are shown with two examples (731 and 732). Format analysis trust factor examples are shown below along with various descriptions, constraints, facts, algorithms, and priorities:
Data Format Violations (731):
Data Format Variety (732):
Another set of column-related atomic trust factors that falls within atomic data profiling trust factors are key analysis trust factors 740. Key analysis trust factors 740 are related to database keys and are shown with four examples (741 to 744). Key analysis trust factor examples are shown below along with various descriptions, facts, and algorithms:
Primary Key Definition (741):
Primary Key Uniqueness (742):
Foreign Key Definition (743):
Referential Integrity (744):
The next set of column-related atomic trust factors that falls within atomic data profiling trust factors are contextual reference trust factors 810. Contextual reference trust factor examples are shown below along with various descriptions, constraints, facts, examples, and algorithms:
Richness of Semantic Definition in the Database (811):
Population to Confirm Validity (812):
The next set of column-related atomic trust factors that falls within atomic data profiling trust factors are process related trust factors 820. Process related trust factor examples are shown below along with various descriptions, constraints, facts, examples, and algorithms:
Currency of Column Analysis (821):
Analysis Rate of Column Analysis (822):
Review Rate of Column Analysis (823):
Actual vs. Sample Data (824):
Data type stability trust factors 830 is another set of atomic trust factors that falls within atomic data profiling trust factors 520. Three examples of data type stability trust factors are shown below along with various descriptions, facts, and algorithms:
Format Stability (831):
Domain/Completeness Stability (832):
Domain/Validity Stability (833):
Other Data Profiling Factors (840)—other data profiling trust factors can be defined as needed by the organization.
Three examples are included within Data Capture trust factors 920. These examples include actor (921) identifying who captured/gathered the data, mechanism/process (922) identifying the mechanism or process used to capture the data, validation (923) identifying validation performed during capture (e.g. data entered as provided or also validated against rules, etc.).
Eight examples are included within Lineage Path trust factors 930. These include degree of formalism (931, e.g. undocumented proprietary code vs. formally defined, documented, tested, controlled . . . ETL jobs, etc.); degree of transformation (932, e.g. the more data is transformed, the more likely it is that defects are introduced, etc.); length of path (933, e.g. the longer the path—meaning the number of data stores and linkages between them—the more likely it is that there will be issues, etc.); degree of auditing and monitoring (934); definition of a system-of-record (935); exception processing (936, e.g. what happens to exceptions, etc.); end-to-end delivery time (937, e.g., the currency of information, etc.); and control of the data along the lineage path (938, governance and security of the data, etc. also related to security factors shown in
At predefined process 1160, facts and algorithms that are used to calculate atomic trust factor scores are identified and stored (see
At predefined process 1180, composite metadata is identified. As previously described, composite metadata can use atomic metadata scores, facts, and other composite metadata in order to generate composite metadata (shown stored in composite metadata data store 1185.
A determination is made as to whether there are more scope items (e.g., database tables, files, etc.) to identify and process (decision 1190). If there are more scope items to identify and process, then decision 1190 branches to “yes” branch 1192 which loops back to identify the next scope item and process it in conformance with the steps outlined above. This looping continues until there are no further scope items that the organization wishes to identify and process, at which point decision 1190 branches to “no” branch 1194 and setup processing ends at 1195. Of course, the organization can choose to return to the processing shown in
At step 1240, the first algorithm that is needed to calculate the selected atomic metadata is selected from available algorithms data store 1165. In one embodiment, multiple algorithms exist for some atomic metadata and in step 1240 the user selects the algorithm most appropriate for the organization. If no algorithm currently exists, or the existing algorithms are not appropriate for the organization, then the user can create a new algorithm or modify an existing algorithm so that it works for the organization. A determination is made as to whether additional algorithms are needed to calculate the selected atomic metadata (decision 1250). If additional algorithms are needed, then decision 1250 branches to “yes” branch 1252 which loops back to select the next algorithm. As shown, the selected one or more algorithms are stored in calculated atomic metadata data store 1175. When no more algorithms are needed to calculate the selected atomic metadata, then decision 1250 branches to “no” branch 1256.
At step 1260, the first threshold that is needed to calculate the selected atomic metadata is selected from available thresholds data store 1170. In one embodiment, multiple thresholds exist for some atomic metadata and in step 1260 the user selects the threshold(s) most appropriate for the organization. If no thresholds currently exists, or the existing thresholds are not appropriate for the organization, then the user can create a new threshold or modify an existing threshold so that it works for the organization. A determination is made as to whether additional thresholds are needed to calculate the selected atomic metadata (decision 1270). If additional thresholds are needed, then decision 1270 branches to “yes” branch 1272 which loops back to select the next threshold. As shown, the selected one or more thresholds are stored in calculated atomic metadata data store 1175. When no more thresholds are needed to calculate the selected atomic metadata, then decision 1270 branches to “no” branch 1274 and processing returns to the calling routine (e.g.,
A determination is made as to whether the user wishes to modify an existing available algorithm selected from available algorithms 1165 or create a new algorithm (decision 1350). If the user wishes to modify or create a new algorithm, then decision 1350 branches to “yes” branch 1355 whereupon, at step 1360, weightings that should be applied to the algorithm (if any) are selected by the user and, at step 1365, the new/modified algorithm is created and stored along with any weightings that are used with the algorithm. The algorithm is stored in custom hierarchical algorithms data store 1340. On the other hand, if the user does not wish to create a new or modified hierarchical algorithm, then decision 1350 branches to “no” branch 1370 bypassing steps 1360 and 1365.
At step 1375, the composite metadata that was constructed using steps 1305 to 1365 is stored in composite metadata data store 1185 in trust index repository 320. The composite metadata is stored along with identification data (name/identifier/description of the composite metadata), the algorithm itself, and the facts and trust factors used in the algorithm. A determination is made as to whether the user wishes to configure additional composite metadata (decision 1380). If the user wishes to create additional composite metadata, then decision 1380 branches to “yes” branch 1385 which loops back to identify the next composite metadata and gather the data needed to generate the newly identified metadata. This looping continues until the user does not wish to identify further composite metadata, at which point decision 1380 branches to “no” branch 1390 and processing returns to the calling routine (e.g.,
At predefined process 1450, events and actions that are based on facts and metadata (both atomic and composite) are established (see
At step 1550, atomic trust factor data is gathered based on the selected fact and the selected atomic trust factor. This data includes the name of the atomic trust factor, the score/value (or algorithm), and the priority (or weighting) to apply to the atomic trust factor. The gathered data is stored in atomic trust factors metadata (data store 1560). At step 1570, the trust factor that was stored in data store 1560 is associated with the selected fact that was stored in data store 1525. A determination is made as to whether there are additional trust factors for this fact category or type of fact (decision 1580). If there are more trust factors to identify, then decision 1580 branches to “yes” branch 1582 which loops back to select the next trust factor and process the selected trust factor accordingly. This looping continues until there are no more trust factors for this type of fact, at which point decision 1580 branches to “no” branch 1586.
A determination is made as to whether there are more facts to identify and process from within selected scope items 1120 (decision 1590). If there are more facts to process, then decision 1590 branches to “yes” branch 1592 which loops back to select the next fact at step 1510 and process it accordingly. This looping continues until there are no further facts to process, at which point decision 1590 branches to “no” branch 1594 and returns to the calling routine (e.g.,
Retuning to
A determination is made as to whether there are more column-based trust factors to select and use for analysis for the selected column of data (decision 1670, see other column-based trust factors in
A determination is made as to whether there are more columns of data to select and analyze using one or more of the column-based trust factors (decision 1680). If there are more columns of data to select, then decision 1680 branches to “yes” branch 1685 which loops back to select the next column of data from fact data 1620 (selected scope items 1120 and selected facts 1525). This looping continues until there are no more columns of data that the user or organization wishes to select and analyze, at which point decision 1680 branches to “no” branch 1690 and processing returns to the calling routine (e.g.,
After the algorithms have been sorted in order to account for any dependencies as described above, at step 1740 the first algorithm is selected from order algorithms data store 1730. The first algorithm selected, as described above, is likely an atomic algorithm that uses facts and atomic metadata in the algorithm rather than using composite scores in the algorithm. At step 1750, the facts and trust metadata needed to compute the selected algorithm are identified. At step 1760, the identified facts and trust metadata are selected from trust index repository 320. As shown, trust index repository includes several data stores including selected facts data store 1525 (facts identified for the organization), column-related atomic trust factors 1660, atomic trust factors 1150, and algorithm results 1775 (results of both atomic algorithms and composite algorithms). At step 1770, the selected algorithm is performed and the results are stored in algorithm results 1775. As previously described, various types of results can be stored in data store 1775 ranging from the actual value of the algorithm that is not set to a particular scale, a coarse-grained result (e.g., “green,” “yellow,” “red,” etc.), and/or a fine-grained result (e.g., a range of 1-9 or 1-100 where higher values (e.g., 7-9 or 70-100) indicate that the data can be trusted, whereas lower values (e.g., 1-3 or 1-30) indicate that the data cannot be trusted.
A determination is made as to whether there are more algorithms to execute that are stored in ordered algorithms data store 1730 (decision 1780). If there are more algorithms to process, then decision 1780 branches to “yes” branch 1782 which loops back to select and process the next algorithm from ordered algorithms data store 1730. This looping continues until there are no more algorithms to process, at which point decision 1780 branches to “no” branch 1786 and processing returns to the calling routine (see
A determination is made as to whether to use a preset range for the selected item (decision 1810). Preset range types include the green-yellow-red type as well as other range types that may be understood by employees of a particular organization. If a preset range type is being used, then decision 1810 branches to “yes” branch 1815 whereupon, at step 1820, the user selects a range type to use from preset range types data store 1825. On the other hand, if the user does not wish to use a preset range type, then decision 1810 branches to “no” branch 1830 whereupon, at step 1835 the user creates a custom range type and at step 1840 the new range type is stored in preset range types data store 1825 so that the newly created range type will be available for future selections.
After a range type has been selected (either a preset or custom range type), then a determination is made as to whether to use preset range values for the range type or custom range values (decision 1845). If a preset range type is being used, then decision 1845 branches to “yes” branch 1848 whereupon, at step 1850, the range values to apply to the selected range type are selected from preset range values data store 1870. For example, if the range type includes three context values (e.g., green-yellow-red) then a possible set of range values would be 7-9 for green, 4-6 for yellow, and 1-3 for red. Likewise, if the range type includes four context values (e.g., “great-good-marginal-poor”), then a possible set of range values would be 8-9 for great, 6-7 for good, 4-5 for marginal, and 1-3 for poor. On the other hand, if the user does not wish to use preset range values, then decision 1845 branches to “no” branch 1852 whereupon, at step 1855, the user provides custom range values for the range type and, at step 1860, the custom range values are stored in data store 1870. For example, if using the green-yellow-red range type but the user wants more stringent values than the default 7-9 for green, 4-6 for yellow, and 1-3 for red, the user can set up more stringent values (e.g., 9 for green, 7-8 for yellow, and 1-6 for red).
At step 1875, the new context range assignment is stored in context range assignment data store 1880 and is associated with the item that was selected back at step 1805. After being established, the context value can now be used in analyses, queries, to monitor events, and the like. In other words, a user can request to be notified when a particular trust index value is “green” rather than having to specify a numeric value. A determination is made as to whether the user wishes to setup more context ranges for additional items (decision 1885). If more context ranges are being established for other items, then decision 1885 branches to “yes” branch 1888 which loops back to select the next item and process the next item. This looping continues until no more context ranges are being setup for items, at which point decision 1885 branches to “no” branch 1892 whereupon processing returns to the calling routine (see
Event notification processing commences at 1940 whereupon, at step 1945, the first event that is being monitored is selected from events and actions data store 1925. At step 1950, the associated fact(s) and/or metadata that correspond to the selected event are selected and the current values of these facts/metadata are retrieved. At step 1955, the retrieved facts/metadata is/are compared with the thresholds retrieved from the selected event from data store 1925. A determination is made as to whether, based on the comparison, the threshold(s) in the selected event have been triggered (decision 1960). If the event has been triggered, then decision 1960 branches to “yes” branch 1962 whereupon, at step 1965, one or more actions that were stored in events and actions data store 1925 are performed (e.g., notifying a user using an email message, etc.). On the other hand, if the thresholds have not been triggered, then decision 1960 branches to “no” branch 1968 bypassing step 1965.
A determination is made as to whether there are more events that are being monitored in events and actions data store 1925 (decision 1970). If there are more events that are being monitored, then decision 1970 branches to “yes” branch 1972 which loops back to select and process the next event from events and actions data store 1925. This looping continues until all events have been processed, at which point decision 1970 branches to “no” branch 1975. At step 1980, processing waits for an amount of time (e.g., five minutes, one day, etc.). At step 1990, processing resets to the first event and loops back to re-process events and actions data store 1925 starting with the first event.
If the user requests further information, then decision 2040 branches to “yes” branch 2045 whereupon, at step 2050, the drill-down request is received from the information consumer about one or more previously returned facts or metadata. At step 2060, facts and other metadata that correspond to the metadata selected by the user are retrieved (e.g., the metadata/facts that were used to compute a composite metadata score, etc.). At step 2070, these underlying facts and metadata are returned to information consumer 360. Processing loops back to decision 2040 to see if the information consumer has further drill down requests. This looping continues as long as the information consumer has further drill down requests. When the information consumer has no further drill down requests, then decision 2040 branches to “no” branch 2075 and processing returns to the calling routine (see
One of the preferred implementations of the invention is a client application, namely, a set of instructions (program code) or other functional descriptive material in a code module that may, for example, be resident in the random access memory of the computer. Until required by the computer, the set of instructions may be stored in another computer memory, for example, in a hard disk drive, or in a removable memory such as an optical disk (for eventual use in a CD ROM) or floppy disk (for eventual use in a floppy disk drive), or downloaded via the Internet or other computer network. Thus, the present invention may be implemented as a computer program product for use in a computer. In addition, although the various methods described are conveniently implemented in a general purpose computer selectively activated or reconfigured by software, one of ordinary skill in the art would also recognize that such methods may be carried out in hardware, in firmware, or in more specialized apparatus constructed to perform the required method steps. Functional descriptive material is information that imparts functionality to a machine. Functional descriptive material includes, but is not limited to, computer programs, instructions, rules, facts, definitions of computable functions, objects, and data structures.
While particular embodiments of the present invention have been shown and described, it will be obvious to those skilled in the art that, based upon the teachings herein, changes and modifications may be made without departing from this invention and its broader aspects. Therefore, the appended claims are to encompass within their scope all such changes and modifications as are within the true spirit and scope of this invention. Furthermore, it is to be understood that the invention is solely defined by the appended claims. It will be understood by those with skill in the art that if a specific number of an introduced claim element is intended, such intent will be explicitly recited in the claim, and in the absence of such recitation no such limitation is present. For non-limiting example, as an aid to understanding, the following appended claims contain usage of the introductory phrases “at least one” and “one or more” to introduce claim elements. However, the use of such phrases should not be construed to imply that the introduction of a claim element by the indefinite articles “a” or “an” limits any particular claim containing such introduced claim element to inventions containing only one such element, even when the same claim includes the introductory phrases “one or more” or “at least one” and indefinite articles such as “a” or “an”; the same holds true for the use in the claims of definite articles.
Li, Chung-Sheng, Milman, Ivan Matthew, Wolfson, Charles Daniel, Sauter, Guenter Anton, Smith, Harald Clyde
Patent | Priority | Assignee | Title |
10061842, | Dec 09 2014 | International Business Machines Corporation | Displaying answers in accordance with answer classifications |
10089409, | Apr 29 2014 | Microsoft Technology Licensing, LLC | Event-triggered data quality verification |
10459607, | Apr 04 2014 | Microsoft Technology Licensing, LLC | Expandable application representation |
10877955, | Apr 29 2014 | Microsoft Technology Licensing, LLC | Using lineage to infer data quality issues |
10885025, | Nov 05 2014 | International Business Machines Corporation | Answer management in a question-answering environment |
11106710, | Dec 09 2014 | International Business Machines Corporation | Displaying answers in accordance with answer classifications |
8577993, | May 20 2011 | International Business Machines Corporation | Caching provenance information |
8909737, | May 20 2011 | International Business Machines Corporation | Caching provenance information |
9256656, | Aug 20 2013 | International Business Machines Corporation | Determining reliability of data reports |
9338137, | Feb 13 2015 | AO Kaspersky Lab | System and methods for protecting confidential data in wireless networks |
9565196, | Nov 24 2015 | International Business Machines Corporation | Trust level modifier |
9635058, | Nov 24 2015 | International Business Machines Corporation | Trust level modifier |
9654514, | Nov 24 2015 | International Business Machines Corporation | Trust level modifier |
9679051, | Nov 05 2014 | International Business Machines Corporation | Answer sequence evaluation |
9703984, | Oct 06 2014 | MARI LLC | One way and two way data flow systems and methods |
9720963, | Nov 05 2014 | International Business Machines Corporation | Answer category data classifying using dynamic thresholds |
9769293, | Apr 10 2014 | Microsoft Technology Licensing, LLC | Slider cover for computing device |
9841874, | Apr 04 2014 | Microsoft Technology Licensing, LLC | Expandable application representation |
9946747, | Nov 05 2014 | International Business Machines Corporation | Answer category data classifying using dynamic thresholds |
Patent | Priority | Assignee | Title |
5191348, | Mar 04 1992 | Radar detector performance verification method and apparatus | |
6480837, | Dec 16 1999 | LinkedIn Corporation | Method, system, and program for ordering search results using a popularity weighting |
6611825, | Jun 09 1999 | The Boeing Company | Method and system for text mining using multidimensional subspaces |
6701305, | Jun 09 1999 | Boeing Company | Methods, apparatus and computer program products for information retrieval and document classification utilizing a multidimensional subspace |
6836463, | Oct 15 1999 | NOKIA SOLUTIONS AND NETWORKS OY | System for communicating labeled routing trees to establish preferred paths and source routes with local identifiers in wireless computer networks |
7240213, | Mar 15 2002 | WATERS EDGE CONSULTING, LLC | System trustworthiness tool and methodology |
7263607, | Jun 12 2003 | Microsoft Technology Licensing, LLC | Categorizing electronic messages based on trust between electronic messaging entities |
7272719, | Nov 29 2004 | KIP Sign P1 LP | Method to control access between network endpoints based on trust scores calculated from information system component analysis |
7480931, | Jul 24 2004 | CORECO IDERA OPS, INC | Volume mount authentication |
7904727, | Nov 29 2004 | KIP Sign P1 LP | Method to control access between network endpoints based on trust scores calculated from information system component analysis |
20020144149, | |||
20030167308, | |||
20040128544, | |||
20050283468, | |||
20060168022, | |||
20060212925, | |||
20060212930, | |||
20060230021, | |||
20060294151, | |||
20070106577, | |||
20070156514, | |||
20070156604, | |||
20070156621, | |||
20070156887, | |||
20070180495, | |||
20080015916, | |||
20080107037, | |||
20080209543, | |||
20080261701, | |||
20090006230, | |||
20090024589, | |||
20090235334, | |||
JP1528498, | |||
JP2005135071, | |||
JP2005322089, | |||
JP2007041835, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Aug 24 2008 | WOLFSON, CHARLES DANIEL | International Business Machines Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 021739 | /0869 | |
Aug 24 2008 | SMITH, HARALD CLYDE | International Business Machines Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 021739 | /0869 | |
Aug 24 2008 | SAUTER, GUENTER ANTON | International Business Machines Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 021739 | /0869 | |
Aug 24 2008 | MILMAN, IVAN MATTHEW | International Business Machines Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 021739 | /0869 | |
Aug 24 2008 | LI, CHUNG-SHENG | International Business Machines Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 021739 | /0869 | |
Oct 23 2008 | SAUTER, GUENTER ANTON | International Business Machines Corporation | CORRECTIVE ASSIGNMENT TO CORRECT THE CORRECT THE EXECUTION DATE PREVIOUSLY RECORDED AT REEL: 021739 FRAME: 0869 ASSIGNOR S HEREBY CONFIRMS THE ASSIGNMENT | 050450 | /0427 | |
Oct 24 2008 | International Business Machines Corporation | (assignment on the face of the patent) | / | |||
Oct 24 2008 | WOLFSON, CHARLES DANIEL | International Business Machines Corporation | CORRECTIVE ASSIGNMENT TO CORRECT THE CORRECT THE EXECUTION DATE PREVIOUSLY RECORDED AT REEL: 021739 FRAME: 0869 ASSIGNOR S HEREBY CONFIRMS THE ASSIGNMENT | 050450 | /0427 | |
Oct 24 2008 | SMITH, HARALD CLYDE | International Business Machines Corporation | CORRECTIVE ASSIGNMENT TO CORRECT THE CORRECT THE EXECUTION DATE PREVIOUSLY RECORDED AT REEL: 021739 FRAME: 0869 ASSIGNOR S HEREBY CONFIRMS THE ASSIGNMENT | 050450 | /0427 | |
Oct 24 2008 | MILMAN, IVAN MATTHEW | International Business Machines Corporation | CORRECTIVE ASSIGNMENT TO CORRECT THE CORRECT THE EXECUTION DATE PREVIOUSLY RECORDED AT REEL: 021739 FRAME: 0869 ASSIGNOR S HEREBY CONFIRMS THE ASSIGNMENT | 050450 | /0427 | |
Oct 24 2008 | LI, CHUNG-SHENG | International Business Machines Corporation | CORRECTIVE ASSIGNMENT TO CORRECT THE CORRECT THE EXECUTION DATE PREVIOUSLY RECORDED AT REEL: 021739 FRAME: 0869 ASSIGNOR S HEREBY CONFIRMS THE ASSIGNMENT | 050450 | /0427 | |
Sep 30 2019 | International Business Machines Corporation | DAEDALUS GROUP LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 051032 | /0784 | |
Dec 30 2019 | International Business Machines Corporation | DAEDALUS GROUP, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 051710 | /0445 | |
Jan 28 2020 | DAEDALUS GROUP, LLC | Daedalus Blue LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 051737 | /0191 | |
Nov 28 2023 | Daedalus Blue LLC | Taiwan Semiconductor Manufacturing Company, Limited | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 066749 | /0668 |
Date | Maintenance Fee Events |
Dec 23 2016 | REM: Maintenance Fee Reminder Mailed. |
May 11 2017 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
May 11 2017 | M1554: Surcharge for Late Payment, Large Entity. |
Nov 02 2020 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Nov 14 2024 | M1553: Payment of Maintenance Fee, 12th Year, Large Entity. |
Date | Maintenance Schedule |
May 14 2016 | 4 years fee payment window open |
Nov 14 2016 | 6 months grace period start (w surcharge) |
May 14 2017 | patent expiry (for year 4) |
May 14 2019 | 2 years to revive unintentionally abandoned end. (for year 4) |
May 14 2020 | 8 years fee payment window open |
Nov 14 2020 | 6 months grace period start (w surcharge) |
May 14 2021 | patent expiry (for year 8) |
May 14 2023 | 2 years to revive unintentionally abandoned end. (for year 8) |
May 14 2024 | 12 years fee payment window open |
Nov 14 2024 | 6 months grace period start (w surcharge) |
May 14 2025 | patent expiry (for year 12) |
May 14 2027 | 2 years to revive unintentionally abandoned end. (for year 12) |