security rating method and a graphical tool for decision making especially for setting priorities which tool comprises at least one diagram (1) having a first scale (2) which reaches in a first direction along a first axis (3), at least a first (4) and a second (5) potential action fields the identified value of which (4a, 4b) is graphically represented on said first scale (2) in said diagram (1) whereby a second scale (6) is arranged along said first axis. Said second scale (6) reaches in an opposite direction to said first scale.

Patent
   7945940
Priority
Sep 21 2000
Filed
Sep 21 2001
Issued
May 17 2011
Expiry
May 09 2024
Extension
961 days
Assg.orig
Entity
Small
2
3
EXPIRED
1. A method for measuring a numerical information security level at an organization for standardized comparison to numerical information security levels of other organizations, the method comprising:
performing measurements on specific information security issues within three different security areas of the organization comprising the security areas logical information security, system information security, and physical information security,
wherein each area is divided into a subset of categories, each category is divided into a subset of checkpoints, and each checkpoint relates to a specific information security issue,
wherein said specific information security issues within the area logical information security comprise guide lines and policies,
wherein said specific information security issues within the area system security comprise technical security of IT-systems, and
wherein said specific information security issues within the area of physical information security comprise burglar alarm and fire alarm;
treating the specific information security issues as checkpoints;
using various sources, including current standards, to gather an information set to state a predefined norm security level;
using the norm security level to create different lower and higher numerical security levels within each checkpoint;
using the measurements and having a rating reviewer give a numerical grade to each checkpoint;
using mathematical operations to calculate a result based on the numerical checkpoint grades; and
using the result to indicate which numerical information security level is fulfilled, wherein the result presents a relative rating regarding information security within each area that is comparable between different organizations and also presents comparable ratings regarding different checkpoints within the organization,
wherein a software tool is used to graphically present the result,
wherein the graphic presentation includes a graphical tool for decision making,
wherein the graphical tool comprises at least one diagram presenting:
a first scale that reaches in a first direction along a first axis,
at least two presentations of the numerical checkpoint grades of a first checkpoint and a second checkpoint respectively, wherein the numerical checkpoint grades are graphically represented along the first scale, and
a second scale arranged along the first axis, wherein said second scale reaches in a direction opposite to the first scale and represents a relative value of input needed to at least one of
increase the grade of each presented checkpoint, and
reach a certain discrete numerical information security level.
2. The method of claim 1, wherein the information set further comprises at least one of guidelines and recommendations regarding security, and wherein the information set is used to set the numerical checkpoint grades.
3. The method of claim 2, wherein the numerical checkpoint grades are presented on an analogue scale, whereby the result is also presented on an analogue scale.
4. The method of claim 2, wherein results from different security areas are compared to evaluate which area needs attention.
5. The method of claim 1, wherein the mathematical operations comprise adding up the numerical checkpoint grades from each checkpoint.
6. The method of claim 1, wherein the numerical information security levels are used to meet the need for different types of information security levels and requirements, to enable one organization type to choose a security level different from another organization type depending on under which regulations, legislation, or branch of trade the one organization type is operating.
7. The method of claim 1, wherein the numerical information security levels are discrete and represent the level of demands that are put onto an organization, both in the amount of demands and the level of demand fulfillment.
8. The method of claim 7, wherein the number of discrete numerical information security levels is at least five.
9. The method of claim 7, wherein the number of numerical checkpoint grades is at least five.
10. The method of claim 7, wherein all demands must be completely fulfilled to pass a certain discrete checkpoint grade.
11. The method of claim 1, wherein a final result is calculated only after having set all numerical checkpoint grades within every security area being measured.
12. The method of claim 1, wherein a checkpoint is represented by a property of an information security issue being measured.
13. The method of claim 11, wherein calculating the final result comprises:
applying a weight to the numerical checkpoint grades to calculate weighted checkpoint results;
calculating, for each area, a weighted area result based on the weighted checkpoint results; and
calculating the final result based on each weighted area result.

This application is a national stage application of international Application No. PCT/EP01/10970, filed Sep. 21, 2001, and claims the benefit of U.S. Provisional Patent Application Ser. No. 60/290,593, filed May 11, 2001, and Sweden Application No. 0003378-7, filed Sep. 21, 2000, all of which are herein incorporated by reference in their entirety.

Traditionally, when mentioning computer security, IT-security has been the most commonly used concept. When we say IT-security we usually mean technical aspects of computer security. Up to this date it is traditionally where most of the work has been concentrated the past few years.

By stating the notion Information Security we actually mean a larger part of the computer security. Information Security does not only cover the technical aspects, but also the administrative and physical aspects. The administrative aspects cover the routines and documentation while the physical aspects cover the buildings and storage of the computers and their information. The Information Security notion cover more areas and more than just the technical aspects.

Everybody knows what good Information Security level is, or at least we have an idea about what the market says is a good level. The level of Information Security that we adopt today is usually based on what we have to protect. If we have a lot of classified information, we have demands of a high Information Security level. If our information is open to everybody, we are usually satisfied with a lower level. But how can we know which level that is appropriate?

What is needed, is a way to express this level of Information Security and invent a new way of factors and definitions, a “common language”. A company A which is going to connect their network to company B needs a way to know that their Information Security level is similar, or at least similar in some specific areas. We need a way to measure the Information Security level at companies A and B, and a method to compare the levels. Company A and B may be operating under different regulations or different legislation, depending on their branch of trade.

We also need a method to cover this situation. Companies A and B will be able to use the method according to the invention to communicate their needs between each other.

But how can Information Security levels be measured? To measure using statistics is one way by sampling values and construct definitions. This is difficult because there is nothing to compare against. The other way to measure is using a relative method. We can say that one protection or implementation is better than another. To use a relative method to measure Information Security forms the basis of the method for rating according to the present invention.

The relative method according to the invention is based on performing measurements on specific Information Security issues within an organisation. These issues are called checkpoints. All checkpoints are divided into three information security areas: logical, physical and system security. A fourth area could be added, cf. below. Logical security cover administrative aspects like polices and guidelines with a perspective on the personnel. Physical security cover all the facilities while system security cover the IT-systems themselves.

The method according to the invention states a norm level for Information Security which is based on the current market situation. This norm is formed by current standards and recommendations for Information Security. The various information sources which are used to construct the norm level within the method is together called the information set. All demands and appliances are derived from the information set to form the norm level.

From the norm level, four other levels are created in the described example. This is made to meet the need for different types of information security levels and requirements. Two lower levels are created from the norm which generally have a smaller amount of checkpoints and the demands are less strict. Above the norm two higher levels are created which generally have a larger amounts of checkpoints and the demands are stricter. Measurements can be performed towards each level by doing analysis and inspections on the three information security areas.

During the analysis and inspections each checkpoint is given a grade. The grade is given on scale which generally indicate the fulfillment towards the demand on the specific checkpoint. When each checkpoint has been given a grade several results are calculated. The mathematical operations to calculate the results consider several organisation properties. To consider the properties when calculating the results within the method is vital to be able to use the results between organisations. The consideration to properties ensure that each result is comparable between organisations.

If the results indicate that the organisation has passed the analysis, a certification organisation has the right to issue a certificate. The certificate will state which level that is passed and the results in each area.

The described method will be developed into new versions to meet the ever changing needs and demands within the information security area. A new version will change demands, add checkpoints and reconstruct the analysis. When to issue a new version is governed by the availability of new sources if information like new legislation, new security techniques, a change of social behaviours etc.

The analysis according to the method is supported by a software tool. This tool can also perform all the result calculations and present the results. Organisations may use the tool to perform the analysis themselves, but in that case the result calculations are locked out. This is because the method must be controlled by one party to insure that results are coherent and thus possible to compare.

The rating method according to the invention will now be described more in detail. First certain definitions and abbreviations will be elucidated. Then follows a part describing the results obtained by means of the method according to the invention.

Definition of Roles

A certain terminology is used in the description of the method, e.g. for defining organisation specific parts and their context. There are also definitions used for the organisation performing the measurement.

Entity The actual organisation that is subjected to the
information security measurement
Entity Owner This is the overall responsible function for
Information security at the Entity or within a group
of entities that shares information
Rating The Independent organisation that conducts the rating
Reviewer measurement process. This is also the license company
that gets the license to use the method, license taker.
(ref. To as Reviewer)
Reviewer The Independent organisation team which always
team should at least consist of three persons. Each
person should have his or her major knowledge within
some of the three information security areas.
License provides the license and also has the basis for determine
Provider the final results (result method part).
Client The person/function at the Entity owner that has ordered
the rating.

Method Areas—Definition of Information Security

Within the method, three areas are defined on which to perform the measurements. These areas are:

Logical Administrative information security like policies, guidelines etc.
Security Logical security focuses on personnel and documentation
within the organisation measured.
Physical Information security regarding facilities, fire and burglar
Security protection etc. Physical security focuses on the personnel as
well as the facilities within the organisation measured.
System Information security regarding network, operating systems and
Security applications. System security focuses on personnel, and the
level of IT-systems' technical security within the organisation
measured.

Definition of Measurement Levels

The basis and uniqueness for the method are the pre-set levels of information security. The levels are derived form a predefined scale and a norm set as level 3. A higher number on the scale equals a larger set of information security analysis. A higher number on the scale also represents a higher set of demands to be fulfilled. Present levels of the concept using the method are:

Veriscan 1 Controlled Information Security
Veriscan 2 Information Security Aware
Veriscan 3 Market adapted Information Security (current market norm)
Veriscan 4 High Information Security
Veriscan 5 Very High Information Security

Demands are relatively set from a standard level (norm). The middle level within the total number of levels represents the norm. From the norm level, lower levels and higher levels are set. In the five levels shown above in this example, the norm is based at level 3 (Veriscan 3). The norm is derived from demands set by dependant and independent organisations through standards, guidelines and policies. Other market demands within the information security area also have impact on the construction of the norm level. The middle level will always be named “Market adopted Information Security” to reflect the level of demands that are put onto organisations, both in amount of demands and the level of demand fulfilment.

The statement above provides that the levels described always must be developed to meet the market demands within the Information Security Area. This also means that each level must have it's own version number of method used. The version number describes future developments of the method.

Definition of Information Set

In order to set and develop the different levels information is gathered from various sources. In this example the following is a list of such sources:

Several of the information sources mentioned above contain information that is similar or identical to information from other sources on the list. Several books include the topics covered within standard from another point of view. Information about information security within literature has more often a more aggressive approach towards problems covered. This approach is used within the method to further develop the various version numbers of a level.

Definition of Method Terminology

Each information security area (Logical, Physical and System) is divided into measurement categories. The categories reflect certain sub areas within an information security area. A category is further divided into checkpoints. The checkpoints are the actual measurements performed by analysis, interviews, inspection or tests etc. The definitions are stated as follows:

Level The actual rating level measured, e.g. Veriscan 1 to 5
Area The three Information security areas; Logical, Physical and
System
Category A service or specific part in the organisation Information
security
Checkpoint A specific function within a category

The connections between these definitions can also be viewed in FIG. 2.a.
Definition of Method Weights

The method consists of weights used during calculations of the results in the method result part. The weights are used within the method to solve the following problems:

Which type of entity is measured?

Who is the person respondent answering questions during interviews (his or hers responsibility)?

How many of the entity personnel are going to be interviewed?

Who is the person from the license provider doing the actual measurement?

What is the current level of threats against the measured entity?

What is the consequence on a specific resource?

What is the probability that some thing happens on some information resource?

Is a checkpoint generally more important than another one?

Is a category generally more important than another one?

What are the general efforts that the entity has to mobilise to solve an information security problem?

Each weight is explained in the following.

Threat Weight:

The threat weight represents a general level of threat against the checkpoint measured within a scale from 1 to 5. A higher threat weight represents a high threat and a lower weight represent a low threat. The actual threat weight on a specific checkpoint is deduced from a generalisation performed by the license provider.

Probability Weight:

The probability weight represents a general level of probability that a certain action will be fulfilled if the threat on the checkpoint is activated on a scale of 1 to 5. A lower weight represents a low probability and high weight represents a high probability. The actual probability weight on a specific checkpoint is deduced from a generalisation performed by the license provider.

Consequence Weight:

The consequence weight represents a general level of consequence for the entity measured if both the threat and probability of checkpoint has activated on a scale of 1 to 5. A lower weight represents a low consequence and high weight represents a high consequence. The actual consequence weight on a specific checkpoint is deduced from a generalisation performed by the license provider.

Checkpoint Weight:

By adding the threat, probability and consequence weights to each other and divide with the numbers of factors used, a factor called the checkpoint weight is created. The checkpoint weight represents the importance of the checkpoint towards other checkpoints.

Category Weight:

A set of checkpoints form a category. The category weight is like the checkpoint weight built on three factors: threat, probability and consequence. There is a major difference between the checkpoint weight and the category weight in the manner that the actual category weight can be derived from two different aspects. The first type of category weight is derived from a generalisation on the three factors threat, probability and consequence. This generalisation is performed during measurement setup in discussions with the entity owner.

The other type of category weight is derived from a mean value from all the checkpoints threat, probability and consequence factors.

Which approach that is used is decided upon measurement setup between the license provider and the entity owner. Within the method these two approaches on category weight are called pre-set category weight and derived category weight.

Position Weight:

Upon measurement interviews will be performed with people who have different types of responsibility. To limit variations, that may be due to the fact that the actual sample do not reflect the organisation structure of the Entity, the Position Weight is used. Like in the previous weights a 5-graded scale is used, where 1 represents a low position weight and 5 represent a high position weight. All position weights are explained below:

The position weight is used in the method to grade answers, inspection or analysis. It's vital that all the position weights together don't give a false or weighted result. For example, a person having a high position weight will have higher impact on the checkpoint result than a person with a low position weight.

The license provider adds a restriction on position weights of the entity saying: There must always be enough lower position weights to cover a higher position weight. For example this means that if there is one person participating who has a position weight of 4, there must be at least 4 people with position weight of 1 to cover.

The example described above gives the notion that this may not always be true. There may not be enough persons to cover a higher position weight. Therefore, the method uses fictitious persons to cover. These fictitious persons are based on a mean value of the results from the other persons within the same position weight.

Time Effort Weight:

Each checkpoint has some drawback if the measurement gives a result that is below the norm, 3. To be able to measure the effort in time that the entity has to provide to correct (i.e. pass) the checkpoint from a very low state, the time effort weight is introduced into the method. The time effort weight uses a scale form 0 to 5 where a low time effort weight represent that the entity has to put aside a minimum of time to solve the problem. A high time effort weight represent that the entity has to put aside a considerably amount of time to solve the problem. The complete scale of time effort weight is presented below.

To be able to measure the effort in costs that the entity has to provide to correct (i.e. pass) the checkpoint from a very low state, the cost effort weight is introduced into the method. The cost effort weight uses a scale form 0 to 5 where a low cost effort weight represent that the entity has to put aside a minimum of funds to solve the problem. A high cost effort weight represent that the entity has to put aside a considerably amount of funds to solve the problem. The complete scale of time effort weight is presented below.

The time effort weight and the cost effort weight form the actual effort weight for a particular checkpoint. This is done by adding the participating types of effort weights to each other and then by dividing with the number of effort weights. By stating this last condition the effort weight may include further effort factors in later versions.

Entity Template:

Together the effort weight, threat, probability and consequence form several factors between 1 to 5 for each measured checkpoint. All the factors for each checkpoint form a matrix called the entity template. The entity template forms the basics for importance and the different effort between checkpoints.

The license provider form the basic entity template for each measured level. Upon measurement setup the entity owner has the right to adjust the basic entity template to meet the measured entity's needs or structure. The license provider has set a constriction on modifications on the template, which include that no factor may increase more than one and no factor may decrease more than one.

If an entity will choose to alter the basic entity template, the altered version will be used in all the sub entities within the measured organisation.

The license provider also has the possibility to alter the entity template to meet certain conditions stated for different types of entities. There may be an entity template for health care and another for process industry, for example.

In order to have an open comparison independent of variations of the actual entity should generally an altered entity template not be considered and the basic entity template should be used.

Definition of Checkpoint Grades

Each checkpoint that is measured is given a grade. The grade is set upon a specific checkpoint during interviews, inspections, system analysis or documentation analysis. The grade is constructed with a six graded scale, where 0 is the lowest grade and 5 the highest. Like in the method level norm, a norm grade is used on a specific checkpoint. This norm is also set to 3. A grade lower than 3 represent a checkpoint that is not passed, and likewise a grade higher than or equal to 3 represent a passed checkpoint. All present checkpoint grades and their meaning are stated below. The meaning of the actual grade is different depending upon the information security area measured.

Grade0 There are severe/crucial shortages or lack of information
(Failed) resources concerning:
Knowledge
Documents or document content
System implementations or configurations
Facilities, facility construction or Implementations
Grade 1 There are major shortages of Information resources
(Failed) concerning:
Knowledge
Documents or document content
System Implementations or configurations
Facilities, facility construction or Implementations
Grade 2 There are minor shortages of Information resources
(Failed) concerning:
Knowledge
Documents or document content
System Implementations or configurations
Facilities, facility construction or Implementations
Grade 3 There are no shortages of Information resources concerning:
(Passed) Knowledge
Documents or document content
System Implementations or configurations
Facilities, facility construction or Implementations
Grade 4 There are no shortages of Information resources concerning:
(Passed) (and checkpoints passed with commendation):
Knowledge
Documents or document content
System implementations or configurations
Facilities, facility construction or Implementations
Grade 5 There are no shortages of Information resources concerning
(Passed) (and checkpoints passed with high commendation):
Knowledge
Documents or document content
System implementations or configurations
Facilities, facility construction or implementations

The grades can also be seen as an analogue scale using natural numbers adapted to each information security area as follows:

##STR00001##

The definitions on how each grade is set upon a specific checkpoint will be discussed later in this document.

Definition of Checkpoint Demands

The approach of grading is achieved by stating a specific demand on each checkpoint. The demand may vary from a specific text that must exist within the controlled entity's security policy or that a specific type of door must be used in the computer room. Within the system security area for example, must the e-mail service be configured in a way that will minimise a security breach. All these conditions form together the method demands. As stated earlier the demands will vary from each information security level. All checkpoint demands are derived from the information set.

Definition of Result Grades 1 to 5

If the demand on a specific checkpoint is fulfilled, a checkpoint grade of at least 3 will be given. To give a checkpoint a grade of 3 is actually not a major problem because the demand is derived from the information set, which it's based upon standards like BS7799, SS3492 etc. To give a grade of 0 is not either a difficult task, because this grade states a total lack of information or protection. The tricky part is to grade a checkpoint either lower (1 or 2) or higher (4 or 5) than 3.

This latter part is up to the rating reviewer. It's his or hers responsibility to give a correct grade on the checkpoint. The actual grade properties have been stated earlier and those statements form the basics within the method to give a correct grade. To pass a certain checkpoint the demands must be completely fulfilled. On each demand there is also a negative and a positive aspect defined and described. These aspects will act as a grading guide on each checkpoint and help the reviewer towards giving a correct grade during the analysis.

The positive and negative aspects are stated in a generalised form. For example a checkpoint about fire protection in the computer room will at rating level of 3 have the demand stating that a fire extinguisher of the proper type should be placed easily accessible and signs have to be in place. A typical positive aspect on this checkpoint will be if the extinguisher is placed in an excellent position, it has been controlled recently and there are signs which are both obvious and of a excellent quality. A typical negative aspect will be if the extinguisher is not of the proper type (extinguish electrical fires), it's not obvious where to find it and signs may be misplaced. The amount of positive and negative aspects on a specific checkpoint will guide the reviewer to the correct grade.

But there is a potential risk that one checkpoint may be given different grades depending on who the rating reviewer is.

To meet this potential risk the method states that each checkpoint, where there are doubts must be controlled within the rating reviewed team. The method states that all checkpoints, even if there is only the slightest of doubts upon the grade, should always be reviewed by the rating reviewers team. It's then up to the team to set the final grade. By adopting this technique within the method, there is only a minor chance of faulty grades.

When all grades have been set for all three information security areas, the basis for calculating the final results are in place.

Method Development and Versions

Norm Level Development

To put all the demands and set of fulfillments into the method, the license provider determines the set of criteria forming the norm. This is done by analysis of the various sources mentioned in the previous chapter 2.2. All the analysis is performed by identifying the common parts between each information resource. The common parts are stated as the backbone for the norm level, Veriscan 3. This is done on each information security area: Logical, Physical and System.

The backbone information resource is evaluated against current market information security demands, fulfillments and actual market situation. The information set produced after evaluation form the full set of information used in the norm level, Veriscan 3.

On the full set of information the license provider, performs various tasks to group, divide and adopt the information set into the method. This step creates the areas, categories and checkpoints of the norm as described in chapter 2.2.

As time goes by new features and factors will change the global information security market. New technologies will be adopted and many security issues will be both changed and revised. For example, there will be new inventions to meet threats from hackers and crackers and the method must have a built in support to deal with these kind of situations. This means that the checkpoints, weights, demands and so on will have to change over time.

To meet the ever changing information security market the method uses a version numbering. The method version numbering uses three number stated as a.b.c. The a represents a major version/revision, the b represent the market version and the c stands for the internal working revision number. Analysis towards an entity can only be performed on a market version.

The method states that a new major version can only be released if all information security areas are revised onto the corresponding major version, i.e. 1.b, 2.b, 3.b etc. This means if some area is to be revised onto a new major version the other areas must also be revised.

The market versions are the actual versions which are used to perform analysis and measurements. These versions do not have to be at the corresponding level at each information security area. For example, it's possible to use a different market version on the physical area and another market version on the system security area. Internal working revision numbers are only used by the license provider, these revisions are never used in actual analysis or measurements.

When a new major development, a new security technique has been adopted or a major shift in social behaviour within the information security areas, the market version have to shift. This is a trigger in the method to develop a new major version. When to develop new market versions, is mainly controlled by involving new standards, guidelines or other information security information sets. A new market version must also be developed if transferring a certain security feature from a higher rating level to a lower. The whole process of developing new versions and revisions is up to the license providers ability to cover the information security market and to be up to date with the day to day information security situation.

Measurement Basic Methods

In the following is described the main tasks during the actual measurement performed towards the entity. This is a principle overview. Detailed information about what and who is set in method software program depending on the level selected.

Document Analysis

Document analysis are performed by the rating reviewer team. The reviewer uses material produced by the license provider to perform the analysis. To analyse the documentation the reviewer uses the statements within the method. Demands on context and text are derived from the information set used at the measured level, i.e. rating 1 to 5. Document analysis are performed within each information security area. The grades set upon the analysis are derived from the following:

All interviews are performed by the rating reviewer team. This is done by asking the respondent the set of questions applicable for his or hers position weight. The material used is applicable for the measured level. Interviews can be performed by a sole reviewer team member. Interviews are performed within each information security area. The grades upon interviews is always set according to:

All inspections are performed by the rating reviewer team. Inspections mainly concern facilities in the context of information security. A checklist is used for the measured level. Inspections can be performed by a sole reviewer team member. Inspections are only performed within the physical information security area. The grades upon inspections are set upon:

The IT-System analysis is performed using a software toolbox according to the actual situation of Entity IT-systems. Different modules are used in the toolbox depending on what systems are in use and also the level chosen.

Differences of Methods Depending on Level

There is also a difference to the extension of the measurements depending on which rating level that is measured. This can be seen in table 2a.

The extension of method used during the measurement is determined by which rating level that is used. Interviews will be used more extensively and they will be more comprehensive the higher the measurement level is. Documentation analysis will also be conducted more thoroughly the higher the measurement level is. Inspections will also follow this pattern. System analysis will be performed in somewhat other manner. The higher the measurement level, the amount of system checks will be extended.

Generating Method Results

To generate the actual results from the checkpoint grades the method states that several mathematical calculations will be performed. These calculations will use all the prior discussed methodology and factor like weights and templates. All different types of calculated results will be covered in this chapter.

Result Definitions and Abbreviations

Grades

CpG=Checkpoint Grade

CpWG=Checkpoint Weighted Grade

Weights

CpW=Checkpoint Weight

CaW=Category Weight

EfW=Effort Weight

Primary Results

CpR=Checkpoint Result

CpWR=Checkpoint Weighted Result

ArWR=Area Weighted Result

Secondary Results

CaR=Category Result

CaWR=Category Weighted Result

ArR=Area Result

ToR=Total Result

ToWR=Total Weighted Result

Gaps

SeG=Security Gap

PrG=Priority Gap

The Result Generating Principle

The primary results are the results used for evaluation for level fulfilment and the basis for official statement. The primary results are the CpR and the CpWR and ArWR. All other results are secondary results and are used as guides or key values in the evaluation of the results.

The primary result generating formula takes each CpR to the weighted CpWr by using the weights and transforms this to the ArWR. The level fulfillment is then based upon that ArWR is passed for all three areas; Logical, Physical and System.

Generating Method Results

The Non Weighted Checkpoint Result (Primary)

The non weighted checkpoint result (CpR) is the raw result set by the reviewer team. There are no weights used on this result and therefore the value of the result has no properties connecting it to the entity template or to the actual respondent type. This result represent if all checkpoints where equally important and if all respondents has the same information security responsibility.

The Weighted Checkpoint Result (Primary)

This result is calculated by consider different factors defined previously to affect the importance of the checkpoint. The checkpoint results are calculated based on the entity template. The weighted checkpoint result (CpWR) is the basis for the calculation of the total performance of the analysis and subsequently the results which the rating is based upon.

The Category Results (Secondary)

Category results are contraction of the results of the checkpoints that are within that category and represents a mean value. The result of a category is merely a result to seen stand alone as it is not a calculated part of the total result. The category results can be of two types:

Non Weighted Category Results (Secondary)

The raw (unweighted) category result (CaR) is generated by mean value of the non weighted Checkpoint result (CpR). These results has no correspondence to the entity template or respondent types.

Weighted Category Results (Secondary)

The weighted category result (CaWR) is based on the mean value of the weighted checkpoint results (CpWR). Thus the CpWR follows the entity template and the actual types of the respondents.

The Area Results

The area result is a contraction of the checkpoint belonging to an area and represents a mean value. An area result is a stand alone result and can be of two types.

The Non Weighted Area Result (Secondary)

The raw (unweighted) area result (ArR) is generated by all non weighted results of the checkpoints belonging to the area in question and are a mean value. These results has no correspondence to the entity template or respondents types.

The Weighted Area Result (Primary)

The weighted area result is generated by all weighted results of the checkpoints belonging to the area in question and are a mean value.

The weighted Area result (ArWR) is based on the weighted calculated results and are by that based on the entity template and the actual types of the respondents.

The Level Fulfillment (Primary)

The total result is if each of the three areas have passed a value of three. The total result is presented as “Passed” or “Failed” the rating on the actual level depending on that all areas shall have performed with grade three or higher on each the weighted area result (ArWR). To have passed no checkpoints with a high checkpoint weight shall have a result that is less than 1.

Additional information can be obtained by looking at a total result generated by the total mean values of all the checkpoint results. These additional total results can be of two types:

Non Weighted Total Result (Secondary)

Also here a raw (non weighted) total result (ToR) can be calculated based on all non weighted checkpoint results. These results has no correspondence to the entity template or the respondents types.

Weighted Total Result (Secondary)

The weighted total result (ToWR) is also calculated based on all weighted checkpoint results. This result is based upon the entity template. These results are corresponding to the entity template and the respondents types.

Gaps

The gaps are results that are distances between values set in the method and results generated by the analysis. Definitions of gaps can also be seen in FIG. 4b.

Security Gap

The security gap (SeG) is calculated for each checkpoint by comparing the checkpoint result to the norm value of three that equals the result needed to pass the demands on the checkpoint. The checkpoint result is subtracted by 3 and if the result is negative is there a security gap equal to that figure.

Priority Gap

The priority gap (PrG) is calculated by taking all checkpoint results that have a security gap and adding the security gap for the checkpoint and the difference between the actual effort weight for the checkpoint and the maximum value of five.

Method Processing

In the following will be described how to work with and implement the method. The method processing is how the results are gathered. As the method is software supported by a method software program, this is also described under this heading as well as the software toolbox for system security testing.

General Work Flow

The measurement follows a number of process steps/phases as can generally be seen in FIG. 3.a. This can also be described more in depth by a phase description. It is important to again point out that as the basis for the result is the actual performance that the Entity shows for each checkpoint, and the major part of the work is carried through by gathering and determine the performance of each checkpoint included in template.

Phase 1—Pre Study, Setting the Level Template

This phase is to determine what actual level that is preferred by the Client and requirements that the client may have. It is also to determine minor adjustments of the method, such as terminology. Further it will state what kind of documentation that is expected from the client as well as how organisation and system layout will set the parameters for interviews and system check modules. A list of Actions will be set for the Entity to fulfil.

Phase 2—Set Up of the Method Software Program

The Reviewer receives the method software program and the software toolbox. The method software program will be set with parameters for basic terminology of the Entity and data for interviews as well as system check modules.

Phase 3—Documentation Analysis

The Reviewer following the guide in the software program sets the actual measurement and results on the Checkpoints considering documentation and the results (0-5) for each Checkpoint are put in.

Phase 4—Interview Analysis

The Reviewer performs the interviews of different personnel within the entity and the results (0-5) for each Checkpoint are put into the software program.

Phase 5—Physical Inspections

Physical inspections are conducted at the premises of the Entity and results (0-5) for each Checkpoint are put into the software program.

Phase 6—System Analysis

The modules for systems tests in the software toolbox are conducted and the results (0-5) for each Checkpoint are set in the method software program.

Phase 7—Result Calculation

The result file is sent to license provider for calculation of the results according to the set up determined in the pre study phase.

Phase 8—Result

The final result is sent to the Reviewer in electronic format.

Phase 9—Report

The reviewer chooses standard report that will suit the Client directly from the software program.

Phase 10—Presentation

The reviewer will present the results to the Client. The client will also get the electronic copy.

Phase 11—Analysing the Result

The Client may reconstruct the result in a fashion that suits him and can from the electronic copy sort the result in many different ways to be able to set priorities for enhanced information security.

The process above and in FIG. 3.b is simplified somewhat and does not cover how the methods differs between different levels and/or areas.

Phase 12—Renewal of Verification

The process is a single loop process but as time will affect the security level should a renewal of the verification be necessary in order to keep the verified level, see the last step. If so the first step “Pre-study” can be much more limited and only cover differences. However if the renewal of the verification is done on a higher level a full pre-study must be carried through.

Method Software Program

The software program supports the method and contains necessary data to perform the measurement. It also serves as the basis for how to handle data between License Provider and Reviewer. The method software program generally support all phases described above.

Data Handling Between Reviewer and Licence Provider

The License Provider will send the method software program to an approved Reviewer when a Client has been contacted. The Reviewer will receive the first part contain all necessary software to be able to carry out the Pre-Study phase.

When the Pre-study Phase is done a file will be sent by the Reviewer to the Licence Provider generating the set up for the Client.

The method software program for the specific client will be sent to the Reviewer, that will put in all the results gathered and when finished a file will be sent to the Licence provider. The file will be the base for calculating the final result and these will than be sent to the Reviewer to present this to Client as the last pieces in the method software program.

It is important to point out that the License Provider will always have the mathematical criteria's for determine the final result and thus ensure that the result is done correctly.

Method Support

The method software program will contain all questions and guides of how to perform the data gathering and setting the raw results. It will also contain features that will speed up the process of gathering the data.

The method support software program contains also the values for effort weights for each checkpoint.

Result Presentation and Analysing the Result

The method software program will generate a number of standard reports to choose from. This will include both written and graphical presentations.

The report will be in a form of a database and also enable the Client and/or Entity to revise the result in a way that suits further work and actions to improve the information security.

Software Toolbox—IT-System Security Analysis

The idea with using a software toolbox is that IT-systems are somewhat insecure because the rapid development of hacker and cracker tools on the Internet.

The software toolbox contains a major part of hacker and cracker tools and other system analysis tools and it will enable modifications to adapt to be used for the actual analysis. The toolbox is built in a modular system to be adaptive to actual system. Here the variation could vary greatly when it comes to applications. In order to handle that situation are the applications dived into standard application and extended application. The standard applications are the ones that are known and which information about threats etc is available to anybody. Extended application could be unique, new or old application where information is not available. These have to be dealt with depending on what functions they may have. If they are considered to be of importance from a security point of view they will be tested and moved to as a standard application for the next test.

Method Result Presentation

In the following will be described how to interpret the results. The results are presented in two media. One media is written report and the other media is a software. The results are presented as text and also in graphical format.

Presentation of the Total Result

The presentation of the total result is presented in written form as “Passed” or “Failed” on the level measured. If passed the Entity Owner will receive an Certificate stating that the Entity measured has for example passed the Veriscan 3 (Or applicable level). Additional comments on major issues are addressed. As the total results is based on the results of each area is the total result also presented in a graphical form including the three areas, see FIG. 4a.

Presentation of Category Results

The category results are presented in a graphical format showing the weighted and non weighted results for each category within an area for all three areas. Written major comments are added.

Presentation of Checkpoint Results

For each area and under each category are the checkpoint results listed as verbal comments and connected to these are also the graphical presentation. For each checkpoint result is also the effort weight presented, see FIG. 4b.

In the graphical presentation can also the Security Gap (SeG) clearly be seen and also the Priority Gap (PrG).

Evaluation of Results

In the following will described how to interpret the various results and how to use them separately and to use them together.

General Principle

The methodology is based on the principle of that the total security level is depending on that the three sub areas has an equally sufficient security level. Within each area is then the result generated by mathematical formulas adding up the results from each checkpoint within the area. For each category is also the results mathematically calculated from the results from each checkpoint as well as there is also a total result generated by calculating all checkpoints but this is additional information.

Thus is can be stated that the method's results are gained from measuring the smallest part, the checkpoint, and summarise these into the areas. If all areas have passed a critical value is the level approved and the organisation has passed the test on the decided and measured level and can be given its certificate/Rating for that level.

As all results are numerical from total rating to area, category and checkpoint they are comparable within the level. As they are in numerical form they are possible to compare between different analysis conducted, if consideration are taken into versions and that comparison is made on the same type of value (weighted or non weighted).

Evaluate the Result on the Total Level

The result of the level measured requires that all three areas has a result of three or higher to be approved on the level. If approved the entity will be rated according to the level measured, Veriscan 1 to 5.

As the lower level are a sub set of the above level it may be possible to receive a lower rating than on the actual level measured, e.g. if not approved on measured level Veriscan 3 but approved on level 2 the organisation will be rated 2. But it is not possible to have a higher rating than the level that was measured due to that higher levels contains more extended demands and more checkpoints, e.g. if measured on level three and approved with a high margin it is still not possible to receive a higher Rating than Veriscan 3.

If a result of less that 1 is achieved on particular important checkpoints this will be stated. This may be a checkpoint necessary to have a passed result.

Taking the Secondary Total Analysis results both in weighted and non weighted form will indicate the following:

If a ToWR is higher than or in the region of 3 but the analysis is not passed, there is minor achievement needed to pass on the level and subsequently if the ToWR is lower there is major amendments necessary.

If the ToR is higher than the ToWR this shows that the entity is not good at important checkpoints. If the ToR is lower than ToWR is the organisation generally performs better on important checkpoints. This information is suitable for top management presentation and comments.

Evaluate the Result of an Area

The result of each area is on an analogue scale of 0-5, and that the result of 3 or higher for each area means that the area has been approved according to the criteria set for the level. As the method is based upon that the security demands on each area are balanced within a level, it can easily be evaluated which are needs attention if a non-approved result is achieved.

Further more the results of each area can easily be distributed to responsible managers within the entity is so preferred by the client.

For example:

Logical Security Result: 2,3
Physical Security Result: 3,6
System Security Result: 2,9

The conclusion is that the Management has to look into the Logical Security Area, as the most action needs to be taken there. The System Security Area needs just some minor action to pass and the Physical Area does not need any further investments at the moment.

Within an Area the results on each checkpoint could also be evaluated. However this should be done in two different ways. First the comparison between if the performance have been generally good or poor on important checkpoints. This should be done by comparing the non weighted and weighted results on the Checkpoints of the Area, e.g. comparing Area Result without Category weight (AeR) and Area result with checkpoint weight (AeRW).

If the weighted Area result (AeWR) is lower than the non weighted area result (AeWR), the entity has mainly poor results on particular checkpoints that are more important than others, i.e. have a higher checkpoint weight. If the (AeWR), is higher than the (AeR), the organisation has reached good results in checkpoints that are generally more important then others. The second evaluation is of course to see what checkpoints that have a non-approved result and also to see what is the performance on the most important ones.

Evaluate the Result of a Category

The results of a Category is also on a analogue scale of 0-5, and that the result of 3 or higher for each area means that the area has been approved according to the criteria set for the level.

A result of X means that the specific Category has not been measured.

Evaluation should be made taking the results of categories within an area and compare weighted and non weighted results. This is additional information and could serve as a guide for management on a not to detailed level.

This is done by comparing the weighted result (CaWR) and compare it to the non weighted result of the same category (CaR). If the weighted result (CaWR) is higher than the non weighted one (CaR) is the organisation better on important categories.

Particular attention should be taken to Categories not measured.

Evaluate the Result of a Checkpoint

The results of a checkpoint is also on a analogue scale of 0-5, and that the result of 3 or higher for each area means that the area has been approved according to the criteria set for the level.

A result of X means that the specific Checkpoint has not been measured.

As the Checkpoint results are the basis both for the results but also for the actions can the evaluation of this bee done in many ways. When evaluating the Checkpoint is it also possible to use the Effort weights and also the Position weights to get more information of taking the proper actions for enhanced information security. However these results may not bee necessary to view for top management but can be used for the personnel directly responsible within the Entity for certain information security areas and categories.

The first evaluation should be done so that the security gap is considered, meaning that checkpoints with a low result should be looked into. These should then be divided into different groups. The first group is the ones with low Effort weight, which means that actions can be taken directly to correct them.

Secondly are the ones with higher effort weights where work should be initiated and evaluation of proper actions should start. Third is the group that has medium effort weights, which should be looked into individually.

Naturally the checkpoint results could be sorted within Areas and/or Categories if so preferred.

It is also important to point out that as the effort weights are general there are variations depending on the Entity resources. For example if the entity has a large IT department the costs may be lower to make correction as less eternal purchase of Consultants may be necessary.

Further Position weights could be analysed to evaluate where necessary training of information security is needed, e.g. if generally personnel with high weight have poor results but not personnel with lower, management and IT personnel needs more training than the standard users.

Particular attention should be taken to checkpoints not measured.

FIG. 1a

Shows entities with described needed measured information security level. Two entities shares or exchange information that is considered valuable. The entity to the left has a certificate on Veriscan level 3 but does not know what information security level the right Entity has. Insecurity arises and the transaction may not take place. If however the right entity also had a corporate certificate on a Veriscan level suitable to them, the left entity can evaluate if they can handle the information properly. The corporate certificate has enable trust and the transaction may take place. Comparison between entities has been enabled when it comes to such a complex area as information security by simply stating a figure created related to a norm.

FIG. 2a

FIG. 2a shows the principle of the breakdown of terms within an area.

Table 2b

Table 2b shows the method differences between the levels.

FIG. 3a

This figure shows the method steps.

FIG. 4a

FIG. 4a shows the total results in graphical form.

FIG. 4b

FIG. 4b shows a graphical result table

FIG. 5

FIG. 5 is a schematic diagram showing an exemplary result of Veriscan Rating 3, including ratings for logical, physical, system, and operational security.

FIG. 6

FIG. 6 is a schematic diagram showing an exemplary hierarchy of a total result, including a total based on organizational, logical, physical, and system checkpoints.

The concept is described in a complementary way by means of 65 slides of a slide show enclosed on separate pages numbered 1-33.

The method according to the invention could also be extended to operational risk rating.

Operational Risk is here defined as an organisation total risk of loosing immaterial assets.

As information is a major part of immaterial assets is it possible to add a fourth area to the three areas for information security to control and rate also Operational Risk. This Fourth area is subsequently only added if Operational Risk Rating is selected rather than Information Security Rating.

This is called Veriscan Rating of Operational Security.

Definition Organisational Security

Operational risks is defined as other risks of errors, mistakes crime or accidents that causes directly or indirectly losses.

Organisational Security risks are defined as risks as above caused by the organisations structure and functions and environment. Environment could be linked both to factors that the organisation controls as well as the environment in which the organisation operates that is outside of the organisation's control.

The method is the same as for measuring Information Security.

The categories and checkpoints are of course of a different type. Categories are such as: Legal, Image and PR, Control Systems, Business Strategies, Authorities, Financial, Market Mix, Spirit, Product & Service Structure.

Measurement of Organisational Risk

The measurement is basically the same as for the other areas and only abbreviations will be presented in these chapter.

Measurement Basic Methods

The measurement are using the same methods as for the another three areas with one slightly different aspect and that is that the measurement of IT/Systems is naturally not used.

For measuring Financial Risk, may also sector or other methods be used, as these are from the beginning in mathematical form that can be transferred to the relative system in this method.

Area Results

The area result is presented separately, and also in a graphical form including the other three areas.

Level Fulfilment

To add organizational Risk to the other three areas states that the measurement of fulfillment is done for Operational Security and not for information security. But as information security is part of the Operational Security must all areas have been measured and have passed result to be rated for Operational Security. This is the identical principal as Informational Security Rating but there are four areas to consider.

Method Processing

The method processing for Operational Risk Measurement is done in addition to the information security measurement and follows it's workflow. Only differences will be stated in this chapter.

The major difference is that from start it must be clear if it is the information security that should be measured with Logical, Physical and System areas or operational Security which ads the fourth area Organisational Risk.

The workflow is in principle the same with the exception of that System analyses is not conducted.

There will be additions to all other phases to cover the forth area. Generally the fourth area means that each phase is spitted up with two focuses. One is the same to cover Information Security and the second focus is the Organisation Security measurement. The pre-study for example will be extended as a much more documentation and more people have to interview and this must be set from start.

Method Result Presentation

The result is presented in the same manor but with the addition of the fourth area. Please see below. The symbol for passed rating for Operational Security is different as for information security Rating and the name are also different in order to not misjudge the different measurements.

Lundmark, Lorentz, Gustafsson, Mikael, Branzell, Jan, Rodrick, Anders

Patent Priority Assignee Title
8539586, May 19 2006 Method for evaluating system risk
9626515, Dec 30 2014 Samsung Electronics Co., Ltd. Electronic system with risk presentation mechanism and method of operation thereof
Patent Priority Assignee Title
5440498, May 06 1993 Method for evaluating security of protected facilities
6950944, Dec 18 2000 Security code activated access control system
7552480, Apr 23 2002 Citibank, N.A. Method and system of assessing risk using a one-dimensional risk assessment model
/////
Executed onAssignorAssigneeConveyanceFrameReelDoc
Sep 21 2001Veriscan Security Aktiebolag(assignment on the face of the patent)
Mar 19 2003GUSTAFSSON, MIKAELVeriscan Security AktiebolagASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0144190530 pdf
Mar 19 2003BRANZELL, JANVeriscan Security AktiebolagASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0144190530 pdf
Mar 20 2003RODRICK, ANDERSVeriscan Security AktiebolagASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0144190530 pdf
Mar 20 2003LUNDMARK, LORENTZVeriscan Security AktiebolagASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0144190530 pdf
Date Maintenance Fee Events
Dec 24 2014REM: Maintenance Fee Reminder Mailed.
Feb 20 2015M2551: Payment of Maintenance Fee, 4th Yr, Small Entity.
Feb 20 2015M2554: Surcharge for late Payment, Small Entity.
Jan 07 2019REM: Maintenance Fee Reminder Mailed.
Jun 24 2019EXP: Patent Expired for Failure to Pay Maintenance Fees.


Date Maintenance Schedule
May 17 20144 years fee payment window open
Nov 17 20146 months grace period start (w surcharge)
May 17 2015patent expiry (for year 4)
May 17 20172 years to revive unintentionally abandoned end. (for year 4)
May 17 20188 years fee payment window open
Nov 17 20186 months grace period start (w surcharge)
May 17 2019patent expiry (for year 8)
May 17 20212 years to revive unintentionally abandoned end. (for year 8)
May 17 202212 years fee payment window open
Nov 17 20226 months grace period start (w surcharge)
May 17 2023patent expiry (for year 12)
May 17 20252 years to revive unintentionally abandoned end. (for year 12)