10
Expected benefits of information security investments Julie J.C.H. Ryan a, *, Daniel J. Ryan b a Department of Engineering Management and System Engineering, The George Washington University, Washington, DC 20052, USA b Information Resources Management College, National Defense University, Washington, DC 20319, USA article info Article history: Received 27 January 2005 Revised 11 June 2006 Accepted 3 August 2006 Keywords: Security Information security Attack probabilities Return-on-investment Benefits of security investments abstract Ideally, decisions concerning investments of scarce resources in new or additional proce- dures and technologies that are expected to enhance information security will be informed by quantitative analyses. But security is notoriously hard to quantify, since absence of activity challenges us to establish whether lack of successful attacks is the result of good security or merely due to good luck. However, viewing security as the inverse of risk en- ables us to use computations of expected loss to develop a quantitative approach to mea- suring gains in security by measuring decreases in risk. In using such an approach, making decisions concerning investments in information security requires calculation of net ben- efits expected to result from the investment. Unfortunately, little data are available upon which to base an estimate of the probabilities required for developing the expected losses. This paper develops a mathematical approach to risk management based on Kaplan–Meier and Nelson–Aalen non-parametric estimators of the probability distributions needed for using the resulting quantitative risk management tools. Differences between the integrals of these estimators evaluated for enhanced and control groups of systems in an informa- tion infrastructure provide a metric for measuring increased security. When combined with an appropriate value function, the expected losses can be calculated and investments evaluated quantitatively in terms of actual enhancements to security. ª 2006 Elsevier Ltd. All rights reserved. 1. Introduction Making decisions concerning investments in information se- curity requires calculation of net benefits expected to result from the investment. Gordon and Loeb (2002) suggest that expected loss provides a useful metric for evaluating whether an investment in information security is warranted. They pro- pose that since expected loss is the product of the loss v that would be realized following a successful attack on the systems comprising our information infrastructure and the probability that such a loss will occur, one way of accomplishing such calculations is to consider for an investment i the probabilities p 0 and p i of the losses occurring with and without the investment, respectively. The expected net benefit of the investment i is, then, E NB ½i¼ p 0 n p i n i ¼ p 0 p i n i: (1) A positive expected net benefit characterizes an attractive investment opportunity. In addition to being subject to a number of simplifying as- sumptions, such as the loss being constant rather than a func- tion of time, the ability to use this equation depends upon the ability to obtain probability distributions for information secu- rity failures. The notion of an information security ‘‘failure’’ is itself a concept that requires careful consideration. ‘‘Failure’’ may not necessarily mean the catastrophic destruction of in- formation assets or systems. Failure in this context means some real or potential compromise of confidentiality, integrity or availability. An asset’s confidentiality can be compromised by illicit access even if the integrity and availability of the * Corresponding author. E-mail addresses: [email protected] (J.J.C.H. Ryan), [email protected] (D.J. Ryan). available at www.sciencedirect.com journal homepage: www.elsevier.com/locate/cose 0167-4048/$ – see front matter ª 2006 Elsevier Ltd. All rights reserved. doi:10.1016/j.cose.2006.08.001 computers & security 25 (2006) 579–588

Expected benefits of information security investments

  • Upload
    ndu

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

ava i lab le at www.sc ienced i rec t . com

journa l homepage : www.e lsev ie r . com/ loca te /cose

c o m p u t e r s & s e c u r i t y 2 5 ( 2 0 0 6 ) 5 7 9 – 5 8 8

Expected benefits of information security investments

Julie J.C.H. Ryana,*, Daniel J. Ryanb

aDepartment of Engineering Management and System Engineering, The George Washington University, Washington, DC 20052, USAbInformation Resources Management College, National Defense University, Washington, DC 20319, USA

a r t i c l e i n f o

Article history:

Received 27 January 2005

Revised 11 June 2006

Accepted 3 August 2006

Keywords:

Security

Information security

Attack probabilities

Return-on-investment

Benefits of security investments

a b s t r a c t

Ideally, decisions concerning investments of scarce resources in new or additional proce-

dures and technologies that are expected to enhance information security will be informed

by quantitative analyses. But security is notoriously hard to quantify, since absence of

activity challenges us to establish whether lack of successful attacks is the result of good

security or merely due to good luck. However, viewing security as the inverse of risk en-

ables us to use computations of expected loss to develop a quantitative approach to mea-

suring gains in security by measuring decreases in risk. In using such an approach, making

decisions concerning investments in information security requires calculation of net ben-

efits expected to result from the investment. Unfortunately, little data are available upon

which to base an estimate of the probabilities required for developing the expected losses.

This paper develops a mathematical approach to risk management based on Kaplan–Meier

and Nelson–Aalen non-parametric estimators of the probability distributions needed for

using the resulting quantitative risk management tools. Differences between the integrals

of these estimators evaluated for enhanced and control groups of systems in an informa-

tion infrastructure provide a metric for measuring increased security. When combined

with an appropriate value function, the expected losses can be calculated and investments

evaluated quantitatively in terms of actual enhancements to security.

ª 2006 Elsevier Ltd. All rights reserved.

1. Introduction

Making decisions concerning investments in information se-

curity requires calculation of net benefits expected to result

from the investment. Gordon and Loeb (2002) suggest that

expected loss provides a useful metric for evaluating whether

an investment in information security is warranted. They pro-

pose that since expected loss is the product of the loss v that

would be realized following a successful attack on the systems

comprising our information infrastructure and the probability

that such a loss will occur, one way of accomplishing such

calculations is to consider for an investment i the probabilities

p0 and pi of the losses occurring with and without the

investment, respectively. The expected net benefit of the

investment i is, then,

ENB½i� ¼ p0n� pin� i ¼�p0 � pi

�n� i: (1)

A positive expected net benefit characterizes an attractive

investment opportunity.

In addition to being subject to a number of simplifying as-

sumptions, such as the loss being constant rather than a func-

tion of time, the ability to use this equation depends upon the

ability to obtain probability distributions for information secu-

rity failures. The notion of an information security ‘‘failure’’ is

itself a concept that requires careful consideration. ‘‘Failure’’

may not necessarily mean the catastrophic destruction of in-

formation assets or systems. Failure in this context means

some real or potential compromise of confidentiality, integrity

or availability. An asset’s confidentiality can be compromised

by illicit access even if the integrity and availability of the

* Corresponding author.E-mail addresses: [email protected] (J.J.C.H. Ryan), [email protected] (D.J. Ryan).

0167-4048/$ – see front matter ª 2006 Elsevier Ltd. All rights reserved.doi:10.1016/j.cose.2006.08.001

c o m p u t e r s & s e c u r i t y 2 5 ( 2 0 0 6 ) 5 7 9 – 5 8 8580

asset are preserved. Assets may have their integrity compro-

mised even if their confidentiality and availability are un-

changed. Obviously, destruction of an asset compromises its

availability, even when confidentiality and integrity may be

inviolate. Degradation of performance can be a form of failure,

even when the system continues to operate correctly, albeit

slowly. Successful installation of malicious code can be

a form of information security failure even when the mali-

cious code has not yet affected system performance or com-

promised the confidentiality, integrity or availability of

information assets. Consequently, detection of failures for

use in the models discussed in this paper requires that, as

part of our experimental design, we carefully define what

types of failures are to be considered. Nevertheless, it is only

by examining failures that we can begin to understand the

actual security states of our information infrastructures.

Counting vulnerabilities we have patched or numbers of

countermeasures implemented may provide evidence to indi-

rectly reassure us that we are making progress in protecting

our valuable information assets, but only metrics that

measure failure rates can inform us how well we are actually

abating risk. Security is the inverse of risk, and risk is

measured by expected loss.

Unfortunately, little data are available upon which to base

an estimate of the probabilities of failure that are required for

expected loss calculations (Ryan and Jefferson, 2003). Mariana

Gerber and Rossouw von Solms (2001), in describing the use of

quantitative techniques for calculation of annual loss expec-

tancies, say dryly, ‘‘The only factor that was somewhat

subjective was determining the likelihood of a threat mani-

festing.’’ Other authors explicitly or implicitly assume the

availability of these probability distributions in models they

propose to apply to risk management, cost-benefit or return-

on-investment decisions (e.g., see Carroll, 1995; Ozier, 1999;

Wei et al., 2001; Iheagwara, 2004; Cavusoglu et al., 2004, among

others). John Leach (2003) says, ‘‘The data is there to be gath-

ered but it seems we are not in the practice of gathering it. We

haven’t started to gather it because we haven’t yet articulated

clearly what questions we want the gathered data to answer.’’

This paper will develop a mathematical approach to risk man-

agement that will clarify the data that need to be collected,

and will explore methods for non-parametric estimation of

the probability distributions needed for using the resulting

quantitative risk management tools.

2. Failure time distributions

The mathematics of failure time distributions is explored

thoroughly in several texts. (See Kalbfleisch and Prentice,

2002; Collett, 2003; Bedford and Cooke, 2001; Crowder, 2001;

Therneau and Grambsch, 2000 for excellent coverage of the

field.) In the case of information security, the exploration of

time to failure is interesting in that it provides a contextual

basis for investment decisions that is easily understood by

managers and financial officers. Showing an economic benefit

over a given period of time can be used to help them to under-

stand investment strategies that take into account capital

expenditures, risk mitigation, and residual risk in an opera-

tional environment.

If T is a non-negative random variable representing times

of failure of individuals in a homogenous population, T can

be specified using survivor functions, failure functions, and

hazard functions. Both discrete and continuous distributions

arise in the study of failure data.

The survivor function is defined for discrete and continuous

distributions by

SðtÞ ¼ PrðT � tÞ; 0 < t < N: (2)

That is, S(t) is the probability that T exceeds a value t in its

range. S(t) is closely related to several other functions that

will prove useful in risk assessment and management, in-

cluding the failure function F(t), which determines the cumu-

lative probability of failure, and its associated probability

density f(t), and the hazard function h(t), which provides

the instantaneous rate of failure, and its cumulative function

H(t). The way in which these functions relate is determined

as follows.

The failure function or cumulative distribution function

associated with S(t) is

FðtÞ ¼ PrðT < tÞ ¼ 1� SðtÞ: (3)

S(t) is a non-increasing right continuous function of t with

S(0)¼ 1 and limt/NSðtÞ ¼ 0.

The probability density function f(t) of T is

fðtÞ ¼ dFðtÞdt¼ d½1� SðtÞ�

dt¼ � dSðtÞ

dt: (4)

Now f(t) gives the density of the probability at t, and so,

fðtÞdzPrðt � T < tþ dÞ ¼ SðtÞ � Sðtþ dÞ: (5)

for small values of d, providing that f(t) is continuous at t. Also,

f(t)� 0,Z N

0

fðtÞ ¼ 1; and SðtÞ ¼Z N

t

fðsÞds: (6)

The hazard function of T is defined as

hðtÞ ¼ limd/0þPrðt � T < tþ djT � tÞ=d: (7)

The hazard is the instantaneous rate of failure at t of indi-

viduals that have survived up until time t. From Eq. (7) and def-

inition of f(t),

hðtÞ ¼ fðtÞSðtÞ: (8)

Integrating with respect to t

SðtÞ ¼ exp

"�Z t

0

hðsÞds

#¼ exp½�HðtÞ�; (9)

where HðtÞ ¼R t

0 hðsÞds ¼ �log SðtÞ is the cumulative hazard

function. Then

fðtÞ ¼ hðtÞexp½�HðtÞ�: (10)

If T is a discrete random variable, and takes on values

a1< a2</ with a probability given by the function

fðaiÞ ¼ PrðT ¼ aiÞ; i ¼ 1;2;.; (11)

then the survivor function is given by

SðtÞ ¼Xjjaj>t

f�aj

�: (12)

c o m p u t e r s & s e c u r i t y 2 5 ( 2 0 0 6 ) 5 7 9 – 5 8 8 581

The hazard function at ai is the conditional probability of

failure at ai provided that the individual has survived to ai.

hi ¼ PrðT ¼ aijT � aiÞ ¼fðaiÞS�a�i�; where S

�a�i�¼ limt/a�

iSðtÞ: (13)

Then the survivor function is

SðtÞ ¼Y

jjai�t

ð1� hiÞ; (14)

and the probability density function is

fðaiÞ ¼ hi

Yi�1

j�1

ð1� hiÞ: (15)

3. Empirical estimates of survivor andfailure distributions

Failure data are not generally symmetrically distributed,

usually tending to be positively skewed with a longer tail

to the right of the peak of observed failures. Hiromitsu

Kumamoto and Ernest J. Henley (1996) show that if the

data were not censored, we could use an empirical distribu-

tion function to model survivor, failure and hazard

distributions.

Let N be the number of individual systems in the study, and

let n(t) be the number of failures occurring prior to time t.

Then, nðtþ dÞ � nðtÞ is the number of systems that can be

expected to fail during the time interval [t, tþ d). N� nðtÞ is

the number of systems still operational at time t.

An empirical estimator SðtÞ for the survivor function SðtÞ is

given by

SðtÞ ¼ Number of individuals operational at time tNumber of individual systems in the study

¼ N� nðtÞN

:

(16)

So, an empirical estimator FðtÞ for FðtÞ is

FðtÞ ¼ 1� SðtÞ ¼ nðtÞN: (17)

Then, we see that

fðtÞ ¼ dFðtÞdt

xFðtþ dÞ � FðtÞ

d¼ nðtþ dÞ � nðtÞ

dN: (18)

For sufficiently small d, the empirical estimator hðtÞ for the

instantaneous failure rate (hazard function) hðtÞ should ap-

proximately satisfy

hðtÞd¼Number of failures during ½t; tþ dÞ

Number of systems still operational at time t

¼ nðtþ dÞ � nðtÞN� nðtÞ

¼�

dnðtþ dÞ � nðtÞ

dN

���N� nðtÞ

N

¼ fðtÞdSðtÞ

:

So, hðtÞ ¼ fðtÞ = SðtÞÞ���

, as we would expect from Eq. (8).

Now, consider the notional system failure data given in

Table 1. The entries in the table are the times, in days follow-

ing initiation of operations, at which the computer systems

we have been tracking were observed to fail due to attacks

on information assets they are creating, storing, processing

or communicating. Thus, two systems fail in the first day,

a second fails on the 19th day, and so on until the final failure

on the 126th day. Graphs of the empirical survivor and empir-

ical failure functions are shown in Fig. 1.

The data in Table 1 consist of failure times due to suc-

cessful attacks on systems being studied. Usually, however,

some of the systems entered into our studies either survive

beyond the time allotted for the study and are still going

strong when we cease to track them, or they fail during

the study period for reasons unrelated to security – perhaps

due to reliability problems not resulting from successful at-

tacks, or due to human errors rather than attacks. In either

case, studies that collect failure time data usually have such

cases, and the systems that survive or that fail for reasons

other than those related to the purpose of the study are

said to be ‘‘right censored.’’ Other types of censoring can

occur as well (see Collett, 2003, pp. 1–3; Kalbfleisch and

Prentice, 2002, pp. 52–54).

Unfortunately, the presence of censored data makes use of

the empirical estimators impossible, because the definitions

of the functions do not allow information provided by a system

for which survival time is censored prior to time t to be used in

calculating the functions at t. Fortunately, other estimators

have been developed that can take advantage of censored

data and provide valid and useful estimations of the functions

we need in determining the advantage of a proposed

investment.

Table 1 – Notional computer system failure data

0.5 0.5 19.5 20 23.5 23.5 25.5 25.5 27.5 27.5

27.5 28.5 29 29 29.5 30 30.5 30.5 31.5 33

33 33 33.5 33.5 34 34.5 34.5 34.5 35.5 35.5

36 36 37 37 37 37.5 37.5 38.5 38.5 38.5

38.5 39.5 39.5 39.5 40 40 40.5 40.5 40.5 41

41 42 42 42.5 42.5 43 43 43 43 44.5

44.5 44.5 44.5 45 45 46 46 46 46.5 46.5

47 48 48.5 48.5 48.5 48.5 49 50 50 50

51 51 52 52 53 53 53.5 54.5 55 56

57 57 59.5 61 100 102 109 121 121 126

Values represent days.

c o m p u t e r s & s e c u r i t y 2 5 ( 2 0 0 6 ) 5 7 9 – 5 8 8582

4. Estimates of survival and failuredistributions with censoring

The 100 entries in Table 2 are notional times at which com-

puter systems we have been tracking were observed to fail

due to attacks on information assets they are creating, storing,

processing or communicating. Thus, two systems fail on the

first day, another fails on the 19th day, and so forth until six

systems remain operational when the experiment is termi-

nated on the 100th day. The six surviving systems are right-

censored. Systems that fail on the 27th, 28th, and the other

entries marked as negative times, are also censored, repre-

senting failures due to causes unrelated to successful attacks

such as, perhaps, reliability failures.

We will explore two estimators that provide useful repre-

sentations of the probability distributions underlying such

data.

5. Kaplan–Meier estimators

One such estimator is the product–limit estimator of the

survivor function derived by Kaplan and Meier (1958) using

the method of maximum likelihood, although the estimator

Survival and Failure Distributions

0

0.2

0.4

0.6

0.8

1

1.2

Days

S(t)

F(t)

0

8.5 17 25 34 42 51 59 68 76 85 93 102

111

119

128

Fig. 1 – Empirical survival and failure functions for system

failure data.

was known and used earlier. Let t0 represents the start of

our study, and let t1< t2</< tk, be a complete list of times

at which individual systems being tracked in the study fail.

To construct this estimator, we construct a series of time

intervals, each of which starts at the time ti at which one

or more systems in the study fails, and ends immediately

before the next time of failure tiþ1. Thus, at least one,

and perhaps several, system(s) fail(s) at exactly the start

of the interval, and no other systems fail during the inter-

val [ti, tiþ1). If individuals are censored at times during the

interval [ti, tiþ1), their censoring does not change the con-

struction of the interval, but should one or more systems

be censored at exactly the start of the interval, we will as-

sume that their censoring actually occurs immediately after

ti, and not at ti.

The number of systems that remain functional just before

time tj is denoted as nj, and dj will represent the number of sys-

tems in the study that fail at tj. The probability of failure at tj is

estimated by dj/nj, while the probability of surviving is

1� ðdj=njÞ ¼ ðnj � djÞ=ðnjÞ. Since there are no deaths between

tj and tjþ1, the probability of survival in that interval is unity,

and the joint probability or surviving the interval from tj to

tjþ1 is ðnj � djÞ=ðnjÞ.We assume for now that the failures of the systems in the

study are independent of one another, so the estimated survi-

vor function at any time t, where tj� t< tjþ1, j¼ 1,2,.,k� 1,

where tkþ1¼N, will be the estimated probability of surviving

beyond tj, which is the probability of surviving through the

jth interval and all the preceding intervals. This is the product

SKMðtÞ ¼Yj

i¼1

ni � di

ni

; for tj � t < tjþ1; j ¼ 1;2;.; k� 1; (20)

and we have SKMðtÞ ¼ 1 when t< t1, and SKMðtÞ ¼ 0 when tk� t.

Actually, if the largest recorded time t* is a censored survival

time, SKM(t) is undefined for t> t*.

Note that if there are no censored failure times in the data

set, then nj � dj ¼ njþ1 in Eq. (20), and

SKMðtÞ ¼Yk

i¼1

niþ1

ni¼ nkþ1

n1: (21)

Since n1¼N, and nkþ1 is the number of systems that survive

beyond tkþ1, the Kaplan–Meier estimate of the survivor func-

tion is identical to the empirical estimator of the survival in

the absence of censored data.

Table 2 – Notional computer system failure data

0.5 0.5 19.5 20 23.5 23.5 25.5 25.5 27.5 27.5

�27.5 28.5 29 �29 29.5 30 30.5 30.5 31.5 33

33 33 33.5 �33.5 34 34.5 34.5 34.5 35.5 35.5

36 �36 37 37 37 37.5 37.5 38.5 38.5 38.5

38.5 39.5 39.5 39.5 40 40 40.5 40.5 40.5 41

�41 42 42 42.5 42.5 43 43 43 43 44.5

44.5 44.5 44.5 45 �45 46 46 46 46.5 46.5

47 48 48.5 48.5 48.5 48.5 49 50 50 50

51 �51 52 52 53 �53 53.5 54.5 55 56

57 7 59.5 61 �100 �100 �100 �100 �100 �100

The sign � indicates censoring. Values represent days.

c o m p u t e r s & s e c u r i t y 2 5 ( 2 0 0 6 ) 5 7 9 – 5 8 8 583

The estimator for the failure function that corresponds to

Eq. (20) is

FKMðtÞ ¼ 1� SKMðtÞ: (22)

It is natural to estimate the hazard function by the ratio of

the number of systems that fail at a given time divided by the

number at risk at that time. More formally, because

HðtÞ ¼ �log SðtÞ, we have

HKMðtÞ ¼ �log SKMðtÞ ¼ �Xj

i¼1

log

ni � di

ni

; for tj � t < tjþ1: (23)

and

hKMðtÞ zHKM

�tj

��HKM

�tj�1

��tjþ1 � tj

� zdj

nj

�tjþ1 � tj

�; for tj � t < tjþ1:

(24)

This equation can be applied to all intervals except the

interval that begins at t

k

, since that interval is infinite in

length.

Returning our attention to the data in Table 2, we set out

the calculations in Table 3 for the Kaplan–Meier estimators

of the survivor, failure and hazard functions. Fig. 2 shows

the Kaplan–Meier estimator for the survivor function.

Table 3 – Kaplan–Meier calculations

j tj tjþ1� tj dj cj (nj� dj)/nj SKM(t) FKM(t) hKM(t) fKM(t)

0 0 0.5 0 0 1 1 0 0.0000 0.0000

1 0.5 19 2 0 0.9800 0.9800 0.0200 0.0011 0.0011

2 19.5 0.5 1 0 0.9897 0.9699 0.0301 0.0206 0.0200

3 20 3.5 1 0 0.9896 0.9598 0.0402 0.0030 0.0029

4 23.5 2 2 0 0.9789 0.9396 0.0604 0.0106 0.0100

5 25.5 2 2 0 0.9785 0.9194 0.0806 0.0109 0.0100

6 27.5 1 2 1 0.9780 0.8992 0.1008 0.0225 0.0202

7 28.5 0.5 1 0 0.9886 0.8890 0.1110 0.0227 0.0202

8 29 0.5 1 1 0.9885 0.8787 0.1213 0.0233 0.0205

9 29.5 0.5 1 0 0.9882 0.8684 0.1316 0.0235 0.0204

10 30 0.5 1 0 0.9881 0.8581 0.1419 0.0238 0.0204

11 30.5 1 2 0 0.9759 0.8374 0.1626 0.0244 0.0204

12 31.5 1.5 1 0 0.9877 0.8270 0.1730 0.0082 0.0068

13 33 0.5 3 0 0.9615 0.7952 0.2048 0.0769 0.0612

14 33.5 0.5 1 1 0.9867 0.7846 0.2154 0.0263 0.0206

15 34 0.5 1 0 0.9863 0.7739 0.2261 0.0267 0.0207

16 34.5 1 3 0 0.9583 0.7416 0.2584 0.0417 0.0309

17 35.5 0.5 2 0 0.9706 0.7198 0.2802 0.0571 0.0411

18 36 1 1 1 0.9848 0.7089 0.2911 0.0147 0.0104

19 37 0.5 3 0 0.9524 0.6752 0.3248 0.0923 0.0623

20 37.5 1 2 0 0.9667 0.6527 0.3473 0.0317 0.0207

21 38.5 1 4 0 0.9310 0.6076 0.3924 0.0678 0.0412

22 39.5 0.5 3 0 0.9444 0.5739 0.4261 0.1071 0.0615

23 40 0.5 2 0 0.9608 0.5514 0.4486 0.0741 0.0409

24 40.5 0.5 3 0 0.9388 0.5176 0.4824 0.1176 0.0609

25 41 1 1 1 0.9783 0.5064 0.4936 0.0204 0.0103

26 42 0.5 2 0 0.9545 0.4834 0.5166 0.0851 0.0411

27 42.5 0.5 2 0 0.9524 0.4603 0.5397 0.0889 0.0409

28 43 1.5 4 0 0.9000 0.4143 0.5857 0.0650 0.0269

29 44.5 0.5 4 0 0.8857 0.3670 0.6330 0.2162 0.0793

30 45 1 1 1 0.9677 0.3551 0.6449 0.0286 0.0102

31 46 0.5 3 0 0.8966 0.3184 0.6816 0.1875 0.0597

32 46.5 0.5 2 0 0.9231 0.2939 0.7061 0.1333 0.0392

33 47 1 1 0 0.9583 0.2816 0.7184 0.0345 0.0097

34 48 0.5 1 0 0.9565 0.2694 0.7306 0.0714 0.0192

35 48.5 0.5 4 0 0.8182 0.2204 0.7796 0.3333 0.0735

36 49 1 1 0 0.9444 0.2082 0.7918 0.0435 0.0091

37 50 1 3 0 0.8235 0.1714 0.8286 0.1500 0.0257

38 51 1 1 1 0.9286 0.1592 0.8408 0.0556 0.0089

39 52 1 2 0 0.8333 0.1327 0.8673 0.1250 0.0166

40 53 0.5 1 1 0.9000 0.1194 0.8806 0.1429 0.0171

41 53.5 1 1 0 0.8750 0.1045 0.8955 0.0769 0.0080

42 54.5 0.5 1 0 0.8571 0.0895 0.9105 0.1667 0.0149

43 55 1 1 0 0.8333 0.0746 0.9254 0.0909 0.0068

44 56 1 1 0 0.8000 0.0597 0.9403 0.1000 0.0060

45 57 2.5 2 0 0.5000 0.0298 0.9702 0.1000 0.0030

46 59.5 1.5 1 0 0.5000 0.0149 0.9851 0.0952 0.0014

47 61 1 0 0.0000 0.0000 1.0000 0.0000 0.0000

c o m p u t e r s & s e c u r i t y 2 5 ( 2 0 0 6 ) 5 7 9 – 5 8 8584

6. Nelson–Aalen estimators

The Nelson–Aalen estimator was first proposed by Nelson

(1969, 1972). Altshuler (1970) also derived the estimator. It

has been shown that the Nelson–Aalen estimate of the sur-

vivor function is always greater than the Kaplan–Meier

estimate at a specified time. For small samples, the Nelson–

Aalen estimator is better than the Kaplan–Meier estimator

(Collett, 2003, p. 22).

The Nelson–Aalen estimator is given by

SNAðtÞ ¼Yk

j¼1

exp��dj=nj

�: (25)

Since expð�dj=njÞz1� ðdj=njÞ ¼ ðnj � djÞ=nj, whenever dj is

small compared to nj, which it is except toward the end

of the study, the Kaplan–Meier estimate given by Eq.

(20) closely approximates the Nelson–Aalen estimate in

Eq. (25).

Then, as usual, we can obtain the Nelson–Aalen failure

function directly from the survivor estimator:

FNAðtÞ ¼ 1� ~SðtÞ: (26)

The cumulative hazard H(t) at time t is, by definition, the

integral of the hazard function. Because HðtÞ ¼ �log SðtÞ, we

find

HNAðtÞ ¼ �log SNAðtÞ ¼Xr

j¼1

dj

nj; (27)

the cumulative sum of estimated probabilities of failure in the

first r time intervals. Since the differences between adjacent

values of HNAðtÞ are estimates of the hazard function hNAðtÞafter being divided by the time interval (Collett, 2003, p. 33).

Thus,

hNAðtÞ ¼ dj

nj

�tjþ1 � tj

�; (28)

exactly as in Eq. (24).

The calculations for the Nelson–Aalen estimators for the

data from Table 2 are shown in Table 4, and a graph of the

survivor function is shown in Fig. 3. To obtain the probability

density functions associated with the Kaplan–Meier and

Nelson–Aalen estimators, we can use Eq. (8), or Eq. (16).

The Kaplan-Meier Estimate of the Survivor Function

0

0.2

0.4

0.6

0.8

1

1.2

0 6 12 18 24 30 36 42 48 54 60 66 72 78 84 90 96

Time in days

S(t)

Fig. 2 – The Kaplan–Meier estimator for the survivor func-

tion underlying the data in Table 2.

7. The advantage of an investmentin information security

We hope and expect, of course, that an investment in infor-

mation security will provide us with some advantages. If it

did not, we would be foolish to make the investment. More

specifically, we expect that an investment in information se-

curity will result in greater freedom from successful attacks

on the systems being protected, so that the systems survive

longer before succumbing to an attack. Since no security is

perfect, we know that eventually the systems will succumb,

but our investment should delay that time.

The result of an investment should, then, be to move the

survivor curve to the right. Suppose that we track 100 systems

which are protected by an investment in additional informa-

tion security that mirrors the investment proposed for our

information infrastructure. This might occur contemporane-

ously with the gathering of the data in Table 2, so that the sys-

tems represented in Table 2 represent a control group. Thus,

in the enhanced group, one system fails on the first day,

another fails on the fifth day, and so forth until 34 systems

remain operational when the experiment is terminated on

the 100th day. The 34 surviving systems are right-censored.

Systems that fail on the 65th, 93rd, and the other entries

marked as negative times, are also censored, representing

failures due to causes unrelated to successful attacks such

as, perhaps, reliability failures.

Tracking the enhanced systems might also take place later,

and could even be the same systems as were followed in pre-

paring Table 2, suitably repaired following the attacks that

provided the data in Table 2 and enhanced according to the

proposed investment. In any event, suppose that following

protection of the new set of systems with the improved secu-

rity, we observe the attacks described by the failure times in

Table 5. Fig. 4 shows the survivor curve S0 that we experience

without the investment (from Table 2), and the survivor curve

Si that occurs following the investment (from Table 5). The

benefit produced by our investment is the area between the

curves.

Thus, the advantage we gain from our investment i is

given by

AðiÞ ¼Z N

0

SiðtÞ �Z N

0

S0ðtÞ ¼ Ei½T� � E0½T�; (29)

since, as is well known,

E½T� ¼Z N

0

tf ðtÞdt ¼Z N

0

SðtÞdt: (30)

Of course, it may be that SiðtÞ � S0ðtÞ is not always true. The

two curves may cross one or more times, but if AðiÞ > 0 the

benefits of the investment will eventually outweigh any short

term detriments. If, however, we restrict our attention to

the near term, say one year or a few years, it is possible that

AðiÞ > 0 but the value ofR t

0 SiðtÞ �R t

0 S0ðtÞ is less than zero for

t restricted to the period of interest.

Eq. (29) is a useful metric for measuring the advantage

of an investment, but it still fails to address the impact of

a successful attack on those information assets that have

their confidentiality, integrity or availability compromised

at the observed failure times. We know that the impact of

c o m p u t e r s & s e c u r i t y 2 5 ( 2 0 0 6 ) 5 7 9 – 5 8 8 585

Table 4 – Nelson–Aalen calculations

j tj dj nj exp(�dj/nj) SNA(t) FNA(t) hNA(t) fNA(t)

0 0 0 100 1.0000 1.0000 0.0000 0.0000 0.0000

1 0.5 2 98 0.9798 0.9798 0.0202 0.0011 0.0011

2 19.5 1 97 0.9897 0.9697 0.0303 0.0206 0.0200

3 20 1 96 0.9896 0.9597 0.0403 0.0030 0.0029

4 23.5 2 94 0.9789 0.9395 0.0605 0.0106 0.0100

5 25.5 2 92 0.9785 0.9193 0.0807 0.0109 0.0100

6 27.5 2 89 0.9778 0.8989 0.1011 0.0225 0.0202

7 28.5 1 88 0.9887 0.8887 0.1113 0.0227 0.0202

8 29 1 86 0.9884 0.8784 0.1216 0.0233 0.0205

9 29.5 1 85 0.9883 0.8682 0.1318 0.0235 0.0204

10 30 1 84 0.9882 0.8579 0.1421 0.0238 0.0204

11 30.5 2 82 0.9759 0.8372 0.1628 0.0244 0.0204

12 31.5 1 81 0.9877 0.8269 0.1731 0.0082 0.0068

13 33 3 78 0.9623 0.7957 0.2043 0.0769 0.0612

14 33.5 1 76 0.9869 0.7853 0.2147 0.0263 0.0207

15 34 1 75 0.9868 0.7749 0.2251 0.0267 0.0207

16 34.5 3 72 0.9592 0.7433 0.2567 0.0417 0.0310

17 35.5 2 70 0.9718 0.7224 0.2776 0.0571 0.0412

18 36 1 68 0.9854 0.7118 0.2882 0.0147 0.0105

19 37 3 65 0.9549 0.6797 0.3203 0.0923 0.0627

20 37.5 2 63 0.9688 0.6585 0.3415 0.0317 0.0209

21 38.5 4 59 0.9345 0.6153 0.3847 0.0678 0.0417

22 39.5 3 56 0.9478 0.5832 0.4168 0.1071 0.0625

23 40 2 54 0.9636 0.5620 0.4380 0.0741 0.0416

24 40.5 3 51 0.9429 0.5299 0.4701 0.1176 0.0623

25 41 1 49 0.9798 0.5192 0.4808 0.0204 0.0106

26 42 2 47 0.9583 0.4976 0.5024 0.0851 0.0423

27 42.5 2 45 0.9565 0.4759 0.5241 0.0889 0.0423

28 43 4 41 0.9070 0.4317 0.5683 0.0650 0.0281

29 44.5 4 37 0.8975 0.3875 0.6125 0.2162 0.0838

30 45 1 35 0.9718 0.3766 0.6234 0.0286 0.0108

31 46 3 32 0.9105 0.3429 0.6571 0.1875 0.0643

32 46.5 2 30 0.9355 0.3207 0.6793 0.1333 0.0427

33 47 1 29 0.9661 0.3099 0.6901 0.0345 0.0107

34 48 1 28 0.9649 0.2990 0.7010 0.0714 0.0213

35 48.5 4 24 0.8465 0.2531 0.7469 0.3333 0.0844

36 49 1 23 0.9575 0.2423 0.7577 0.0435 0.0105

37 50 3 20 0.8607 0.2086 0.7914 0.1500 0.0313

38 51 1 18 0.9460 0.1973 0.8027 0.0556 0.0110

39 52 2 16 0.8825 0.1741 0.8259 0.1250 0.0218

40 53 1 14 0.9311 0.1621 0.8379 0.1429 0.0232

41 53.5 1 13 0.9260 0.1501 0.8499 0.0769 0.0115

42 54.5 1 12 0.9200 0.1381 0.8619 0.1667 0.0230

43 55 1 11 0.9131 0.1261 0.8739 0.0909 0.0115

44 56 1 10 0.9048 0.1141 0.8859 0.1000 0.0114

45 57 2 8 0.7788 0.0889 0.9111 0.1000 0.0089

46 59.5 1 7 0.8669 0.0770 0.9230 0.0952 0.0073

47 61 1 6 0.8465 0.0652 0.9348 0.0000 0.0000

a successful attack on an information asset varies over

time. A compromise of the confidentiality of the planned

landing on Omaha Beach on D-day would have had

enormous impact on June5, 1944, and a negligible impact

on June 7th. The impact of a successful attack on integrity

is much larger following creation and prior to making and

storing a backup of an information asset than the impact

that would result from a compromise after the backup

is safely stored. Such a collapse of the loss function is

characteristic of information security, although whether

the decline takes place very rapidly, as in the D-day

example, or degrades gradually over time, depends upon

circumstances.

Alternatively, consider the loss functions associated with

an organization that keeps customer accounts in a database

that is updated once per day, say at midnight, based on

a transactions file accumulated throughout the day. The loss

function for the database itself is constant throughout the

day:

0; prior to midnight and the creation of the databasen ¼ n; between midnight and the following midnight0; after midnight when the database has been replaced

by an updated database

:

(31)

If a duplicate copy of the accounts database is made con-

currently with the following midnight’s update, and stored

c o m p u t e r s & s e c u r i t y 2 5 ( 2 0 0 6 ) 5 7 9 – 5 8 8586

securely offsite, the exposure for destruction or corruption of

the accounts database can be as little as the cost of recovering

and installing the backup – much less than the cost of creating

the database from scratch. The loss function for the transac-

tion file, on the other hand, is a sawtooth function, the value

of which is zero at midnight and increases monotonically as

transactions accumulate throughout the day, but which

returns to zero value the following midnight when the trans-

actions are used to update the accounts database and a new

transactions file is created for the following day. The loss

function nðtÞ for all the information assets contained in our in-

formation infrastructure is, of course, a sum of the individual

loss functions for each asset.

Our estimation processes have provided us with the proba-

bility densities f0 and fi we need, so, given a loss function nðtÞrepresenting the loss we would experience from a successful

attack on our unimproved infrastructure as a function of

time, the expected loss at t without the proposed investment is

E0½n� ¼Z N

0

nðtÞf0ðtÞdt; (32)

and, similarly, for the expected loss following making the pro-

posed investment in information security,

Ei½n� ¼Z N

0

nðtÞfiðtÞdt: (33)

Since we expect the loss to be less following our invest-

ment, the expected benefit from our investment is

The Nelson-Aalen Estimate of the Survivor Function

0

0.2

0.4

0.6

0.8

1

1.2

0 6 12 18 24 30 36 42 48 54 60 66 72 78 84 90 96

Time in days

S(t)

Fig. 3 – The Nelson–Aalen estimator for the survivor func-

tion underlying the data in Table 2.

E0 nðtÞ �Ei nðtÞ > 0�½�½ , and the expected net benefit is

BðiÞ ¼ E0½nðtÞ� � Ei½nðtÞ� � i: (34)

Of course, if nðtÞ ¼ n is constant, then

E0½nðtÞ� � Ei½nðtÞ� ¼ E0½n� � Ei½n� ¼ n� n ¼ 0;

so BðiÞ ¼ �i, confirming our intuition that no security is ever

perfect and eventually a compromise will occur. But, if our in-

vestment i is such that the survivor curve moves sufficiently

far to the right, then the occurrence of a successful attack

could be delayed beyond the collapse of the loss function. As

SiðtÞ moves to the right, so does fi(t). If tc is the time at which

the loss function collapses, then, since nðtÞ ¼ 0 for t > tc,

E0½nðtÞ� � Ei½nðtÞ� ¼Z tc

0

nðtÞf0ðtÞdt�Z N

tc

nðtÞf0ðtÞdt

�"Z tc

0

nðtÞfiðtÞdt�Z N

tc

nðtÞfiðtÞdt

#

¼Z tc

0

nðtÞf0ðtÞdt�Z tc

0

nðtÞfiðtÞdt : ð35Þ

Then, if fi has moved to the right, making fi(t) small when

t< tc, the second integral is small, and

BðiÞ¼ E0½nðtÞ� � Ei½nðtÞ� � i

z

Ztc

0

nðtÞf0ðtÞdt� i: ð36Þ

Thus, the collapse of the loss function will make our invest-

ment worthwhile if i is less than the near-term loss expected if

the investment is not made.

8. Future research

To enable the use of expected loss in these mathematical

models for evaluation of proposed investments in information

security, and the use of the metrics in tracking the evolution of

security in information infrastructures as investments are

implemented, more research is needed.

Studies of failure time data can involve epidemiological

studies of the entire information infrastructure, or can use

a cross-sectional study of a representative sub-infrastructure

based on concurrent measurements of samples of individual

systems followed prospectively (a cohort study), or of several

Table 5 – Notional computer system failure data following an investment i in information security

0.5 5.5 26.5 57 57.5 60 60 62 62 65.5

�65.5 67.5 68.5 70.5 72.5 72.5 72.5 72.5 72.5 75

75.5 77 77 78 78 79 80 80 80.5 81

81 82.5 82.5 84.5 84.5 84.5 85.5 85.5 86 86

87.5 87.5 87.5 89 89 89 91 91.5 91.5 91.5

93 93 93.5 �93.5 94.5 94.5 96 96 �96 97.5

98 98 99 99 99 99 �100 �100 �100 �100

�100 �100 �100 �100 �100 �100 �100 �100 �100 �100

�100 �100 �100 �100 �100 �100 �100 �100 �100 �100

�100 �100 �100 �100 �100 �100 �100 �100 �100 �100

The sign � indicates censoring. Values represent days.

c o m p u t e r s & s e c u r i t y 2 5 ( 2 0 0 6 ) 5 7 9 – 5 8 8 587

samples with retrospective measurements (a case–control

study). Alternatively, we can use a randomized controlled

trial to study two or more parallel, randomized cohorts of

systems, one of which (the control group) receives no

security enhancements, and the others of which are en-

hanced by a proposed investment or investments in infor-

mation security. Such studies should be undertaken to

evaluate the overall utility of various approaches to using

failure time data in an information security environment,

and addressing a wide variety of different practices and

technologies to determine if this approach is more effective

in some settings or for some types of practices or

technologies.

9. Conclusion

Too often information security investment decisions are

based on criteria that are at best qualitative, and at worst little

more than fear, uncertainty and doubt derived from anecdotal

evidence. Security is the inverse of risk. Because security is at

its best when nothing happens, it is notoriously difficult to

measure. But we can measure risk by calculating expected

loss, and a reduction in expected loss is a measure of the

change in security posture that accrues to an information

infrastructure following an investment in a new security

practice or technology. Unfortunately, there are little data

available to allow us to understand the probabilities of

successful attacks that we need in order to calculate expected

loss.

We sought a way to obtain the probabilities needed so we

can actually use the expected net benefit Eq. (1). By collecting

data on the experiences of two separate populations, one

which has the benefit of the investment and the other – a con-

trol group – which does not, we can compute the Kaplan–

Meier or Nelson–Aalen estimates of the probability density

functions with and without the investment. Having the prob-

abilities, and knowing the value of our information assets, we

can calculate the respective expected losses, and to use those

Survivor Functions Before and

After an Investment i

0

0.2

0.4

0.6

0.8

1

1.2

Time

Pro

bab

ility

S0(t)

Si(t)

0 7.5 15 22.5 30 37

.5 45 52.5 60 67

.5 75 82.5 90 97

.5

Fig. 4 – The survivor function S0(t) before an investment and

Si(t) following an investment i in Information security.

expected values to measure the advantage that will be

realized from the investment. We will then be able to make

an informed decision as to whether a proposed investment

is wise. Implemented in an operational environment, this

method can change the way security investments are consid-

ered and made.

r e f e r e n c e s

Altshuler B. Theory for the measurement of competing risks inanimal experiments. Mathematical Bioscience 1970;6:1–11.

Bedford Tim, Cooke Roger. Probabilistic risk analysis: foundationsand methods. Cambridge, UK: Cambridge University Press;2001.

Carroll John M. Information security risk management. In:Hutt Arthur, et al., editors. Computer security handbook.3rd ed. NY: John Wiley & Sons; 1995. p. 3.1–320.

Cavusoglu Huseyin, Mishra Birendra, Raghunathan Srinivasan. Amodel for evaluating it security investments. Communica-tions of the ACM 2004;47(7):87–92.

Collett David. Modelling survival data in medical research.2nd ed. Boca Raton: Chapman & Hall/CRC; 2003.

Crowder Martin. Classical competing risks. Washington, DC:Chapman & Hall/CRC; 2001.

Gordon Lawrence A, Loeb Martin P. The economics of informationsecurity investment. ACM Transactions on Information andSystem Security November 2002;5(4):438–57.

Gerber Mariana, von Solms Rossouw. From risk analysisto security requirements. Computers and Security 2001;20:580.

Iheagwara Charles. The effect of intrusion detection manage-ment methods on the return on investment. Computers andSecurity 2004;23:213–28.

Kalbfleisch John D, Prentice Ross L. The statistical analysis offailure time data. 2nd ed. Hoboken, NJ: John Wiley & Sons;2002.

Kaplan EL, Meier P. Nonparametric estimation from incompleteobservations. Journal of the American Statistical Association1958;53:457–81.

Kumamoto Hiromitsu, Henley Ernest J. Probabilistic risk assess-ment for engineers and scientists. 2nd ed. New York: IEEEPress; 1996. p. 266ff.

Leach John. Security engineering and security ROI. p. 2, <http://www.compseconline.com/free_articles/cose_22_6.pdf>; 2003[accessed 7/22/2004].

Nelson Wayne. Hazard plotting for incomplete failure data.Journal of Quality Technology 1969;1:27–52.

Nelson Wayne. Theory and applications of hazard plotting forcensored failure data. Technometrics 1972;14:945–65.

Ozier Will. Risk analysis and assessment. In: Tipton Harold F,Krause Micki, editors. Information security managementhandbook. 4th ed. Boca Raton: Auerbach; 1999. p. 247–85.

Ryan Julie J.C.H., Jefferson Theresa I. The use, misuse, andabuse of statistics in information security research.Managing technology in a dynamic world. In: Proceedings ofthe 2003 American society for engineering managementconference, St. Louis, Missouri, October 15–18, 2003.p. 644–53.

Therneau Terry M, Grambsch Patricia M. Modeling survival data:extending the Cox model. New York: Springer; 2000.

Wei Huaqiang, Frinke Deb, Carter Olivia, Ritter Chris, 2001. Costbenefit analysis for network intrusion detection systems. In:CSI 28th annual computer security conference, Washington,DC, <http://wwwcsif.cs.ucdavis.edu/%7Ebalepin/new_pubs/costbenefit.pdf> [accessed 7/22/2004].

c o m p u t e r s & s e c u r i t y 2 5 ( 2 0 0 6 ) 5 7 9 – 5 8 8588

Julie JCH Ryan is a member of the faculty at George Washing-

ton University in Washington, DC. Earlier she served as Pres-

ident of the Wyndrose Technical Group, Inc., a company

providing information technology and security consulting ser-

vices. She was also a Senior Associate at Booz Allen & Hamil-

ton and a systems analyst for Sterling Software. In the public

sector she served as an analyst at the Defense Intelligence

after beginning her career as an intelligence officer after grad-

uating from the Air Force Academy. She holds a Masters from

Eastern Michigan University and a D.Sc. from George Wash-

ington University.

Daniel J. Ryan is a Professor at the Information Resources

Management College of the National Defense University

in Washington, DC. Prior to joining academia, he served

as Executive Assistant to the Director of Central Intelli-

gence, and still earlier as Director of Information Security

for the Office of the Secretary of Defense. In the private

sector, he served as a Corporate Vice President of SAIC

and as a Principal at Booz Allen & Hamilton. He holds

a Masters in mathematics from the University of Maryland,

an MBA from California State University, and a JD from the

University of Maryland.