20
Risk Analysis, Vol. 8, No. 1, 1988 An Investigation of Uncertainty and Sensitivity Analysis Techniques for Computer Models Ronald L. Iman' and Jon C. Helton' Received November 4, 1986; revised March 23, 1987 Many different techniques have been proposed for performing uncertainty and sensitivity analyses on computer models for complex processes. The objective of the present study is to investigate the applicability of three widely used techniques to three computer models having large uncertainties and varying degrees of complexity in order to highlight some of the problem areas that must be addressed in actual applications. The following approaches to uncertainty and sensitivity analysis are considered: (1) response surface methodology based on input determined from a fractional factorial design; (2) Latin hypercube sampling with and without regression analysis; and (3) differential analysis. These techniques are investi- gated with respect to (1) ease of implementation, (2) flexibility, (3) estimation of the cumulative distribution function of the output, and (4) adaptability to different methods of sensitivity analysis. With respect to these criteria, the technique using Latin hypercube sampling and regression analysis had the best overall performance. The models used in the investigation are well documented, thus making it possible for researchers to make compari- sons of other techniques with the results in this study. KEY WORDS: Latin hypercube sampling; differential analysis; response surface; Monte Carlo: risk assessment. 1. INTRODUCTION Computer models are used in many settings to implement mathematical models for systems that are too complex to analyze directly. Such models have applications in economics, energy exploration, nuclear risk assessment, and many other fields. Typi- cally, these models represent a variety of phenom- ena. For example, the work presented here arises from the Nuclear Regulatory Commission-sponsored MELCOR program at Sandia National Laboratories, which is developing a system of models to represent accident progression, thermal-hydraulic phenomena, radionuclide behavior and transport, and environ- mental consequence analysis for severe reactor acci- Sandia National Laboratories, Division 6415, Albuquerque, New Mexico 87185. *Arizona State University, Tempe, Arizona 85281. dents.(') There are many uncertainties associated with both the development and application of complex models. Understanding of these uncertainties and their causes is required to effectively interpret model predictions. The analysis of uncertainties and sensitivities associated with such models plays an important part in their development and application. Typically the examination of the uncertainty in large systems such as MELCOR is a very complex undertaking and is performed in a sequential manner. The analysis first involves the individual components of the system, and then, at a later stage the system is examined in its entirety. In the first stage, much effort is directed at understanding and simplifying the individual com- ponents in the system. In the second stage, effort is directed at pulling this understanding together for use in an integrated analysis. The purpose of this 71 0272-4332/8R/0300-0071$~ ~/lOl9RS Society for Risk Analysis

An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

  • Upload
    others

  • View
    5

  • Download
    0

Embed Size (px)

Citation preview

Page 1: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

Risk Analysis, Vol. 8, No. 1, 1988

An Investigation of Uncertainty and Sensitivity Analysis Techniques for Computer Models

Ronald L. Iman' and Jon C. Helton'

Received November 4, 1986; revised March 23, 1987

Many different techniques have been proposed for performing uncertainty and sensitivity analyses on computer models for complex processes. The objective of the present study is to investigate the applicability of three widely used techniques to three computer models having large uncertainties and varying degrees of complexity in order to highlight some of the problem areas that must be addressed in actual applications. The following approaches to uncertainty and sensitivity analysis are considered: (1) response surface methodology based on input determined from a fractional factorial design; (2) Latin hypercube sampling with and without regression analysis; and (3) differential analysis. These techniques are investi- gated with respect to (1) ease of implementation, (2) flexibility, (3) estimation of the cumulative distribution function of the output, and (4) adaptability to different methods of sensitivity analysis. With respect to these criteria, the technique using Latin hypercube sampling and regression analysis had the best overall performance. The models used in the investigation are well documented, thus making it possible for researchers to make compari- sons of other techniques with the results in this study.

KEY WORDS: Latin hypercube sampling; differential analysis; response surface; Monte Carlo: risk assessment.

1. INTRODUCTION

Computer models are used in many settings to implement mathematical models for systems that are too complex to analyze directly. Such models have applications in economics, energy exploration, nuclear risk assessment, and many other fields. Typi- cally, these models represent a variety of phenom- ena. For example, the work presented here arises from the Nuclear Regulatory Commission-sponsored MELCOR program at Sandia National Laboratories, which is developing a system of models to represent accident progression, thermal-hydraulic phenomena, radionuclide behavior and transport, and environ- mental consequence analysis for severe reactor acci-

Sandia National Laboratories, Division 6415, Albuquerque, New Mexico 87185.

*Arizona State University, Tempe, Arizona 85281.

dents.(') There are many uncertainties associated with both the development and application of complex models. Understanding of these uncertainties and their causes is required to effectively interpret model predictions.

The analysis of uncertainties and sensitivities associated with such models plays an important part in their development and application. Typically the examination of the uncertainty in large systems such as MELCOR is a very complex undertaking and is performed in a sequential manner. The analysis first involves the individual components of the system, and then, at a later stage the system is examined in its entirety. In the first stage, much effort is directed at understanding and simplifying the individual com- ponents in the system. In the second stage, effort is directed at pulling this understanding together for use in an integrated analysis. The purpose of this

71 0272-4332/8R/0300-0071$~ ~ / l O l 9 R S Society for Risk Analysis

Page 2: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

72 Iman and Helton

study is to investigate three widely used approaches for uncertainty and sensitivity analysis of individual models. Thus, this study focuses on techniques for use in the first stage of an analysis.

For this investigation, it is convenient to think of a model as a function Y = f( X , , . . . , x k , t ) of the independent variables X , , . . . , x k and possibly also of time t . The variables X I , . . . , X , can be used to repre- sent a variety of phenomena within the model. For example, they might represent common properties such as temperature and pressure or other entities such as parameters in statistical distributions, branch points in the evolution of a process, or different submodels within a larger model. Uncertainty analysis involves the determination of the variation or impre- cision in Y that results from the collective variation in the model variables X, , . . . , X k . Summarizing and displaying the uncertainty associated with Y involves many questions, such as (1) what is the range of Y, (2) what are the mean and median of Y, (3) what is the variance of Y, (4) what are the lower and upper 5% quantiles for Y, and (5) are there any discontinui- ties associated with the distribution of Y? A conveni- ent tool for providing such information is the esti- mated cumulative distribution function (cdf) for Y. However, the estimated distribution function of Y can be interpreted in a probabilistic sense only if the model variables X,, . . . , X k have meaningful prob- ability distributions associated with them. Fre- quently, this is not possible as the model input variables may lack an adequate data base. Since the first stage of the uncertainty analysis involves much screening of the model variables, the estimated cumulative distribution function can be used as a summary tool in this part of the analysis without an undue concern about its probabilistic interpretation.

An area closely related to uncertainty analysis is sensitivity analysis. The importance of sensitivity analysis lies in the guidance it provides with respect to the identification of the important contributors to uncertainty in Y. Sensitivity analysis involves the determination of the change in the response of a model to changes in individual model parameters and specifications. Thus, sensitivity analysis is used to identify the main contributors to the variation or imprecision in Y.

The models to which uncertainty and sensitivity analyses are applied are often large and complex and frequently display many of the following properties:

There are many input and output variables. 0 The model is time consuming (i.e., expensive)

to run on a computer.

Alterations to the model are difficult and time consuming. It is difficult to reduce the model to a single system of equations. Discontinuities exist in the behavior of the model. Correlations exist among the input variables, and the associated marginal probability dis- tributions are often nonnormal. Model predictions are nonlinear, multi- variate, time-dependent functions of the in- put variables. The relative importance of individual input variables is a function of time.

For an approach to uncertainty and sensitivity analy- sis to be viable, it must be applicable to models possessing some or all of the preceding characteris- tics.

This study examines the following often used approaches to uncertainty and sensitivity analysis:

Response surface replacement for the com- puter model; Modified Monte Carlo analysis as exem- plified by Latin hypercube sampling; Differential analysis.

These techniques are applied to three models with large uncertainties:

Pathways (a model for environmental radio- nuclide movement); MAEROS (a model for multicomponent aerosol dynamics); DNET (a model for salt dissolution in bed- ded salt formations).

The preceding models were selected because each is well documented, has been extensively used in risk assessment applications, and displays many of the eight model characteristics indicated above.

T h s paper is a condensation of a larger report by the authors.(2) This report contains many exam- ples and analysis details that were eliminated from the present paper to conserve space.

2. METHODS TO BE INVESTIGATED

2.1. Response Surface Replacement

A response surface replacement for a computer model is based on using an experimental design to select a set of specific values and pairings of the

Page 3: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

Uncertainty and Sensitivity Analysis Techniques 73

input variables X,, . . . , X, that are used in making n runs of the computer model. The model output and input Xl,, . . ., X,,, i = 1,. .. , n, are used to esti- mate the parameters of a general linear model of the form

Y=Po+CP,x , (1) i

The estimated model is known as a fitted response surface, and it is this response surface that is used as a replacement for the computer model. All inferences with respect to uncertainty analysis and sensitivity analysis for the computer model are then derived from this fitted model. Selection of the actual design to be used in determining the specific values of the inputs in an analysis can be made only after careful consideration of the model and the variables associ- ated with it. The expense of evaluating the model (and subsequent limitation of the number of com- puter runs) may exert a strong influence on the selection of the design to be used. A good discussion on choosing effective designs for application to com- puter models is presented in Ref. 3. Since this study is not intended to be a detailed investigation of response surface techniques, only an often used ap- proach based on factorial designs will be used for purposes of illustration.

A discussion of the application of a central composite design to a computer model that describes a scrubbing process that removes radioactive material from steam generated during a nuclear reactor melt- down accident is given in Ref. 4. An example of a sophisticated sequential attempt at fitting a response surface to predict peak cladding temperature in a loss of coolant accident at a nuclear reactor appears in Ref. 5. Other examples of the use of fractional factorial designs in uncertainty analysis and sensitiv- ity analysis include Refs. 6-10.

2.2. Modified Monte Carlo (Latin Hypercube Sampling)

A possible alternative to the response surface replacement approach is a type of stratified Monte Carlo sampling known as Latin hypercube sampling (LHS). As originally described,(") LHS operates in the following manner to generate a sample of size n from the k variables X,, ..., X,. The range of each variable is divided into n nonoverlapping intervals on the basis of equal probability. One value from each interval is selected at random with respect to

the probability density in the interval. The n values thus obtained for X, are paired in a random manner with the n values of X,. These n pairs are combined in a random manner with the n values of X, to form n triplets, and so on, until a set of n k-tuples is formed. This set of k-tuples is the Latin hypercube sample. Thus, for given values of n and k, there exist (n! ) , - ' possible interval combinations for a Latin hypercube sample.

Because of the random pairing of intervals in the mixing process, there exists the possibility of inducing undesired pairwise correlations among some of the variables in a Latin hypercube sample, just as there is in simple Monte Carlo sampling. This is more likely to occur if n is small. Such correlations can be avoided by modifying the Latin hypercube sample by restricting the (n!),-' possible interval pairings through the use of a technique introduced by Iman and Conover('2) and implemented by a computer program(13) developed at Sandia National Laboratories. Restricting the pairing in this manner preserves the fundamental nature of Latin hypercube sampling but replaces the random matching of inter- vals with a method that keeps all of the pairwise rank correlations among the k input variables very close to zero, and thus ensures that no unwanted large pairwise correlations will exist between input vari- ables. Additionally, the restricted pairing technique in Ref. 12 can be used to induce correlation when it is desired as is shown in Section 4 of this paper.

In uncertainty analysis associated with Latin hypercube sampling, it is desired to estimate the distribution function and the variance for the par- ticular output variable(s) Y under consideration. Due to the probabilistic nature of Latin hypercube sam- pling, it is possible to estimate these entities directly from the model output associated with the sample just as in simple Monte Carlo sampling.

Discussions of the use of LHS in uncertainty and sensitivity analysis are given in Refs. 14-19. Applications of Latin hypercube sampling appear in Refs. 3 and 20-22. Examples showing good and bad features of LHS in uncertainty analysis appear in Ref. 23.

2.3. Differential Analysis

The last method considered in this study is based on a Taylor series expansion and the associ- ated partial derivatives. With this approach, the de- pendent variable of interest is treated as a function f of the independent variables Xl, ..., X, and a first-

Page 4: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

74 Iman and Helton

order Taylor series expansion for f is constructed about some vector X, = ( X,,, . . . , Xk0) of base-case

variables, are given in Ref. 2. These variables are assumed to behave independently of one another.

values for the variables X = ( X,, . . . , Xk). This Taylor series approximation forms the basis for uncertainty and sensitivity analysis techniques based on differ-

3.2. Selection of the Values of the Input Variables Used in the Analysis

entiation. The first step in a differential analysis is the

generation of the partial derivatives required in the series. If the function f is relatively simple, then it may be possible to generate these derivatives analyti- cally or by simple differencing schemes; however, complex numerical procedures are often required. For uncertainty analysis, the Taylor series approxi- mation can be used in conjunction with Monte Carlo simulation to estimate distribution functions. For sensitivity analysis, the coefficients in a Taylor series can be normalized as will be discussed in Section 3.4.

Differential techniques have been widely used in uncertainty and sensitivity analysis and several intro- ductory treatments are available, e.g., Refs. 24-26. Examples of the use of differential techniques are given in Refs. 27-34. A comparison of LHS and differential analysis appears in Ref. 19.

3. RESULTS BASED ON THE PATHWAYS MODEL

3.1. The Pathways Model

Description of the Pathways Model. The Path- ways model represents environmental radionuclide movement and consists of a system of linear, con- stant coefficient differential equations. The Pathways model has been well documented in a four volume set of NRC reports (Refs. 35-38). The form of the model considered in the present analysis is described in Ref. 22,

Variables Considered in the Analysis of the Path- ways Model. The Pathways model provides multi- variate output for the movement of radionuclides in four environmental components: (1) groundwater, (2) soil, (3) surface water, and (4) sediment. The follow- ing output variables are considered in this analysis:

Response Surface Replacement. For the response surface replacement a fractional factorial design with two levels (low and high) was used to represent each variable. The end points of the ranges were used to represent the low and high values of each variable. A l/2I3 fraction of a 220 factorial design was used to produce a fractional factorial design utilizing 2’ = 128 computer runs, where each level of each variable is used exactly 64 times. The actual design used was the smallest design that would allow for the estimation of all main effects and 91 potentially important interactions indicated in Table 6-3 of Ref. 36.

The Latin Hypercube Sample. For the portion of the analysis utilizing a Latin hypercube sample (LHS), 50 computer runs were made. The restricted pairing procedure of Iman and Conover(12) was utilized with the LHS, which kept all off-diagonal rank correla- tions close to zero. The associated rank correlation matrix for the LHS had a variance inflation factor (see Ref. 39) of 1.07, indicating negligible pairwise correlations within the sample.

The Differential Analysis. The differential analy- sis approach does not have an associated scheme for selecting specific values of the input variables; rather, it provides local information about each input vari- able at particular points that are deemed, a priori, to be of interest. A “base-case” vector consisting of the expected values of the input variables was defined, and the first-order partial derivatives of the depen- dent variables with respect to the independent vari- ables were calculated at this “base-case” value. In addition, partial derivatives of the dependent vari- ables were calculated at each of the 50 LHS input vectors in order to see how much variability would be encountered in local behavior from vector to vector. Thus, some indication of the reliability of extending local information to a global interpretation is provided.

Y, is the amount of Ra226 in soil, Y2 is the con- centration of Ra226 in soil, and Y, is the concentra- tion of Ra226 in surface water. For this study, the input to the Pathways model involves 20 variables that describe various physical phenomena as well as chemical properties of Ra226. Ranges and distribu- tions, along with a more detailed discussion of these

3.3. Uncertainty Analysis for the Pathways Model

LHS is based on a probabilistic input selection technique (as is a simple Monte Carlo procedure). When an output variable is graphed as an empirical cumulative frequency distribution, an estimate of the

Page 5: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

Uncertainty and Sensitivity Analysis Techniques 75

cdf is obtained directly. Model output based on input selected through the use of fractional factorial designs cannot be used to provide a direct estimate of the output cumulative distribution function since the input values are not selected in a probabilistic manner. Rather, the output cdf is estimated by using a Monte Carlo simulation with a fitted response surface. Results of the response surface approach with n = 100 in the Monte Carlo simulation are shown in Fig. 1 for Y, and Y,; Y, was very similar to Y,. Figure 1 also contains estimates based on a LHS utilizing restricted pairing with n = 50. For ease in comparing these estimates, an estimate of “truth” based on a random sample with n = 5 0 0 has also been included in Fig. 1.

The results in Fig. 1 show that the LHS estimate is in close agreement with the random sample esti- mate for both Y, and Y,. The estimate for Y, in Fig. 1 based on the response surface approach is in good agreement with the random sample estimate, but the estimate for Y, is not. The response surface estimate for Y, might be improved in a sequential manner through use of a more sophisticated experimental design, keeping in mind that Y, is only one of four output variables to be fitted. No attempts were made in this direction in this study since we were only attempting to demonstrate what can happen with a well-known and frequently used experimental design.

For the differential analysis (DA), the output cdf is estimated by approximating the underlying model with a first-order Taylor series and then using a Monte Carlo simulation as was done with the response surface fits to yield the estimate of the output cdf. A potential problem with this method lies in the local nature of the Taylor series expansion. To examine this point, the Taylor series expansion at the “base-case’’ vector was used with a Monte Carlo simulation and the results were compared with the direct estimate from LHS. These two results appear in Fig. 2 and are in reasonably good agreement except for the lower 10% of the curves and some noticeable separation in the middle. One might be tempted to say at this point that “base-case” expan- sions give reasonable results. To further illustrate local behavior, we selected four of the 50 LHS input vectors to represent other possible “ base-case” val- ues and used the same Monte Carlo sample with the Taylor series expansion about each of these points. The results are the curves labeled as 1, 2, 3, and 4 in Fig. 3 along with the direct LHS estimate. Of these estimates the one labeled as “2” actually agrees

1 .

.9 z 0 ; . 8 U z

. 7

z 2 .6 c 3 m . 5 LL c ul

0 - . 4

a I . k&lE C W L O OF KSPWSE S U K K E F I T

2. DIRECT ESTIMATE FROH LHS I N = M I c m W

. 1 3. R(u*lotl S W L E ESTIMATE I N = 5 0 0 1

0. 0 . 0 . 4 0 .80 1 . 2 1 . 6 2 . 0

AMOUNT ( C I I OF RA226 I N THE S O I L

0

I .

. 9

. 8

. 7

. 6

. 5

. 4

. 3

. 2

. 1

0 . 2 2 . 5 30 .0 37.5- 4 5 . 0 5 2 . 5 60 .0 6 7 . 5

CONC ( C I / L ) OF RA226 IN SURFACE WC+TER X l O I 5

Fig. 1. Response surface estimates of CDFs for Yl (amount of RA 226 in the soil) and Y, (concentration of RA 226 in the surface water).

better with the LHS estimate than does the “base- case” estimate in Fig. 2. The estimate labeled “1” is quite good in the lower tail but is not close in the upper tail. The estimates labeled “3” and “4” are both extremely poor. Additionally, we caution the reader concerning the difficulty of obtaining agree- ment on a “base-case’’ in many real world situations.

Page 6: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

Iman and Helton

DIRECT ESTIYATE I

rRoYLH8 A+-+

- . 6 0 0 -.ZOO ,200 , 6 0 0 1 . 0 0 1 . 4 0 1 . 8 0

AMOUNT OF RA 226 IN THE SOIL

Fig. 2. “Base-case” Taylor series expansion used to estimate the CDF for Y,.

z L.ooo 0 + .goo0

0 zj . 8 0 0 0

U z .7000 0

3

- I- . 6 0 0 C

m a 9 . 4 ? 0 C

c

0 p , 3 0 0 c

W c a . Z O C :

z c . I C C C v) W

0. o o c - . 6 0 0 0.00 .6CC 1 . 2 0 1 . 8 0 2 . 4 0 3.00 3.60

AMOUNT OF R A 226 IN THE SOIL

Fig. 3. Estimates of the CDF of Y, for various potential “base cases.”

3.4. Sensitivity Analysis for the Pathways Model

There are several methods for quantifying the relative importance of the input variables to a com- puter model. However, these methods do not neces- sarily yield the same conclusions. In this subsection several different methods of quantifying input vari- able importance are presented and compared.

Ranking Input Variables on the Basis of Normal- ized Coefficients. A linear regression model and a finite Taylor series can each be thought of as a linear model of the form appearing in (1.1). The coeffi- cients pi in (1.1) depend on the units used for the input variables; as a result, the p, will change as the units for the variables are changed. Therefore, it is necessary to normalize the coefficients to remove the effect of the units. For linear regression, this involves use of standardized regression coefficients.(40) For the differential analysis, each partial derivative in the Taylor series expansion is standardized as ( a Y/aX,)( ux, /u,,), where ax, is the population stan- dard deviation of X,, and uv is the estimated stan- dard deviation of the output variable. The rankings based on standardized coefficients are given in Table I for each of the three Pathways output variables under the column headings DA (differential analysis), LHS (response surface from Latin hypercube sam- pling), and RS (response surface from the fractional factorial design).

Within Table I, the DA and LHS techniques agree on the order of the top four variables for Y, and Y,, but show considerable disagreement after rank 1 for Y,. The RS technique agrees with DA and LHS on rank 1 for Yl, Y,, and Y3 but shows mod- erate to severe disagreement on the other ranks. Notable is the rank of 4 assigned to X I , under Y, by RS, while the LHS technique assigns rank 20. Also, under Y,, the RS technique assigns rank 17 to X,, , while LHS assigns rank 3. In both of these latter cases the D A technique gives an intermediate rank. In considering the entries in Table I, it is important to keep in mind that disagreements among rankings by the three techniques for variables of lesser impor- tance are of no practical concern since these vari- ables have little or no impact on model output.

A second normalization procedure involves mul- tiplication by X,,/Y0, where X,o and Yo correspond to a base-case run of the computer model. In the case of regression, the coefficients are normalized as p,( X,o/Yo), and in the differential analysis the par- tial derivatives are normalized as (aY/aX,)( X,o/ Yo). The resultant coefficients indicate the effect on the dependent variable of equivalent fractional changes of base-case values for the individual input variables. Such coefficients are frequently referred to as nor- malized sensitivity coefficients. The rankings from this normalization appear in Ref. 2 but are not shown here to conserve space. The normalization rankings in Ref. 2 showed that the best agreement

Page 7: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

Uncertainty and Sensitivity Analysis Techniques 77

Table I. Rankings of the Input Variables to the Pathways Model Based on Standardized Coefficients

Input Yl y2 y3 variable (X,) DA LHS RS DA LHS RS DA LHS RS

1 18 17 17 18 19 18 13 16 10 2 19 14 13 19 9 12 19 19 20 3 10 11 10 8 6 10 3 3 2 4 2 2 3 11 20 8 17 11 17 5 3 3 2 10 17 9 18 18 8

6 6 8 12 4 10 6 14 10 13 7 5 7 11 3 4 5 10 14 19 8 8 9 6 6 8 4 6 5 6 9 9 5 7 I 2 3 16 17 18

10 4 4 5 2 5 2 1 1 1

11 11 10 18 9 3 17 4 4 3 12 17 16 20 17 18 16 12 12 15 13 13 6 8 13 7 14 7 8 I 14 15 19 14 15 15 13 9 15 9 15 16 12 19 16 11 20 11 7 12

16 12 20 4 12 12 7 5 9 5 17 14 13 15 14 13 15 8 20 11 18 1 1 1 1 1 1 15 13 14 19 7 15 9 5 16 11 2 2 4 20 20 18 16 20 14 19 20 6 16

occurs for Y, where all three techniques pick the top three variables in order. Strong agreement also ex- isted between DA and LHS for Y, but these tech- niques disagreed after rank 1 for Y,. The RS tech- nique showed some degree of disagreement with both of these techniques under both Y, and Y,. In the three cases in Table I and the three cases not shown, LHS and DA always agree on the top rank and agree on the order of the top four ranks in three of the six cases. For these same cases, DA and RS agree on rank 1 in four of six cases. The same results also hold true for LHS and RS.

Ranking Input Variables on the Basis of their Contribution to the Variance of the Output. Another way of quantifying the relative importance of the individual input variables is by the percentage contri- bution each makes to the estimated variance of the output variable(s). The percentage contributions from a subset of the input variables (including some inter- action terms) appear in Table I1 along with the model R2 where appropriate. An examination of the values within Table I1 shows an overall good agree- ment on variable ranking for all three methods, par- ticularly for Yl and Y,. For Y,, the best agreement is between the LHS and DA methods but with no great

areas of disagreement. One might expect the LHS and RS methods to show reasonably good agreement depending on the degree of nonlinearity in the out- put; however, agreement of the LHS and RS meth- ods with the DA method may or may not be a desirable objective. The reason for this lies in the local nature of the Taylor series expansion and the selected " base-case.''

4. RESULTS BASED ON THE MAEROS MODEL

4.1. The MAEROS Model

Description of the MAEROS Model. MAEROS is a model that represents multicomponent aerosol dynamic^.(^'.^^) Mathematically, the model is a sys- tem of nonlinear differential equations. The model calculates aerosol composition and mass concentra- tion as a function of particle size and time. Thus, MAEROS is a more complex model than the Path- ways model considered in the previous section. The analysis problem considered in this section is de- scribed in Ref. 43.

Page 8: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

Table Il. Percentage Contribution to the Estimate of the Variance of the Output for Three Different Methods of Estimating the Variance

Iman and Helton

Output variable Input variable LHS RS DA

Yl x'l 6 6 6 x5 6 5 6 x, 3 2 2 XI, 3 2 2 X18 Model R2: 97.4% 94.5%

80 80 71 -

y2

y3

x6

x7

X8 x9

XlO XI1

'1 8 Model R2:

1 3 0 0 3 1

91 98.2%

1 2 1 1 2 0

89 96.3%

1 2 0 0 3 0 90 -

x3 5 8 6 XlO 15 66 71 Xll 8 5 5

9 7 6 8 x16 x19 3 0 0 x, XlO 0 6 0

- Model R2: 97.0% 91.2%

Variables Considered in the Analysis of the MAEROS Model. The output from MAEROS is multivariate as was the case with the Pathways model. For this comparison, four output variables were selected for consideration: total integrated con- centration ( Yl), total integrated deposition ( Y2), total suspended mass (Y,) , geometric expected value of particle diameter (Y,). The variables Y, and Yz are integrated over time (72,000 sec), while Y, is ex- amined primarily at 20 min. However, Y, and Y, are examined at multiple time steps later in this section when illustrating graphs of sensitivity measures over time. The 21 system variables used as input in the analysis are given in Ref. 2 along with the corre- sponding ranges, distributions, and restrictions.

Restrictions on the multivariate structure of the input required a rank correlation of 0.5 between X , and X , and also between X , and X,. The correlation is specified on ranks since this correlation measure is meaningful for both normal and non-normal distri- butions. The variables X , , X, , X , , and X , each had a loguniform distribution. An additional restriction involves the pairs ( X, , X 9 ) and ( X I 2 , XI , ) , for whch the inequalities lOOOX, s X , and X , , I X I , had to be satisfied. The manner in which these inequalities were implemented is explained in the next subsec- tion.

4.2. Selection of the Values of the Input Variables Used in the Analysis

The Response Surface Replacement. A response surface analysis on the MAEROS model was not performed using input values associated with an ex- perimental design. The reason for this decision lies in the complexity of the multivariate structure of the input including correlated variables and conditional distributions as indicated in the previous subsection. It is conceivable that an experimental design could be altered in some manner for use with a multivariate input structure such as specified with MAEROS, but it is doubtful that it could be done and still retain the spirit or intent of the design. A possibility is to ignore the multivariate input structure and generate a response surface utilizing some experimental design in the usual manner. The next step would be to use a Monte Carlo simulation with the response surface with input incorporating the required multivariate input structure. The problem with this approach is that pairs of variables may be created by the design that the modeler knows to be physically impossible or meaningless and for which the model may not run, or if it does run, the results may be useless. These points would influence the fitted response surface and could therefore lead to incorrect response surface

Page 9: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

Uncertainty and Sensitivity Analysis Techniques 79

predictions even when the simulation input is and Green’s function methods (Refs. 51-53). How- meaningful. ever, these methods are often very time consuming to

The Latin Hypercube Sample. The required cor- implement. relations in the MAEROS input are induced by generating a LHS in the usual sense and then con- trolling the individual pairing of variables to produce specific rank correlations as explained in Iman and 4.3. Uncertainty for the MAEROS Mode,

Estimation of the Distribution Function of the Output. Estimates of the output cdf obtained directly

Conover.(12) A LHS of size n = 50 was used with the 21 MAERoS input and produced rank

Of 0*48 and 0.43 for the pairs (x29 x,) from the LHS and indirectly by using a Monte Carlo simulation with the Taylor series ex- and ( x 5 T x6), where the target

were each 0.50. At the same time, of the 187 remain- ing pairs of variables in the rank correlation matrix,

with the largest being a nonsignificant value of 0.23. The condition X12 I X16 was handled by first

generating the required marginal distributions for

pansion at the base-case appear in Figs. and for r2 and r,, respectively. As a check on the adequacy

random sample of size has been included in each of these figures.

In Fig. 4, the LHS and random sample estimates

Only three pairs had larger than O*l0 of these estimates a third estimate based on a simple

‘12 and (l93)I in the usual manner,

Iunifom distributions On the coincide throughout, while the Taylor series expan- sion is not close for cumulative probabilities < 0.4. a transforma-

Of to xG = 0*5(x16 - - xu* In fact, over 35% of the predictions from this latter approach were negative (these values have been set equal to show reasonable agreement between the LHS estimate ( n = 50) and the random sample estimate ( n

After ‘16 is transformed in this the new variable X;r6 satisfies the restriction that Xl, I X;. Moreover, the marginal distribution of XI2 remains uniform on the interval (1,3) but the distribution of

in Fig. 4). Results in ~ i ~ .

xz is uniform, except for Some noticeable separation above the 0.90 is, x;r6 uniform on the (x1293)* It is quantile. The Taylor series expansion procedure

On the Of x12; that

shown that the generation of this conditional distri- bution creates a population correlation of 0.65 be- tween X12 and X;. The actual correlation observed between this pair of variables in the LHS of size 50

agrees with the other two estimates below the o.lo quantile but disagrees significantly for all quantiles above o.lo.

Two summary comments can be made about the was 0*58* The pair (x8, x9) was treated in a results in Figs. 4 and 5. First, the results based on manner.

The Differential Analysis. The differential analy- sis was performed at a “base-case” set of input values consisting of the expected values of the input variables. The’ MAEROS differential analysis was difficult to implement and by the time the differen- tial analysis was completed a separate computer pro- gram had been developed that was considerably more complicated than the original MAEROS program. As the preceding statement might lead one to suspect, the differential analysis required significant human and computer time to implement. Further, owing to the complexity of the implementation, it is also more

I

LHS and random sampling are in reasonably good agreement as was the case with the Pathways model. The second comment concerns the estimate from the Taylor series expansion. Results in Fig. 2 for the Pathways model showed reasonably good agreement with the LHS estimate, but the results in Figs 4 and 5 show areas of disagreement. This points out that the result from the Taylor series expansion may or may not be good. One never knows without additional checking. As indicated in the previous section, this results from trying to extrapolate local information to a global interpretation.

likely to contain errors. In retrospect, it might have been better to have approximated the desired deriva- tives with difference quotients.

It is possible to reduce the amount of computa- tional time required in a differential analysis by the use of specialized numerical procedures. These proce- dures are known as adjoint methods (Refs. 44-50),

4.4. Sensitivity Analysis for the MAEROS Model

In the previous section, the input variables to the Pathways model were ranked using two types of normalized coefficients and contribution to variance.

Page 10: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

80 Iman and Helton

1 .

. 9

. 8

.7

. 6 c 3

P c H1 . 4 0

0 w . 3 a c

I . RANDOM SAMPLE I N = 1 0 0 1 f .2 2 . Lns rk = 50 I t

ul W

3 . BASECASE TAYLOR S E R I E S

0 . ' ' ' ' ' ' ' ' ' ' ' ' 0 . 0 1 0 . 2 0 . 3 0 . 4 0 . 50.

T O T A L I N T E G R A T E D D E P O S I T I O N

Fig. 4. A comparison of three estimates of the CDF for Y, for the h4AEROS model.

These same ranking methods are presented in this subsection for the MAEROS model. In addition, sensitivity measures with time-dependent output utilizing the output variables Y, and Y, are dis- cussed.

Ranking the Input Variables. The results of using the ranking methods listed above were summarized in the previous section for the Pathways model in Tables I and 11. In the interest of shortening the discussion, all of the rankings in this subsection appear simultaneously in Table I11 and are given only for Y,. In addition to the previous methods, the ranking with standardized rank regression coeffi- cients (explained below) is also given in Table 111.

The results in Table I11 for Y, show disagree- ment both within and between ranking criteria. For the standardized coefficients criterion, the top three ranks for both the DA and LHS approaches agree, but there is disagreement on XI, where ranks of 4 and 11 are assigned. This is probably not a major concern since the dominant three variables have been identified. More noticeable is the disagreement be- tween the rankings for normalized coefficients. These rankings disagree considerably after rank 1, and the rankings as a whole disagree with the other ranking criteria, particularly on X18. The rankings under contribution to variance for LHS agree very well with those found under standardized rank regression

1 .

. 9 z 0

c u z

- .7

E! . 6 z c 2

P I-

fE! . 5

II] . 4 0

0

c

I: t m W

w . 3 a - . 2

. I

n " . 0 . 0 0 150. 3 0 0 . 4 5 0 . 6 0 0 . 750. 9 0 0 .

T O T A L SUSPENDED MASS

Fig. 5. A comparison of three estimates of the CDF for Y, for the MAEROS model.

coefficients (SRRC), but the agreement is not quite as good for the DA based rankings. The random sample and LHS rankings under SRRC agree very well with each other.

Use of Sensitivity Measures with Time-Dependent Output. When output is time dependent, the relative importance of the input variables may change with time. To illustrate this, the MAEROS output for Y3 and Y, was recorded at 65 time points. A measure of sensitivity can be calculated for each input variable versus each output variable at each of these 65 time points. The influence of a particular input variable X, on a particular output variable Y is easily sum- marized by making a plot with the measure of sensi- tivity on the vertical axis versus time on the horizon- tal axis. The measures of sensitivity utilized here to demonstrate the value of such plots are the standard- ized regression coefficient and the partial correlation coefficient. First, however, a brief explanation of the relationship between the two measures is provided.

The partial correlation coefficient differs from a simple correlation coefficient in that it measures the degree of linear relationship between the X, and Y after making an adjustment to remove the linear effect of all of the remaining variables. The actual calculation involves finding the inverse of the corre- lation matrix C between the individual X,'s and Y based on n computer runs. The inverse matrix can be

Page 11: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

Uncertainty and Sensitivity Analysis Techniques

c-’=

81

- - I/( 1 - R : , ) c12 . . . ‘lk - b , / ( 1 - R : )

c21 l / ( l - R : l ) ‘ 2 k - b 2 / ( 1 - R : ) . . . . . . ... ...

‘kl ‘ k 2 - . . l / ( l - R : k ) - b k / ( l - R ; )

- - b l / ( l - R : ) - b 2 / ( l - R t ) - . . - b k / ( l - R t ) 1 / ( 1 - R : ) -

Table 111. Rankings of the Input Variables to the MAEROS Model for the Output Variable Y,

coefficients coefficients to variance SRRC

DA LHS DA LHS DA“ LHS” LHS(n=50) RS(n =loo)

Standardized Normalized Contribution

Input variable

1 16 16 15 16 2 7 6 14 11 2.5“ 10 3 2 2 7 4 2.5‘ 2 2 2 4 18 10 18 10 5 14 21 16 21

6 5 4 11 I 4 5 5 7 20 14 20 14 8 13 12 2 5 6 9 9 19 4 15 9

10 19 20 19 20

11 4 11 5 12 12 10 18 9 19 4.5” 13 21 9 21 13 14 15 7 12 6 15 17 17 17 17

4 3 10 8

16 3 3 1 1 4.5” 8 7 17 12 13 8 9 6 9 18 1 1 6 3 1 1 1 1 19 8 8 10 8 7 6 20 11 15 13 18

21 6 5 3 2 3 3 4 R 2 = 75.88

“ Estimate obtained directly from coefficients.

‘ X, and X, jointly contribute 18.6%. ” X,, and X,, jointly contribute 17.2%.

Estimate obtained by regression with LHS output.

written as follows:

Page 12: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

82 Iman and Helton

Px,y and bJ; however, it is important to recognize that they yield different types of information. Standard- ized regression coefficients are derived from a condi- tional univariate distribution, while partial correla- tion coefficients come from a conditional bivariate distribution. Partial correlation coefficients allow one to judge the unique contribution that an explanatory variable can make. Standardized regression coeffi- cients are equivalent to partial derivatives in the standardized model.

Neither of these sensitivity measures provides a reliable measure of sensitivity when the relationship between Xi and Y is basically nonlinear but mono- tonic, or if there are outliers (extreme observations) present. To avoid this problem, each of the individ- ual variables XJ and Y can be replaced by their corresponding ranks from 1 to n and either sensitiv- ity measure can be computed on these ranks. This transformation converts the sensitivity measure from one of linearity between Xi and Y to one of mono- tonicity between XJ and Y. The result of this trans- formation is referred to as either the standardized rank regression coefficient (SRCC) or the partial rank correlation coefficient (PRCC). A computer program(54) for performing such calculations and plotting the results has been developed at Sandia National Laboratories.

Plots of the two sensitivity measures appear in Figs. 6 and 7 for X , and Y,, and X , and Y,. Although having different numerical values, both measures exhibit the same pattern of sensitivity. In Fig. 6, X, has little influence on Y, during the first 10,OOO sec but quickly shows a strong negative in- fluence after 10,OOO sec. In Fig. 7, X, shows a negative influence developing between X, and Y, out to 10,OOO sec and then very quickly changes to a strong positive influence.

The graphs in Figs 6 and 7 show quite clearly the sensitivity of the output to the influence of a single input variable and, in particular, bring out the fact that the sensitivity may be time dependent. One convenient way of summarizing the relative impor- tance of the input variables is to rank each of them on the basis of the absolute value of their sensitivity measure at each of several time steps. Such a summary is presented in Table IV for Y, at 11 different time steps; the summary was identical for both SRRC and PRCC (as was also the case in Table 111). This summary shows that X, is the most influential vari- able through 20 min and then decreases in impor- tance with respect to Y,. In contrast, X, has no

1 I Qo M 52 a4 &o

7Mmi *lcf

Fig. 6. A plot of the SRRC and the PRCC showing the influence of X, on Y, over time.

u)

ob

a6

n

I I 0'

1 1 I J

Qo Y 32 64 no w = Y *lo' Fig. 7. A plot of the SRRC and the PRCC showing the influence

of X, on Y, over time.

influence (rank 19) on Y, at 2 min but starting at 40 min it becomes the most important variable.

5. RESULTS BASED ON THE DNET MODEL

5.1. The DNET Model

Description of the DNET Model. The final model considered is a dynamic network (DNET) flow model

Page 13: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

Uncertainty and Sensitivity Analysis Techniques 83

Table IV. Ranking of the Influence of the Input Variables on Y2 and at 11 Selected Time SteDs

Time (min) X 2 4 6 8 10 20 40 60 120 180 240

1 2 3 4 5 6 7 8 9

10 11 12 13 14 15 16 17 18 19 20 21

4 3 1

13 20 19 6

14 16 15 8 5

11 9

18 2

12 17 21 7

10

4 2 1

13 19 11 8

14 17 6

18 5

15 9

10 3

21 20

7 12 16

5 4 1 7

21 6 8

14 16 9

18 3

13 15 11

2 17 19 12 10 20

5 3 1 6

21 7

10 12 15 8

19 4

14 16 11 2

17 13 9

18 20

5 3 1 7

17 6

12 21 14 9

11 4

13 19 10 2

16 15 8

20 18

7 9 1

15 21

3 13 10 18 8

14 4

19 6

11 2

16 17 12 5

20

12 6 5

10 19 1

16 8

15 20 14 9

21 4

13 3 7

17 11 2

18

20 20 17 17 6 6 6 9 7 8 8 10 8 7 7 12

16 10 9 7 1 1 1 1

11 13 19 6 9 11 13 16

17 15 18 21 21 21 20 20 14 16 12 15 10 9 11 11 15 19 15 14 4 4 3 3

19 17 14 13 3 2 2 2 5 5 5 5

18 14 16 19 12 18 21 18 2 3 4 4

13 12 10 8

used in simulating dissolution in bedded salt forma- tions. DNET simulates several physical processes, including (1) fluid flow, (2) salt dissolution, (3) ther- mal expansion, (4) fracture formation and closure, (5) subsidence, and (6) salt creep. In addition to this multivariate aspect, the output is nonlinear and time dependent. Submodels within DNET are applied sequentially to represent various processes, Because of feedback mechanisms governing the selection of different submodels and the complexity involved in treating various processes, the governing equations are not solved in an implicitly coupled fashion (i.e., simultaneously). This feature makes it difficult to implement a differential analysis. Thus, in this sec- tion only the techniques based on response surface replacement and Latin hypercube sampling are con- sidered. The DNET model has been well documented in Ref. 55.

Variables Considered in the Analysis of the DNET Model. The DNET model provides multivariate out- put for the process of salt dissolution in bedded salt formations. However, the only output variable used in this investigation is the rate of dissolution of a cavity in a bedded salt formation at 20 different time periods from 5 to lo5 yr. The input to DNET for this application consists of the ten independent variables that describe various physical phenomena associated

with the bedded salt formation. The probability dis- tributions and associated ranges are given in Ref. 2.

5.2. Selection of the Values of the Input Variables Used in the Analysis

The Response Surfacs Replacement. For the re- sponse surface analysis, a 1/32 replication of a 21° factorial design was utilized with the input variables. Thus, 32 computer runs are required. This is the smallest number of runs that could be used and still allow for an estimate of each main effect and 15 interactions selected a priori by the model developers as potentially important.

The Latin Hypercube Sample. For the portion of the analysis utilizing a Latin hypercube sample, the value of n = 32 was selected to keep the sample size the same as used with the fractional factorial design. The sample was generated using the restricted pair- ing technique of Iman and Conover('*) to control the correlations between variables within the sample. Examination of the rank correlation matrix for the LHS indicated that 35 of the 45 pairwise entries were < 0.05 in absolute value, 43 of 45 were < 0.10, and the largest element was 0.138. The associated rank correlation matrix for the LHS had a variance infla- tion factor of 1.03.

Page 14: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

84 Iman and Helton

7

5

3

* r

z - 1

0

2 -3

- 5

7

5

3

> I

z Q -1 2

-3

LOGlO X 2 L N X 3

- 5 I LOGlO X 2

..

. . . . . .

‘ 4 4

. 4 4

. 8 8

LN X 8

. .

. .

. . . . . .

LN X 3 L N X 8

Fig. 8. Scatterplots of log X , , In X,, and In X, vs. log Y at 5 yr.

5.3. Scatterplots of the Input-Output Relationships as a Guide to Better Understanding of the Model Behavior

Scatterplots can be a great aid in determining if the model is working as intended, i.e., does the input-output agree with engineering judgment? Ad- ditionally, scatterplots may aid in identifying the need for transformations (such as logarithmic), or when placed side-by-side, may show how several variables jointly influence the output. The DNET

8

4

0

4

8

4

0

4

8

I

LOGlO X 2

I . . . . . . . . . . . .

LOGlO X 2

1

I I

LN X 3

... .... * . ”.

LN X 3

I

L N X 8

. . :. . . . . . . . ’ . . . .

* .

L N X 8

Fig. 9. Scatterplots of log X , , In X , , and In X , vs. log Y at 5000 yr.

model provided a good illustration of the value of scatterplots, as some unexpected model behavior was detected through their use. Side-by-side scatterplots for DNET output appear in Figs. 8 and 9 for X,, X , , and x8 at time steps of 5 and 5000 yr, respectively. The left-hand one-third of each of these graphs shows a scatterplot of log X , versus log r; the middle one- third has In X, versus log r; the right-hand one-third shows In & versus logy. The top half of each figure was based on fractional factorial input, while the bottom half was based on input selected by Latin

Page 15: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

Uncertainty and Sensitivity Analysis Techniques 85

hypercube sampling. With fractional factorial input, each variable takes on only two values; thus, the plot appears as 16 points (half of the 32 runs) above each of the two values for each variable. In the top por- tion of Figure 8, the numerals appearing with the graphs indicate the multiplicity of the various points.

Examination of the top half of Fig. 8 shows that X , is dominant in controlling the value of Y in conjunction with X,. When the low value of X , is present, the output was constant at approximately

regardless of X,. When the high value of X , was used, the output was either lo5 or lo7 depending on whether the high value of X , was paired with the low value of X , or the high value of X,, respectively. The side-by-side scatterplots of X , and X , make this relationship easy to see. Additionally, x8 has no effect at either high or low values.

The plot of X , versus Y in the bottom half of Fig. 8 (5 yr) shows an interesting phenomenon. The points appear in two distinct clusters separated by several orders of magnitude on the Y axis. These clusters are determined by large and small values of X,, which represents the conductivity of a shale layer directly above the bedded salt formation and directly below an aquifer. In a simplified sense, the higher the conductivity of the shale, the greater the water flow through the shale and in turn the greater the dissolu- tion of the salt cavity. Thus, the six points in the cluster at the top of the graph represent “break- through” runs. That is, conditions exist on these runs that allow the salt cavity to undergo rapid dissolution and thereby create a discontinuity in the output. The LHS plot of X , versus Y in Fig. 8 shows that when a “breakthrough” occurs, the rate of dissolution is controlled by X,, the conductivity of the aquifer directly above the shale layer. Further, Fig. 8 indi- cates that x8 has little or no effect on dissolution regardless of whether or not a breakthrough occurs.

Figure 9 (5000 yr) also shows some interesting results. The variable X , is still dominent in determin- ing if a breakthrough occurs. When a breakthrough occurs, X , is the dominant variable, while X , is dominant when a breakthrough does not occur.

The scatterplots in Figs. 8 and 9 make it readily apparent that the occurrence or nonoccurrence of a breakthrough has a direct bearing on any regression based analysis whose purpose is to determine the relative importance and contribution of the input variables. For example, a straight line or even a quadratic fit to the points in the LHS plot of X , versus Y in Fig. 8 will be less than satisfactory. This

is true because while the lower cluster (no break- through) of the graph can be fitted nicely by a quadratic in X,, the upper cluster (breakthrough) shows no relationship to X,. Rather, the behavior of the upper portion seems to be dominated by X,.

While the results of the two sets of scatterplots corresponding to the two approaches give good gen- eral agreement, it is impossible for plots based on two fixed levels to show the true pattern of the input-output relationship for this model. That is, one must assume that the output behaves in a linear fashion between the high and low values of X,, with the exact placement of the linear relationship being dependent upon the value of X,. The discontinuity present in the LHS portion of Fig. 8 cannot be discovered in the fractional factorial portion of the figure, nor can the nonlinear relationship between X , and Y for about 80% of the range of X,. This is a price that must be paid for the simplifying assump- tions that go along with the use of two levels for each variable. At this point, one might consider a modifi- cation of the fractional factorial design to include points other than those found at the two levels. For instance, a central composite design would use 2k + 1 additional points, with one point being in the center for each variable and 2k axial points. For the present analysis, an additional 21 computer runs would be needed. However, there would still be no guarantee that the information available in the lower half of Figs. 8 and 9 would become apparent, whereas the LHS approach would have revealed the same infor- mation with fewer than 32 runs. At the same time, it should be kept in mind that the output variable shown in Figs. 8, 9, and 10 is only one of six that were included in the complete analysis.

5.4. Uncertainty Analysis for the DNET Model

Response surfaces for the fractional factorial design were fitted at each of the time steps and then evaluated 100 times based on random sampling of the inputs. The resulting estimated cdfs for 5 yr and 500 yr appear in Fig. 10 with the label “RS.” For ease in making comparisons, the LHS results also appear (with the label “LHS’). There are actually two cdf estimates based on LHS appearing within both portions of Fig. 10. One is based on 32 runs and the other is based on 100 runs. The close agreement of the LHS estimates within each figure provides an indication of the precision associated with estimates arising from LHS.

Page 16: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

86 Iman and Helton

1

z + z 3

z I-

0 .e

0 .8

.r

0 .(I

= .s z a s n

t z .I

I- .4

.3

.2

c b w o

- 8 -3 -1 I 3 5 7

LOG10 OF SOLUTION RATE

z 0 c

0 w

LOQlO OF SOLUTION RATE

Fig. 10. Response surface and LHS estimates of the CDF for log Y.

Figure 10 shows that the LHS estimates of the output cdf disagree sharply with the response surface estimate obtained from using the fractional factorial input. In fact, most quantile estimates disagree by one or more orders of magnitude. The reason for this disagreement is that the phenomenon of a “break- through” and the corresponding discontinuity in the output mentioned earlier in the discussion of the scatterplots has gone undetected by the analysis of the input-output relationships based on the frac- tional factorial approach. That is, from the top half of Figs. 8 and 9 it can be seen that the low and high

values of the output differ by several orders of mag- nitude over the range of X,. But what of the behav- ior of the output for intermediate values of X,? The bottom half of Figs. 8 and 9 as well as the LHS portion of Fig. 10 make it clear that a discontinuity occurs in the output. This discontinuity is missed by the fractional factorial approach since a linear rela- tionship must be assumed between the two levels used with each input variable. More complex re- sponse surfaces could be estimated, but only at in- creased computer costs, and even then it is likely that such surfaces would fail to accurately depict the discontinuity. This example identifies an underlying problem with the response surface approach, whch is that the DNET computer model is too complex mathematically to be adequately represented by a simple response surface.

6. SUMMARY AND CONCLUSIONS

6.1. Techniques and Models Used in Comparisons

Uncertainty analysis and sensitivity analysis are important elements in the development and imple- mentation of computer models for complex processes. Many different techniques have been proposed for performing uncertainty and sensitivity analyses, and published comparisons have been made of some techniques. It is our observation that such compari- sons are frequently made on unrealistic and artifi- cially simple models. For example, one might see a small number of independent input variables used with a simple function. This approach has merit in allowing comparisons against known answers but fails to show the extendibility of the techniques to complex problems. Thus, a main objective of this study was to investigate three widely used techniques on three models having large uncertainties and vary- ing degrees of complexity, in order to highlight some of the problem areas that must be addressed in actual applications.

We are aware that results presented in a study such as this will not satisfy everyone. For example, questions will arise as to the choice of techniques used in the comparisons. We feel that there is ade- quate justification for the techniques used in the comparisons, but more importantly, given the avail- able documentation on the input and models, it should be possible for other investigators to make comparisons of additional techniques with the results in this study.

Page 17: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

Uncertainty and Sensitivity Analysis Techniques 87

6.2. General Summary of the Techniques

A brief summary of each technique is given in this subsection. A more extensive summary appears in Ref. 2.

Response Surface Replacement Using Fractional Factorial Designs. Fractional factorial designs (FFD) have a proven record of performance and are well documented in the statistical literature. They have provided good results in many experimental situa- tions. They, along with more complex experimental designs, have also been widely used in the construc- tion of response surfaces and are near an optimal choice for input selection if the output behaves in a linear fashion. Even if the output behaves in a non- linear fashion, the FFD can sometimes be modified to give reasonable results by including center points. The problem in using a FFD to produce a response surface replacement for a computer model of the type considered in this paper lies not so much in the choice of the design but rather in the concept of trying to replace the model with a response surface. Generally spealung, the models are too complex mathematically to be adequately approximated with a response surface. Since indirect estimates of the output cdf(s), the variance of the output variable(s), and variable ranking by contribution to variance are derived from the response surface when using input from a FFD, an inadequate response surface can generate misleading uncertainty analysis and sensitiv- ity analysis results.

The presence of multivariate output creates problems when using a FFD in a sequential manner. Generally speaking, in a sequential approach, the point pairing process is altered at each step in the sequence on the basis of the effect on a single re- sponse. However, when a model such as Pathways shows 19 of the 20 inputs as being important in influencing at least one of the 12 outputs,(22) it becomes difficult to proceed in a sequential manner with a FFD.

The FFD response surface approach provided both good and poor estimates of the output cdfs for the simple Pathways model, but did provide a reli- able ranking of the input variables. However, in evaluating rankings for the Pathways model, it should be kept in mind that one input variable tended to dominate each output variable. The response surface approach was not used with the MAEROS model owing to the complexity of the multivariate input structure. The response surface approach provided

an unreliable estimate of the output cdf with the DNET model largely because of a discontinuity in the output with respect to one of the inputs.

Differential Analysis and Local Behavior. A dif- ferential analysis is intended to provide information with respect to small perturbations about a point. Excellent information is provided about variable be- havior and influence about this point, and such in- formation is quite useful in a variety of applications. Problems arise, however, in an uncertainty analysis or in a sensitivity analysis when large uncertainties are present and attempts are made to extend the results from the small perturbations in the input variables, for which the differential analysis is in- tended, to a broader or global interpretation. For example, estimates of the cdf and variance can be obtained indirectly by using a Monte Carlo simula- tion with the Taylor series expansion about a base- case point. The results may or may not be sensitive to the choice of this base-case point.

The differential analysis was relatively easy to implement with the Pathways model, and a Monte Carlo simulation of the Taylor series expansion about the base-case point produced a cdf in reasonably good agreement with the unbiased estimate as shown in Fig. 2. However, as illustrated in Fig. 3, other choices of base-case values gave widely varying re- sults. The input variable ranking proved to be in good agreement with the other techniques, but as was mentioned previously, one input variable tended to dominate each output variable. The implementation of the differential analysis with the MAEROS model proved to be another matter entirely. It was ex- tremely difficult and time consuming to do, requiring about six man-months of effort. Moreover, even though the cdf estimates shown in Figs. 4 and 5 all utilized a Taylor series expansion about the same base-case point, the results ranged from reasonably good to poor for two different output variables. The variable rankings shown in Table I11 display quite a bit of disparity among the rankings. Due to the complexity of the feedback mechanism governing the selection of different submodels, a differential analy- sis was not performed with the DNET model.

Latin Hypercube Sampling. The implementation of LHS is similar to that of simple random sampling in that both have a probabilistic basis. In fact, for large sample sizes there is little difference between the two techniques. However, the original intent of LHS was to make more efficient use of computer runs than random sampling for smaller sample sizes.

Page 18: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

88 Iman and Helton

If the output is a monotone function of the input (as has often been our experience with computer models), then LHS has been shown by Iman and Conover(”) to be more efficient than simple random sampling (i.e., LHS provides a smaller variance for the estima- tor). If monotonicity is not satisfied, then LHS may or may not be more efficient than simple random sampling. Since LHS has a probabilistic basis, it can provide direct estimates for the cdf and variance. When random pairing is used with LHS, the estimate of the cdf is unbiased while the variance estimate has a small, bounded, but unknown bias. When LHS is used in conjunction with the pairing technique of Iman and Conover,(’*) correlated multivariate struc- tures for the input variables can be input to the computer model in the proper form, something that cannot be done with random pairing in the LHS. However, the property of unbiasedness no longer applies. It is felt that the amount of bias is negligible for the type of problems considered in this study.(s6)

LHS was used with all three computer models and produced good estimates of the cdf throughout as measured by comparisons with results from large random samples. For example, for both the Pathways model and the MAEROS model, the LHS estimates with n=50 showed good agreement with random sample estimates with n =loo. In the case of the DNET model, the LHS estimate with n = 32 was compared against a LHS with n =lo0 in order to illustrate the small variability associated with LHS estimates. The presence of a discontinuity in the output from DNET was clearly identifiable from the LHS cdf estimate, illustrating the usefulness of LHS in mapping the input space to the output space.

6.3. Concluding Remarks

On the basis of the results presented in this paper and several years of experience working in this area, we recommend use of the LHS approach. We feel that it is easy to use, has a great deal of flexibil- ity, is applicable to many different modeling situa- tions, and provides reliable results.

The best use of experimental designs in the type of analysis considered in this study is in a screening role. However, the use of response surface replace- ments for the computer models is not recommended as the underlying models are often too complex to be adequately represented by a simple function.

A differential analysis provides good local infor- mation about the inputs, but does not extend well to

a global interpretation. A very real problem with a differential analysis lies in the difficulty of imple- mentation.

The role of probability distributions to describe the input is worth noting. In a screening analysis, the exact form of the distribution is of less importance than is the use of a physically reasonable range and an input selection procedure that thoroughly explores this range. However, if a meaningful cdf of the output variables is desired, then care must be taken to correctly describe the input distributions as well as their underlying correlation structure.

Many models considered in practice are far more complex than the three models used in this compari- son. For example, the MELCOR computer program indicated in the introduction is such a model (ap- proximately 150,000 lines of coding). The analysis of a complex model must be carefully planned. A strategy for performing the analysis must be devel- oped that takes into account the scale, structure, and constraints associated with the particular problem under consideration. Among other things, such a strategy must include mechanisms for propagating uncertainties between different submodels. Clearly, an understanding of the individual submodels gained through analyses of the type indicated in this paper will facilitate the development of an overall analysis strategy. Examples of uncertainty/sensitivity analyses performed for complex systems using LHS are given in Refs. 21 and 57. Unfortunately, the orchestration of an analysis for a complex system depends on many properties unique to the particular problem under consideration; this makes it difficult to pro- vide a general prescription for performing such an analysis.

ACKNOWLEDGMENTS

This work was supported by the U.S. Nuclear Regulatory Commission and performed at Sandia National Laboratories which is operated for the U.S. Department of Energy under contract No. DE- AC04-76DP00789.

REFERENCES

1. J. L. Sprung, D. C. Aldrich, D. J. Alpert, and G. G. Weigand, “Overview of the MELCOR Risk Code Development Pro- gram,” Proceedings of the International Meeting on Light Wuter Reactor Severe Accident Evaluation, Cambridge, Massachusetts (1983). pp. TS-10.1-1 to TS-10.1-8.

Page 19: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

Uncertainty and Sensitivity Analysis Techniques 89

2. R. L. Iman and J. C. Helton, “A Comparison of Uncertainty and Sensitivity Analysis Techniques for Computer Models,” NUREG/CR-3904, SAND84-1461, Sandia National Labora- tories, Albuquerque, New Mexico (1985).

3. D. J. Downing, R. H. Gardner, and F. 0. Hoffman, “Re- sponse Surface Methodologies for Uncertainty Analysis in Assessment Models,” Technometrics 27(2), 151-163 (1985).

4. T. A. Bishop, “The Use of Sensitivity Analysis in Experimen- tal Planning and Design,” Paper presented at the 1983 Annual Meeting of the American Statistical Association, Toronto, Canada.

5. G. P. Steck, D. A. Dahlgren, and R. G. Easterling, “Statistical Analysis of LOCA, FY75 Report,” Technical Report SAND75-0653, Sandia National Laboratories, Albuquerque, New Mexico (1975).

6. N. D. Cox, “Comparison of Two Uncertainty Analysis Meth- ods,” Nuclear Science and Engineering 64,258-265 (1977).

7. R. K. Steinhorst, H. W. Hunt, G. S. Innis, and K. P. Haydock, “Sensitivity Analysis of the ELM Model,” Grassland Simula- tion Model, G. S . Innis (ed.) (Springer-Verlag, New York,

8. J. P. C. Kleijnen, “Regression Metamodels for Generalizing Simulation Results,” IEEE Transactions on Systems, Man, and Cybernetics, SMC-9, 93-96 (1979).

9. D. H. Nguyen, “The Uncertainty in Accident Consequences Calculated by Large Codes Due to Uncertainties in Input,” Nuclear Technology 49, 80-91 (1980).

10. P. Baybutt, D. C. Cox, and R. E. Kurth, “Method for Uncer- tainty Analysis of Large Computer Codes, with Application to Light-Water Reactor Meltdown Accident Consequence Evaluation,” Informal Report, Battelle Columbus Laborato- ries, Columbus, Ohio (1982).

11. M. D. McKay, W. J. Conover, and R. J. Beckman, “A Comparison of Three Methods for Selecting Values of Input Variables in the Analysis of Output from a Computer Code,” Technometrics 21(2), 239-245 (1979).

12. R. L. Iman and W. J. Conover, “A Distribution-Free Ap- proach to Inducing Rank Correlation Among Input Variables,” Communications in Statistics B11(3), 311-334 (1982).

13. R. L. Iman and M. J. Shortencarier, “A FORTRAN 77 Program and User’s Guide for the Generation of Latin Hyper- cube and Random Samples for Use with Computer Models,” NUREG/CR-3624, SAND83-2365, Sandia National Labora- tories, Albuquerque, New Mexico (1984).

14. R. L. Iman, “A Matrix-Based Approach to Uncertainty and Sensitivity Analysis for Fault Trees,” Risk Analysis 7, 21-33 (1987).

15. R. L. Iman and W. J. Conover, “Small Sample Sensitivity Analysis Techniques for Computer Models, with an Applica- tion to Risk Assessment,” Communications in Statistics A9(17), 1749-1842. “Rejoinder to Comments,” 1863-1874 (1980).

16. R. L. Iman and W. J. Conover, “Sensitivity Analysis Tech- niques: Self-Teaching Curriculum,” NUREG/CR-2350, SAND81-1978, Sandia National Laboratories, Albuquerque, New Mexico (1982).

17. R. L. Iman, J. C. Helton, and J. E. Campbell, “An Approach to Sensitivity Analysis of Computer Models, Part 1. Introduc- tion, Input Variable Selection and Preliminary Variable As- sessment,” Journal of Qualify Technology 13(3), 174-183 (1981).

18. R. L. Iman, J. C. Helton, and J. E. Campbell, “An Approach to Sensitivity Analysis of Computer Models, Part 2. Ranking of Input Variables, Response Surface Validation, Distribution Effect and Technique Synopsis,” Journal of Qualify Technology

19. W. V. Harper and S. K. Gupta, “Sensitivity/Uncertainty Analysis of a Borehole Scenario Comparing Latin Hypercube Sampling and Deterministic Sensitivity Approaches,” BMI/ONWI-516, Battelle Memorial Institute, Columbus, Ohio (1983).

1978), pp. 231-255.

13(4), 232-240 (1981).

20. J. C. Helton and R. L. Iman, “Sensitivity Analysis of a Model for the Environmental Movement of Radionuclides,” Health Physics 42(5), 565-584 (1982).

21. R. M. Cranwell, J. E. Campbell, J. C. Helton, R. L. Iman, D. E. Longsine, N. R. Ortiz, G. E. Runkle, and M. J. Shortencarier, “Risk Methodology for Geologic Disposal of Radioactive Waste: Final Report,” NUREG/CR-2452, SAND81-2573, Sandia National Laboratories, Albuquerque, New Mexico (1987).

22. J. C. Helton, R. L. Iman, and J. B. Brown, “Sensitivity Analysis of the Asymptotic Behavior of a Model for the Environmental Movement of Radionuclides,” Ecological Mod- elling 28, 243-278 (1985).

23. R. J. Beckman and D. N. Whiteman, “Uncertainty Analysis: Good News and Bad News,” Proceedings of the Ninth Annual Statistics Symposium on National Energy Issues, October, Rockville, Maryland 103-112 (1983).

24. R. Tomovic, Sensitiuify Analysis of Dynamic Systems (Mc- Graw-Hill, New York, 1963).

25. R. Tomovic and M. Vukobratovic, General Sensitivity Theory (Elsevier, New York, 1972).

26. P. M. Frank, Introduction to System Sensitivity Theory (Academic Press, New York, 1978).

27. S. Morisawa and Y. Inoue, “On the Selection of a Ground Disposal Site by Sensitivity Analysis,” Health Physics 26,

28. R. W. Atherton, R. B. Schainker, and E. R. Ducot, “On the Statistical Analysis of Models for Chemical Kinetics,” AICHE Journal 21, 441-448 (1975).

29. R. P. Dickinson and R. J. Gelinas, “Sensitivity Analysis of Ordinary Differential Equation Systems-A Direct Method,” Journal of Computational Physics 21, 123-143 (1976).

30. K. W. Lee, J. A. Gieseke, and L. D. Reed, “Sensitivity Analysis of the HAARM-3 Code,” NUREG/CR-0527, BMI- 2008, Battelle Columbus Laboratories, Columbus, Ohio (1978).

31. M. E. Cunningham, C. R. Ham, and A. R. Olsen, “Uncer- tainty Analysis and Thermal Stored Energy Calculations in Nuclear Fuel Rods,’’ Nuclear Technology 47, 457-467 (1980).

32. A. M. Dunker, “Efficient Calculation of Sensitivity Coeffi- cients for Complex Atmospheric Models,” A tmospheric En- vironment 15, 1155-1161 (1981).

33. M. Koda, “Sensitivity Analysis of the Atmospheric Diffusion Equation,” Atmospheric Environment 16, 2595-2601 (1982).

34. J. Barhen, D. G. Cacuci, J. J. Wagichal, M. A. Bjerke, and C. B. Mullins, “Uncertainty Analysis of Time-Dependent Nonlinear Systems: Theory and Application to Transient Thermal Hydraulics,” Nuclear Science and Engineering 81,

35. J. C. Helton and P. C. Kaestner, “Risk Methodology for Geologic Disposal of Radioactive Waste: Model Description and User Manual for Pathways Model,” NUREG/CR-1636, Vol. 1, SAND78-1711, Sandia National Laboratories, Al- buquerque, New Mexico (1981).

36. J. C. Helton and R. L. Iman, “Risk Methodology for Geologic Disposal of Radioactive Waste: Sensitivity Analysis of the Environmental Transport Model,’’ NUREG/CR-1636, Vol. 2, SAND79-1393, Sandia National Laboratories, Albuquerque, New Mexico (1980).

37. J. C. Helton, J. B. Brown, and R. L. Iman, “Risk Methodol- ogy for Geologic Disposal of Radioactive Waste: Asymptotic Properties of the Environmental Transport Model,” NUREG/CR-1636, Vol. 3, SAND79-1908, Sandia National Laboratories, Albuquerque, New Mexico (1981).

38. J. B. Brown and J. C. Helton, “Risk Methodology for Geo- logic Disposal of Radioactive Waste: Effects of Variable Hy- drologic Patterns on the Environmental Transport Model,” NUREG/CR-1636, Vol. 4, SAND79-1909, Sandia National Laboratories, Albuquerque, New Mexico (1981).

39. D. W. Marquardt and R. D. Snee, “Ridge Regression in Practice,” The American Statistician 29(1), 3-20 (1975).

251-261 (1974).

23-44 (1982).

Page 20: An Investigation of Uncertainty and Sensitivity Analysis …mycourses.ntua.gr/.../WRM_Part_B_Makropoulos/iman1988.pdf · 2017-12-16 · Risk Analysis, Vol. 8, No. 1, 1988 An Investigation

Iman and Helton

40. N. R. Draper and K Smith, Applied Regressiou Analysis, Second Edition (Wiley, New York, 1981).

41. F. Gelbard, “MAEROS User Manual,” NUREG/CR-1391, SANDBO-0822, Sandia National Laboratories, Albuquerque, New Mexico (1982).

42. F. Gelbard and J. H. Seinfeld, “Simulation of Mttlticompo- Rent Aerosol Dynamics,” JWMI of Cdluid m d Interfuce

43. J. C. Helton, R. L. Iman, J. D. Johnson, and C. D. Leigh, “Uncertainty and Sensitivity Analysis of a Model for Multi- component Aerosol Dynamics;” Nuclear Technobgy 73(2),

44. M. Koda, A. H. Dogru, and J. H. Seinfeld, “Sensitivity Analysis of Partial Differential Equations with Apphcation to Reaction and Diffusion Processes,” J w n d o/ Computational Physics 38, 259-282 (1979).

45. D. G. Cxuci, C. F. Weber, E. M. W o w , and J. H. Marable, “Sensitivity Theory for Genera4 Systems of Nonlinear Equa- tions,“ Nrleur Science and Engineering 75, 88-110 (1980).

06. D. G. C a c r i , “Sensitivity Theory for Nodinear Systems. I. Nonlinear Functional Analysis Approach,” Journal of Matkematical Physics 22, 2794-2802 (1981).

47. D. G. Cacuci, “Sensitivity Theory for Nonlinear Systems. 11. Extensions to Additional Classes of Responses,” Journal of Mathematical Physics 22, 2803-2812 (1981).

48. D. G. Cacuci, P. J. Maudlin, and C. U. Parics, “Adjoint Sensitivity Analysis of Extremum-Type Responses in Reactor safety,” Nuclear Science and Emginecring, 83, 112-135 (1983).

49. M. G. Piepho, J. B. Cady, and M. A. Kenton, “An Importance Theory for Nonlinear O r d i n q Differential Equations,” Re- port No. HEDL-TIM 81-24, Hanford Engineering Develop- ment Laboratory, Richland, Washington (1981).

50. M. C. G. Hall, D. G. Cacuci, and M. E. Schlesigwr, “Sensitiv- iky Analysis of a Radiative-Convective Model by the Adjoint

S ~ i e n ~ e 78,485-501 (1980).

320-342 (1986).

Method,” Journal of the Atmospheric Sciences 39. 2038-2050 (1982).

51. J.-T. Hwang, E. P. Dougherty, S. Rabitz, and H. Rabitz, “The Green’s Function Method of Sensitivity Analysis in Chemical Kinetics,” Journal of Chemical Physics 69, 5180-5191 (1978).

52. E. P. Dougherty, J.-T. Hwang, and H. Rabitz, “Further Devel- opments and Applications of the Green’s Function Method of Sensitivity Analysis in Chemical Kinetics,” Journal of Chem-

53. D. K. Dacol and H. Rabitz, ‘‘Arbitrary Order Functional Sensitivity Densities for Reaction-Diffusion Systems.” Journul of Chemical Physics 78, 4905-4914 (1983).

54. R. L. Iman, M. J. Shortencarier, and J. D. Johnson, “A FORTRAN 77 Program and User’s Guide for the Calculation of Partial Correlation and Standardized Regression Coeffi- cients,” NUREG/CR-4122, SAND85-0044, Sandia National Laboratories, Albuquerque, New Mexico (1985).

55. R. M. Cranwell, J. E. Campbell, and S. E. Stuckwisch, “ h s k Methodology for Geologic Disposal of Radioactive Waste: The DNET Computer Code User’s Manual,” NUREG/CR- 2343, SAND81-1663. Sandia National Laboratories. Al- buquerque, New Mexico (1982).

56. W. R. Schucany, A. Thombs, and K. Cunningham, “Gener- ating Jointly Distributed Variates by Restricted Random Sam- pling,” Technical Report SMU-D5-TR-197, Southern Method- ist University, Dallas, Texas (1986).

57. C. N. Amos, A. S. Benjamin, G. J. Boyd, J. M. Griesmeyer. F. E. Haskin, J. C. Helton, D. M. Kunsman, S. R. Lewis, and L. N. Smith, “Evaluation of Severe Accident Risks and the Potential for Risk Reduction: Peach Bottom, Unit 2 (Draft for Comment),” NUREG/CR-4551, SAND86-1309, Vol. 3, Sandia National Laboratories, Albuquerque, New Mexico (1987).

ical physic^ 71, 1794-1808 (1979).