71
EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A thesis submitted to The Johns Hopkins University in conformity with the requirements for the degree of Master of Science in Engineering. Baltimore, Maryland May, 2007 c Sonjoy Das 2007 All rights reserved

AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

EFFICIENT CALCULATION OF FISHER

INFORMATION MATRIX: MONTE CARLO APPROACH

USING PRIOR INFORMATION

by

Sonjoy Das

A thesis submitted to The Johns Hopkins University in conformity with the

requirements for the degree of Master of Science in Engineering.

Baltimore, Maryland

May, 2007

c© Sonjoy Das 2007

All rights reserved

Page 2: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Abstract

The Fisher information matrix (FIM) is a critical quantity in several aspects of

mathematical modeling, including input selection and confidence region calculation.

Analytical determination of the FIM in a general setting, specially in nonlinear mod-

els, may be difficult or almost impossible due to intractable modeling requirements

and/or intractable high-dimensional integration.

To circumvent these difficulties, a Monte Carlo (MC) simulation-based technique,

resampling algorithm, based on the values of log-likelihood function or its exact

stochastic gradient computed by using a set of pseudo data vectors, is usually rec-

ommended. The current work proposes an extension of the resampling algorithm in

order to enhance the statistical qualities of the estimator of the FIM. This modified

resampling algorithm is useful in those cases where the FIM has a structure with

some elements being analytically known from prior information and the others being

unknown. The estimator of the FIM, obtained by using the proposed algorithm, si-

multaneously preserves the analytically known elements and reduces variances of the

estimators of the unknown elements by capitalizing on information contained in the

ii

Page 3: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

known elements. Numerical illustrations considered in this work show considerable

improvement of the new FIM estimator (in the sense of mean-squared error reduction

as well as variance reduction) based on the modified MC based resampling algorithm

over the FIM estimator based on the current MC based resampling algorithm.

Advisor: Prof James C. Spall. The Johns Hopkins University.

iii

Page 4: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Acknowledgements

I would like to express my sincere gratitude and appreciation to my advisor,

Professor James C. Spall, for his insightful suggestions, technical guidance, patience,

support and encouragement throughout the course of this work. I also like to formally

express my sincere thanks and appreciation to the advisor of my doctoral program,

Professor Roger Ghanem, at the Department of Civil and Environmental Engineer-

ing, University of Southern California, for his constant support and allowing me to

continue this work within the already busy schedule of my doctoral work, without

which this work would definitely have not been completed. In addition, I thank Pro-

fessor Daniel Q. Naiman for his support to my endeavor for the successful completion

of this work.

I am thankful to Ms. Kristin Bechtel, Academic Program Coordinator of the

department, for her constant help in numerous department and university formalities

essential for completion of this work.

iv

Page 5: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Contents

Abstract ii

Acknowledgements iv

List of Tables vii

List of Figures viii

1 A Glimpse of Fisher Information Matrix 11.1 Motivating Factors . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21.2 Fisher Information Matrix: Definition and Notation . . . . . . . . . . 41.3 Estimation of Fisher Information Matrix . . . . . . . . . . . . . . . . 5

1.3.1 Current Re-sampling Algorithm: No use of Prior Information . 6

2 Improved Re-sampling Algorithm 142.1 Additional Notation . . . . . . . . . . . . . . . . . . . . . . . . . . . 152.2 Modified Re-sampling Algorithm using Prior Information . . . . . . . 162.3 Theoretical Basis for the Modified Re-sampling Algorithm . . . . . . 20

2.3.1 Case 1: only the measurements of L are available . . . . . . . 202.3.2 Case 2: measurements of g are available . . . . . . . . . . . . 37

3 Numerical Illustrations and Discussions 463.1 Example 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 463.2 Example 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 483.3 Discussions on results . . . . . . . . . . . . . . . . . . . . . . . . . . . 52

4 Conclusions and Future Research 574.1 Summary of Contributions Made . . . . . . . . . . . . . . . . . . . . 584.2 Suggestions for Future Research . . . . . . . . . . . . . . . . . . . . . 58

Bibliography 60

v

Page 6: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Vita 63

vi

Page 7: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

List of Tables

3.1 Example 1 – MSE and MSE reduction of FIM estimates based on N =2000 pseudo data (results on variance are reported within parentheses). 54

3.2 Example 1 – MSE comparison for Fn and Fn for only the unknownelements of Fn(θ) according to F

(given)n based on N = 100000 pseudo

data (similar results on variance are reported within parentheses, A =∑pi=1

∑j∈Ici var[Fij|θ] and B =

∑pi=1

∑j∈Ici var[Fij|θ]). . . . . . . . . 55

3.3 Example 2 – MSE and MSE reduction of FIM estimates based onN = 40000 pseudo data (results on variance are reported within paren-theses). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55

3.4 Example 2 – MSE comparison for Fn and Fn for only the unknownelements of Fn(θ) according to F

(given)n based on N = 40000 pseudo

data (similar results on variance are reported within parentheses A =∑pi=1

∑j∈Ici var[Fij|θ] and B =

∑pi=1

∑j∈Ici var[Fij|θ]). . . . . . . . . 55

3.5 Example 1 – Variance comparison between theoretical results (correctup to a bias term as discussed in Section 2.3) and simulation results(with N = 2000). . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56

vii

Page 8: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

List of Figures

1.1 Fisher information matrix with known elements as marked; void partconsists of unknown elements. . . . . . . . . . . . . . . . . . . . . . . 3

1.2 Schematic model for forming FIM estimate, FM,N ; adapted from [13]. 10

2.1 Schematic of algorithm for computing the new FIM estimate, Fn. . . 19

3.1 F(given)30 for Example 2 with known elements as marked; void part con-

sists of unknown elements to be estimated by using MC based resam-pling algorithms. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50

viii

Page 9: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Chapter 1

A Glimpse of Fisher Information

Matrix

The Fisher information matrix (FIM) plays a key role in estimation and identifica-

tion [12, Section 13.3] and information theory [3, Section 17.7]. A standard problem

in the practical application and theory of statistical estimation and identification is

to estimate the unobservable parameters, θ, of the probability distribution function

from a set of observed data set, {Z1, · · · ,Zn}, drawn from that distribution [4]. The

FIM is an indicator of the amount of information contained in this observed data set

about the unobservable parameter, θ. If θ is an unbiased estimator of θ∗ represent-

ing the true value of θ, then the inverse of the FIM computed at θ∗ provides the

Cramer-Rao lower bound of the covariance matrix of θ, implying that the difference

between cov(θ) and the inverse of the FIM computed at θ∗ is a positive semi-definite

1

Page 10: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

matrix. Therefore, the name information matrix is used to indicate that a larger

FIM (in the matrix sense – positive semi-definiteness as just mentioned) is associated

with a smaller covariance matrix (i.e., more information), while a smaller FIM is

associated with a larger covariance matrix (i.e., less information). Some important

areas of applications of FIM include, to name a few, confidence interval computation

of model parameter [2, 7, 16], configuration of experimental design involving linear [9]

and nonlinear [12, Section 17.4] models, and determination of noninformative prior

distribution (Jeffreys’ prior) for Bayesian analysis [8].

1.1 Motivating Factors

However, analytical determination of the FIM in a general setting, specially in

nonlinear models [16], may be a formidable undertaking due to intractable modeling

requirements and/or intractable high-dimensional integration. To avoid this analyti-

cal problem, a computational technique based on Monte Carlo simulation technique,

called resampling approach [12, Section 13.3.5], [13], may be employed to estimate

the FIM. In practice, there may be instances when some elements of the FIM are an-

alytically known from prior information while the other elements are unknown (and

need to be estimated). In a recent work [4, 5], a FIM of size 22× 22 was observed to

have the structure as shown in Figure 1.1.

In such cases, the above resampling approach, however, still yields the full FIM

without taking any advantage of the information contained in the analytically known

2

Page 11: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

1 4 8 12 16 20 22

1

4

8

12

16

20

22

Known elements = 54

Figure 1.1: Fisher information matrix with known elements as marked; void partconsists of unknown elements.

elements. The prior information related to known elements of the FIM is not incor-

porated while employing this algorithm for estimation of the unknown elements. The

resampling-based estimates of the known elements are also “wasted” because these

estimates are simply replaced by the analytically known elements. The issue yet to

be examined is whether there is a way of focusing the averaging process (required

in the resampling algorithm) — on the elements of interest (unknown elements that

need to be estimated) — that is more effective than simply extracting the estimates

of those elements from the full FIM estimated by employing the existing resampling

algorithm.

The current work presents a modified and improved (in the sense of variance re-

duction) version of the resampling approach for estimating the unknown elements

of the FIM by “borrowing” the information contained in the analytically known el-

ements. The final outcome exactly preserves the known elements of the FIM and

3

Page 12: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

produces variance reduction for the estimators of the unknown elements that need to

be estimated.

The main idea of the current work is presented in chapter 2.2 and the relevant

theoretical basis is highlighted in section 2.3. The proposed resampling algorithm,

that is similar in some sense to the one for Jacobian/Hessian estimates presented

earlier [14], modifies and improves the current resampling algorithm by simultaneously

preserving the known elements of the FIM and yielding better (in the sense of variance

reduction) estimates of the unknown elements. The effectiveness and efficiencies of the

new resampling algorithm is illustrated by numerical examples in chapter 3. Chapter 4

infers the conclusions from the current work.

1.2 Fisher Information Matrix: Definition and No-

tation

Consider a set of n random data vector {Z1, · · · ,Zn} and stack them in Zn,

i.e., Zn = [ZT1 , · · · ,ZT

n ]T . Here, the superscript T is transpose operator. Let the

multivariate joint probability density or mass (or hybrid density/mass) function (pdf)

of Zn be denoted by pZn(·|θ) that is parameterized by a vector of parameters, θ =

[θ1, · · · , θp]T ∈ Θ ⊆ Rp in which Θ is the set consisting of allowable values for θ.

The likelihood function of θ is then given by `(θ|Zn) = pZn(Zn|θ) and the associated

log-likelihood function L by L(θ|Zn) ≡ ln `(θ|Zn).

4

Page 13: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Let us define the gradient vector g of L by g(θ|Zn) = ∂L(θ|Zn)/∂θ and the

Hessian matrix H by H(θ|Zn) = ∂2L(θ|Zn)/∂θ ∂θT . Here, ∂(·)/∂θ is the p×1 column

vector of elements ∂(·)/∂θi, i = 1, · · · , p, ∂(·)/∂θT = [∂(·)/∂θ]T , and ∂2(·)/∂θ ∂θT

is the p × p matrix of elements ∂2(·)/∂θi ∂θj, i, j = 1, · · · , p. Then, the p × p FIM,

Fn(θ), is defined [12, Section 13.3.2] as follows,

Fn(θ) ≡ E[g(θ|Zn) · gT (θ|Zn)

∣∣ θ]

= −E [H(θ|Zn)|θ] , (1.1)

provided that the derivatives and expectation (the expectation operator E is with

respect to the probability measure of Zn) exist. The equality ‘=’ in (1.1) is followed

[1, Proposition 3.4.4], [12, p. 352 − 353] by assuming that L is twice differentiable

with respect to θ and the regularity conditions [1, Section 3.4.2] hold for the like-

lihood function, `. The Hessian-based form above is more amenable to the practi-

cal computation for FIM than the gradient-based form that is used for defining the

FIM. It must be remarked here that there might be cases when the observed FIM,

−H(θ|Zn) ≡ −∂2L(θ|Zn)/∂θ ∂θT , at a given data set, Zn, is preferable to the actual

FIM, Fn(θ) = −E[H(θ|Zn)|θ], from an inference point of view and/or computational

simplicity (see e.g., [6] for scalar case, θ = θ1).

1.3 Estimation of Fisher Information Matrix

The analytical computation of Fn(θ), in a general case, is a daunting task since

it involves evaluation of the first or second order derivative of L and expectation of

5

Page 14: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

the resulting multivariate, often non-linear, function that become difficult or next

to impossible. This analytical problem can be circumvented by employing a recently

introduced simulation-based technique, resampling approach [12, Section 13.3.5], [13].

This algorithm is essentially based on producing a set of large (say N) number of

Hessian estimates from either the values of the log-likelihood function or (if available)

its exact stochastic gradient that, in turn, are computed from a set of pseudo data

vector, {Zpseudo(1), · · · ,Zpseudo(N)}, with each Zpseudo(i), i = 1, · · · , N , being digitally

simulated from pZn(·|θ) and statistically independent of each other. The set of pseudo

data vector acts as a proxy for the observed data set in the resampling algorithm.

The average of the negative of these Hessian estimates is reported as an estimate of

the Fn(θ).

A brief outline of the existing resampling algorithm is provided in the following

subsection.

1.3.1 Current Re-sampling Algorithm: No use of Prior In-

formation

For i-th pseudo data, let the k-th estimate of the Hessian matrix, H(θ|Zpseudo(i)),

in the resampling algorithm, be denoted by H(i)k . Then, H

(i)k , as per resampling

scheme, is computed as [13],

H(i)k =

1

2

δG(i)k

2 c

[∆−1

k1 , · · · , ∆−1kp

]+

(δG

(i)k

2 c

[∆−1

k1 , · · · , ∆−1kp

])T

, (1.2)

6

Page 15: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

in which c > 0 is a small number, δG(i)k ≡ G(θ+c∆k|Zpseudo(i))−G(θ−c∆k|Zpseudo(i))

and the perturbation vector ∆k = [∆k1, · · · , ∆kp]T is a user-generated random vec-

tor statistically independent of Zpseudo(i). The random variables, ∆k1, · · · , ∆kp, are

mean-zero and statistically independent and, also the inverse moments, E[|1/∆km|],

m = 1, · · · , p, are finite.

The symmetrizing operation (the multiplier 1/2 and the indicated sum) as shown

in (1.2) is useful in optimization problems to compute a symmetric estimate of the

Hessian matrix with finite samples [11]. This also maintains a symmetric estimate of

Fn(θ), which itself is a symmetric matrix.

Depending on the setting, G(·|Zpseudo(i)), as required in δG(i)k , represents the k-

th direct measurement or approximation of the gradient vector g(·|Zpseudo(i)). If the

direct measurement or computation of the exact gradient vector g is feasible, G(θ±

c∆k|Zpseudo(i)) represent the direct k-th measurements of g(·|Zpseudo(i)) at θ ± c∆k.

If the direct measurement or computation of g is not feasible, G(θ ± c∆k|Zpseudo(i))

represents the k-th approximation of g(θ ± c∆k|Zpseudo(i)) based on the values of

L(·|Zpseudo(i)).

If the direct measurements or computations of g are not feasible, G in (1.2) can

be computed by using the classical finite-difference (FD) technique [12, Section 6.3]

or the simultaneous perturbation (SP) gradient approximation technique [10], [12,

Section 7.2] from the values of L(·|Zpseudo(i)). For the computation of gradient ap-

proximation based on the values of L, there are advantages to using one-sided [12,

7

Page 16: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

p. 199] SP gradient approximation (relative to the standard two-sided SP gradient

approximation) in order to reduce the total number of function measurements or

evaluations for L. The SP technique for gradient approximation is quite useful when

p is large and usually superior to FD technique for estimation of Fn(θ) when the

resampling approach is employed to estimate the FIM. The formula for the one-sided

gradient approximation using SP technique is given by,

G(1)(θ ± c∆k|Zpseudo(i))

= (1/c)[L(θ + c∆k ± c∆k|Zpseudo(i))− L(θ ± c∆k|Zpseudo(i))

]

∆−1k1

...

∆−1kp

, (1.3)

in which superscript (1) in G(1) indicates that it is one-sided gradient approximation

(G = G(1)), c > 0 is a small number and ∆k = [∆k1, · · · , ∆kp]T is generated in the

same statistical manner as ∆k, but otherwise statistically independent of ∆k and

Zpseudo(i). It is usually recommended that c > c. Note that G(1)(·) in (1.3) not only

depends on θ, c, ∆k and Zpseudo(i) but also depends on c and ∆k that, however, have

been suppressed for notational clarity.

At this stage, let us also formally state that the perturbation vectors, ∆k and ∆k,

satisfy the following condition [10, 11], [12, Chapter 7].

C.1: (Statistical properties of the perturbation vector) The random vari-

ables, ∆km (and ∆km), k = 1, · · · , N , m = 1, · · · , p, are statistically indepen-

dent and almost surely (a.s.) uniformly bounded for all k, m, and, are also

8

Page 17: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

mean-zero and symmetrically distributed satisfying E[|1/∆km|] < ∞ (and

E[|1/∆km|] < ∞).

If the pseudo data vectors are expensive to simulate relative to the Hessian esti-

mate, then it is preferred to generate several Hessian estimates by generating more

than one (say, M > 1) perturbation vectors, ∆1, · · · ,∆M , for each pseudo data vector,

Zpseudo(i), and, thus, computing M Hessian estimates, H(i)1 , · · · , H(i)

M , by employing

(1.2) using these M perturbation vectors and the pseudo data vector, Zpseudo(i). Let

the average of, say, k Hessian estimates, H(i)1 , · · · , H(i)

k , for the i-th pseudo data,

Zpseudo(i), be denoted by H(i)k . The computation of H

(i)k can be conveniently carried

out by using the standard recursive representation of sample mean in contrast to

storing the matrices and averaging later. The recursive representation of H(i)k is given

by,

H(i)k =

k − 1

kH

(i)k−1 +

1

kH

(i)k , k = 1, · · · ,M, (1.4)

in which at the first step (k = 1), H(i)0 in the right-hand-side can be set to a matrix

of zeros or any arbitrary matrix of finite-valued elements to obtain the correct result.

If the sample mean of total M Hessian estimates is now denoted by H(i) ≡ H(i)M

(for the sake of simplification of notation), then an estimate, FM,N , of Fn(θ) is com-

puted by averaging H(1), · · · , H(N) and, taking the negative value of the resulting

average. A recursive representation identical to (1.4) for computing the sample mean

of H(1), · · · , H(N) is also recommended to avoid storage of the matrices . The current

resampling algorithm is schematically shown in Figure 1.2.

9

Page 18: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

(1)pseudoZ

Pseudo dataInput

H1

( )1^ ^H ( )

M

1, . . . . .,

^H1

( )N ^H( )

M

N

(1)H−

( )H

N−

.

.

.

.

.

.

.

.

.

− H(

) iN

egat

ive

aver

age

of

Z pseudo( )N , . . . . .,

Hessian estimates iHk

^ ( ) AverageH^ k

( )iof =iH

− ( )

FIM estimate,FM,N−

Figure 1.2: Schematic model for forming FIM estimate, FM,N ; adapted from [13].

However, it should be noted that M = 1 has certain optimality properties [13]

and it is assumed throughout the work presented in this thesis that for each pseudo

data vector, only one perturbation vector is generated and, thus, only one Hessian

estimate is computed. The current work can, however, be readily extended to the

case when M > 1. Therefore, from now on, the index of the pseudo data vector

will be changed from i to k. Consequently, the pseudo data vector will be denoted

by Zpseudo(k), k = 1, · · · , N , the difference in gradient and the Hessian estimate

in (1.2), respectively, will simply be denoted by δGk and Hk, k = 1, · · · , N , and

the notation of the one-sided gradient approximation in (1.3) will take the form of

G(1)(θ ± c∆k|Zpseudo(k)). Finally, the following simplification of notation for the

estimate of the FIM will also be used from now on,

Fn ≡ F1,N . (1.5)

10

Page 19: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

The set of simplified notations as just mentioned essentially conforms to the work

in [12, Section 13.3.5] that is a simplified version of [13] in which the resampling

approach was rigorously presented.

Let us also assume that the moments of ∆km and 1/∆km (and, of ∆km and 1/∆km)

up to fifth order exist (this condition will be used later in Section 2.3. Since ∆km

(and ∆km) is symmetrically distributed, 1/∆km (and 1/∆km) is also symmetrically

distributed implying that

I: (Statistical properties implied by C.1) All the odd moments of ∆km and

1/∆km (and of ∆km and 1/∆km) up to 5-th order are zeros, E[(∆km)q] = 0

and E[(1/∆km)q] = 0 (E[(∆km)q] = 0 and E[(1/∆km)q] = 0) , q = 1, 3, 5.

The random vectors ∆k (and ∆k) are also independent across k. The random vari-

ables ∆k1, · · · , ∆kp (and ∆k1, · · · , ∆kp) can also be chosen identically distributed. In

fact, independent and identically distributed (i.i.d.) (across both k and m) mean-

zero random variable satisfying C.1 is a perfectly valid choice for ∆km (and ∆km). In

particular, Bernoulli ±1 random variable for ∆km (and ∆km) is a valid — but not the

necessary — choice among other probability distributions satisfying C.1.

It must be noted from (1.2) and (1.3) that the current resampling algorithm would

will satisfactorily provided the following condition is satisfied.

C.2: (Smoothness of L) The log-likelihood function, L, is twice continuously

differentiable with respect to θ and all the possible second order derivatives

are a.s. (a.s. with respect to the probability measure of Zpseudo(k)) continuous

11

Page 20: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

and uniformly (in k) bounded ∀θ ∈ Sθ ⊂ Θ ⊆ Rp in which Sθ is a small

region containing θ that represents the value of θ at which the FIM needs

to estimated. (For example, if |∆kl| ≤ ρ and |∆kl| ≤ ρ, ∀k, l, for some small

numbers ρ, ρ > 0, then it is sufficient to consider Sθ as a hypercube with length

of its side being 2(cρ + cρ) and center being θ in Rp.)

In all the gradient approximation techniques as mentioned earlier (FD technique,

two-sided SP technique and one-sided SP technique), the gradient approximation,

G(θ|Zpseudo(k)), has OZ(c2) bias provided the following is satisfied.

C.2′ (Smoothness of L) The log-likelihood function, L, is thrice continuously

differentiable with respect to θ and all the possible third order derivatives are

a.s. continuous and bounded ∀θ ∈ Sθ.

The exact form of the bias terms are, however, different for different techniques [12,

Section 6.4.1, 7.2.2]. The subscripts in OZ(·) explicitly indicate that the ‘big-O’ term

depends on Zpseudo(k). This term is such that |OZ(c2)/c2| < ∞ a.s. as c −→ 0. The

explicit expression of bias, when one-sided SP technique is employed, is reported in

Section 2.3.1. In the case where the direct measurements or computations of g are

feasible, G is an unbiased gradient estimate (measurement) of g.

It is shown in the next chapter that the FIM estimate in (1.5) is accurate within

an (O(c2) + O(c2)) bias in the case where only the measurements of the log-likelihood

function L are available and within an O(c2) bias in the case where the measurements

12

Page 21: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

of the unbiased exact stochastic gradient vector g are available provided the following

condition (stronger than C.2 and C.2′) is satisfied.

C.3: (Smoothness of L) The log-likelihood function, L, is four times continu-

ously differentiable with respect to θ and all the possible fourth order deriva-

tives are a.s. continuous and uniformly (in k) bounded ∀θ ∈ Sθ.

Here, the ‘big-O’ terms, O(c2) and O(c2), satisfy the conditions, |O(c2)/c2| < ∞ as

c −→ 0 and |O(c2)/c2| < ∞ as c −→ 0. The FIM estimate can be made as accurate

as desired through reducing c and c and increasing the number of Hk values being

averaged.

In the next chapter, a modification is proposed in the existing resampling algo-

rithm that would enhance the statistical characteristics of the estimator of Fn(θ), a

part of which is analytically known a priori.

13

Page 22: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Chapter 2

Improved Re-sampling Algorithm

Let the k-th estimate of Hessian matrix, H(θ|Zpseudo(k)), per proposed resampling

algorithm be denoted by Hk. In this chapter, the estimator, Hk, is shown separately

for two different cases: Case 1 - when only the measurements of the log-likelihood

function L are available and, Case 2: when measurements of the exact gradient vector

g are available. To contrast the two cases, the superscript (L) is used in H(L)k and

H(L)k to represent the dependence of Hk and Hk on L measurements for Case 1 and,

similarly, the superscript (g) in H(g)k and H

(g)k for Case 2. (The superscripts, (L)

or (g), should not be confused with the superscript, (i), as appeared in (1.2) which

referred to the pseudo data vector. The superscript (i) was eventually suppressed in

consideration of M being 1 as indicated at p. 10.)

14

Page 23: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

2.1 Additional Notation

Denote the (i, j)-th element of Fn(θ) by Fij(θ) (non-bold character and suppress-

ing the subscript, n, in the symbolic notation representing the element of Fn(θ)) for

simplification of notation. Let Ii, i = 1, · · · , p, be the set of column indices of the

known elements of Fn(θ) for the i-th row and Ici be the complement of Ii. Consider

a p× p matrix F(given)n whose (i, j)-th element, F

(given)ij , is defined as follows,

F(given)ij =

Fij(θ), if j ∈ Ii

0, otherwise

, i = 1, · · · , p. (2.1)

Consider another p× p matrix Dk defined by,

Dk =

∆k1

∆k1

∆k1

∆k2· · · ∆k1

∆kp

∆k2

∆k1

∆k2

∆k2· · · ∆k2

∆kp

......

. . ....

∆kp

∆k1

∆kp

∆k2· · · ∆kp

∆kp

=

∆k1

...

∆kp

[∆−1k1 , · · · , ∆−1

kp ]

= ∆k [∆−1k1 , · · · , ∆−1

kp ]. (2.2)

together with a corresponding matrix Dk based on replacing all ∆ki in Dk by the cor-

responding ∆ki (note that Dk is symmetric when the perturbations are i.i.d. Bernoulli

distributed).

15

Page 24: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

2.2 Modified Re-sampling Algorithm using Prior

Information

The new estimate, Hk, is extracted from Hk0 that is defined next separately for

Case 1 and Case 2:

Case 1: only the measurements of the log-likelihood function L are available,

H(L)k0 = H

(L)k − 1

2

[DT

k (−F(given)n ) Dk +

(DT

k (−F(given)n ) Dk

)T]

. (2.3)

Case 2: measurements of the exact gradient vector g are available,

H(g)k0 = H

(g)k − 1

2

[(−F(given)

n ) Dk +((−F(given)

n ) Dk

)T]. (2.4)

The estimates, H(L)k and H

(g)k , are readily obtained from, respectively, H

(L)k0 in (2.3)

and H(g)k0 in (2.4) by replacing the (i, j)-th element of H

(L)k0 and H

(g)k0 with known values

of −Fij(θ), j ∈ Ii, i = 1, · · · , p. The new estimate, Fn, of Fn(θ) is then computed

by averaging the Hessian estimates, Hk, and taking negative value of the resulting

average. For convenience and, also since the main objective is to estimate the FIM,

not the Hessian matrix, the replacement of the (i, j)-th element of Hk0 with known

values of −Fij(θ), j ∈ Ii, i = 1, · · · , p, is not required at each k-th step. The matrix,

Fn0, can be first obtained by computing the (negative) average of the matrices, Hk0,

and subsequently, the (i, j)-th element of Fn0 can be replaced with the analytically

known elements, Fij(θ), j ∈ Ii, i = 1, · · · , p, of Fn(θ) to yield the new estimate, Fn.

Consequently, replacing needs to be done only once at the end of averaging. However,

if a better estimate of Hessian matrix is also required (e.g., in optimization [11, 14]),

16

Page 25: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

the replacing as mentioned above for Hessian estimates is recommended at each k-th

step.

The matrices, H(L)k in (2.3) or H

(g)k in (2.4), need to be computed by using the

existing resampling algorithm implying that H(L)k ≡ Hk or H

(g)k ≡ Hk as appropriate

with Hk being given by (1.2). Note that F(given)n as shown in the right-hand-sides

of (2.3) and (2.4) is known by (2.1). It must be noted as well that the random

perturbation vectors, ∆k in Dk and ∆k in Dk, as required in (2.3) and (2.4) must be

the same simulated values of ∆k and ∆k used in the existing resampling algorithm

while computing the k-th estimate, H(L)k or H

(g)k .

Next a summary of the salient steps, required to produce the estimate Fn (i.e.,

with appropriate superscript, F(L)n or F

(g)n ) of Fn(θ) per modified resampling algo-

rithm as proposed here, is presented below. Figure 2.1 is a schematic of the following

steps.

Step 0. Initialization: Construct the matrix F(given)n as defined by (2.1) based

on the analytically known elements of FIM. Determine θ, the sample

size n and the number (N) of pseudo data vectors that will be gener-

ated. Determine whether log-likelihood L(·) or gradient vector g(·) will

be used to compute the Hessian estimates Hk. Pick a small number c

(perhaps c = 0.0001) to be used for Hessian estimation (see (1.2)) and

if required another small number c (perhaps c = 0.00011) for gradient

approximation (see (1.3)). Set k = 1.

17

Page 26: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Step 1. At the k-th step perform the following tasks,

a. Generation of pseudo data: Based on θ, generate by MC sim-

ulation technique the k-th pseudo data vector of n pseudo mea-

surements Zpseudo(k).

b. Computation of Hk: Generate the k-th perturbation vector ∆k

by satisfying C.1 and if required the other perturbation vector

∆k by satisfying C.1 for gradient approximation. Using the k-th

pseudo data vector Zpseudo(k) and the perturbation vector(s) ∆k

or/and ∆k, evaluate Hk (i.e., H(L)k or H

(g)k ) by using (1.2).

c. Computation of Dk and Dk: Use ∆k or/and ∆k, as generated

in the above step, to construct the matrices, Dk or/and Dk, as

defined in section 2.1.

d. Computation of Hk0: Modify Hk as produced in Step 1b by

employing (2.3) or (2.4) as appropriate in order to generate Hk0

(i.e., H(L)k0 or H

(g)k0 ).

Step 2. Average of Hk0: Repeat Step 1 until N estimates, Hk0, are produced.

Compute the (negative) mean of these N estimates. (The standard

recursive representation of sample mean can be used here to avoid the

storage of N matrices, Hk0). The resulting (negative) mean is Fn0.

Step 3. Evaluation of Fn: The new estimate, Fn, of Fn(θ) per modified re-

18

Page 27: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

sampling algorithm is simply obtained by replacing the (i, j)-th element

of Fn0 with the analytically known elements, Fij(θ), j ∈ Ii, i = 1, · · · , p,

of Fn(θ). To avoid the possibility of having a non-positive semi-definite

estimate, it may be desirable to take the symmetric square root of the

square of the estimate (the sqrtm function in MATLAB may be useful

here).

(1)pseudoZ H1

^

H^N

Cur

rent

re−

sam

plin

gal

gorit

hm

HHessian

estimates, k^

Pseudo dataInput

(giv

en)

F nU

se

Hessianestimates, Hk

~

H1

~

HN

~H~

kN

egat

ive

aver

age

of

.

.

.

.

.

.

.

.

.

Z pseudo( )N

New FIMestimate,Fn

~

Figure 2.1: Schematic of algorithm for computing the new FIM estimate, Fn.

The new estimator, Fn, is better than Fn in the sense that it would preserve

exactly the analytically known elements of Fn(θ) as well as reduce the variances of

the estimators of the unknown elements of Fn(θ).

The next section sketches the theoretical basis of the modified resampling algo-

rithm as proposed in this section.

19

Page 28: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

2.3 Theoretical Basis for the Modified Re-sampling

Algorithm

It should be remarked here for the sake of clarification that the dependence of Hk

and Hk on θ, Zpseudo(k), ∆k, c or/and ∆k and c are suppressed throughout this work

for simplification of notation. Now, for further notational simplification, the subscript

‘pseudo’ in Zpseudo(k) and the dependence of Z(k) on k would be suppressed (note that

Zpseudo(k) is identically distributed across k because Zpseudo(k) ∼ pZn(·|θ)). Since, ∆k

is usually assumed to be statistically independent across k and an identical condition

for ∆k is also assumed, the dependence on k would also be suppressed in the future

development.

Let the (i, j)-th element of Hk and Hk be, respectively, denoted by Hij and Hij

with the appropriate superscript. In the following, the mean and variance of both

Hij and Hij are computed for two cases in two different subsections. It is also shown

that the variance of the new estimator, Hij, is less than the variance of the existing

estimator, Hij, yielding a better estimate of Fn(θ) in the process.

2.3.1 Case 1: only the measurements of L are available

The mean and variance of H(L)ij are computed first followed by the mean and

variance of H(L)ij . The subsequent comparison of variances will show the superiority

of H(L)k , which leads to the superiority of F

(L)n .

20

Page 29: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

In deducing the expressions of mean and variance in the following, the conditions,

C.1 and C.3, are required along with the following condition as stated below. (The

condition, C.2, only ensures that employing the resampling algorithm would yield

some estimate but it cannot guarantee the boundedness of several ‘big-O’ terms as

mentioned earlier or as will be encountered while deducing the expressions of mean

and variance below.)

C.4: (Existence of covariance involving Hlm, Hlm,s and Hlm,rs) All the com-

binations of covariance terms involving Hlm(θ|Z), Hlm,s(θ|Z) and Hlm,rs(θ|Z),

l, m, s, r = 1, · · · , p, exist ∀θ ∈ Sθ (see C.2 for definition of Sθ).

Mean and Variance of H(L)ij

It is assumed here that the gradient estimate is based on one-sided gradient ap-

proximation using SP technique given by (1.3). Based on a Taylor expansion, the i-th

component of G(1)(θ|Z), i = 1, · · · , p, that is an approximation of the i-th component,

gi(θ|Z) ≡ ∂L(θ|Z)/∂θi, of g(θ|Z) based on the values of L(·|Z), is given by,

G(1)i (θ|Z) =

L(θ + c∆|Z)− L(θ|Z)

c ∆i

(2.5)

=1

c ∆i

[(L(θ|Z) + c

p∑

l=1

gl(θ|Z)∆l +1

2c2

p∑

l=1

p∑m=1

Hlm(θ|Z)∆m∆l

+1

6c3

p∑

l=1

p∑m=1

p∑s=1

∂3L(θ|Z)

∂θs∂θm∂θl

∆s∆m∆l

)− L(θ|Z)

]

=∑

l

gl(θ)∆l

∆i

+1

2c∑

l,m

Hlm(θ)∆m∆l

∆i

+1

6c2

l,m,s

∂Hlm(θ)

∂θs

∆s∆m∆l

∆i

, (2.6)

21

Page 30: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

in which Hlm(θ|Z) ≡ ∂2L(θ|Z)/∂θl ∂θm is the (l, m)-th element of H(θ|Z), θ =

λ(θ + c∆) + (1 − λ)θ = θ + cλ∆ (with λ ∈ [0, 1] being some real number) denotes

a point on the line segment between θ and θ + c∆ and, in the expression after

the last equality, the condition on Z is suppressed for notational clarity and, also

the summations are expressed in abbreviated format where the indices span their

respective and appropriate ranges. This convention would be followed throughout

this work unless the explicit dependence of Z and an expanded summation format

need to be invoked for the sake of clarity of the context to be discussed.

Now note that

E[G(1)i (θ|Z)|θ,Z] =

l

gl(θ|Z)E

[∆l

∆i

]+ 0 +

1

6c2

l,m,s

∂Hlm(θ|Z)

∂θs

E

[∆s∆m∆l

∆i

]

= gi(θ|Z) + OZ(c2) (2.7)

in which the point of evaluation, θ, is suppressed in the random ‘big-O’ term,

OZ(c2) = (1/6)c2∑

l,m,s[∂Hlm(θ|Z)/∂θs]E[(∆s∆m∆l)/∆i], for notational clarity. The

subscript in OZ(·) explicitly indicate that this term depends on Z satisfying |OZ(c2)/c2|

< ∞ almost surely (a.s.) (a.s. with respect to the joint probability measure of Z)

as c −→ 0 by C.1 and C.2′. The first equality follows from the fact that the second

term in (2.6) vanishes upon expectation since it involves either E[∆r] or E[1/∆r],

r = 1, · · · , p, both of which are zero by implication I. The second equality follows by

the condition C.1.

Now Gi(θ ± c∆|Z) ≡ G(1)i (θ ± c∆|Z), that is required for computing H

(L)ij , need

to be calculated. Suppressing the condition on Z (again for simplification of no-

22

Page 31: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

tation, specially when the dependence is not important in the context of the topic

being discussed) and using the abbreviated summation notation as indicated earlier,

G(1)i (θ±c∆) with subsequent and further Taylor expansion can be written as follows,

G(1)i (θ ± c∆)

=∑

l

gl(θ ± c∆)∆l

∆i︸ ︷︷ ︸1st part

+1

2c∑

l,m

Hlm(θ ± c∆)∆m∆l

∆i︸ ︷︷ ︸2nd part

+1

6c2

l,m,s

∂Hlm(θ ± c∆)

∂θs

∆s∆m∆l

∆i︸ ︷︷ ︸3rd part

=

[∑

l

gl(θ)∆l

∆i

+ (±c)∑

l,m

Hlm(θ)∆m∆l

∆i

+1

2(±c)2

l,m,s

∂Hlm(θ)

∂θs

∆s∆m∆l

∆i

+1

6(±c)3

l,m,s,r

∂2Hlm(θ±1 )

∂θr∂θs

∆r∆s∆m∆l

∆i

]+

[1

2c∑

l,m

Hlm(θ)∆m∆l

∆i

+1

2c(±c)

l,m,s

∂Hlm(θ±2 )

∂θs

∆s∆m∆l

∆i

]+

[1

6c2

l,m,s

∂Hlm(θ)

∂θs

∆s∆m∆l

∆i

+1

6c2(±c)

l,m,s,r

∂2Hlm(θ±)

∂θr∂θs

∆r∆s∆m∆l

∆i

]. (2.8)

Here, θ±1 and θ±2 denote some points on the line segments between θ and θ± c∆, θ±

denote some points on the line segments between θ and θ ± c∆, the superscript, ±

sign, in θ±1 , θ±2 and θ±

corresponds to whether θ + c∆ or θ− c∆ is being considered

for the argument of G(1)i (·) in (2.8).

Given Gi(θ ± c∆|Z) ≡ G(1)i (θ ± c∆|Z) by (2.8), the (i, j)-th element of H

(L)k can

be readily obtained from,

H(L)ij =

1

2

[Gi(θ + c∆|Z)−Gi(θ − c∆|Z)

2 c ∆j

+Gj(θ + c∆|Z)−Gj(θ − c∆|Z)

2 c ∆i

]. (2.9)

Consider a typical term (say, the first term in the right-hand-side of (2.9)) and denote

it by J(L)ij (J is to indicate Jacobian for which the symmetrizing operation should not

23

Page 32: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

be used). By the use of (2.8), J(L)ij , then can be written as,

J(L)ij ≡ Gi(θ + c∆|Z)−Gi(θ − c∆|Z)

2 c ∆j

=

[∑

l,m

Hlm(θ|Z)∆m

∆j

∆l

∆i

+O∆,∆,Z(c2)

]

︸ ︷︷ ︸1st part

+[O∆,∆,Z(c)

]︸ ︷︷ ︸

2nd part

+[O∆,∆,Z(c2)

]︸ ︷︷ ︸

3rd part

.(2.10)

The subscripts in O∆,∆,Z(·) explicitly indicate that the random ‘big-O’ terms depend

on ∆, ∆ and Z. By the use of C.1 and C.3, it should be noted that these terms are

such that |O∆,∆,Z(c2)/c2| < ∞ almost surely (a.s.) (a.s. with respect to the joint

probability measure of ∆, ∆ and Z) as c −→ 0 and, both |O∆,∆,Z(c)/c| < ∞ a.s. and

|O∆,∆,Z(c2)/c2| < ∞ a.s. as c −→ 0. The explicit expressions of the ‘big-O’ terms in

(2.10) are shown below separately,

O∆,∆,Z(c2) =1

12c2

l,m,s,r

(∂2Hlm(θ+

1 |Z)

∂θr∂θs

+∂2Hlm(θ−1 |Z)

∂θr∂θs

)∆r∆s∆m

∆j

∆l

∆i

,

O∆,∆,Z(c) =1

4c∑

l,m,s

(∂Hlm(θ+

2 |Z)

∂θs

+∂Hlm(θ−2 |Z)

∂θs

)∆s

∆j

∆m∆l

∆i

,

O∆,∆,Z(c2) =1

12c2

l,m,s,r

(∂2Hlm(θ

+|Z)

∂θr∂θs

+∂2Hlm(θ

−|Z)

∂θr∂θs

)∆r

∆j

∆s∆m∆l

∆i

.

The effects of O∆,∆,Z(c2) are not included in O∆,∆,Z(c). The reason for showing

O∆,∆,Z(c) separately in (2.10) is that this term vanishes upon expectation because it

involves either E[∆r] or E[1/∆r], r = 1, · · · , p, both of which are zero by implication

I and rest of the terms appeared in O∆,∆,Z(c) do not depend on ∆ (as mentioned

earlier below (2.8) that θ±2 in O∆,∆,Z(c) depend only on θ and θ ± c∆). The other

24

Page 33: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

terms, O∆,∆,Z(c2) and O∆,∆,Z(c2), do not vanish upon expectation yielding,

E[J(L)ij |θ] = E

[∑

l,m

Hlm(θ|Z)∆m

∆j

∆l

∆i

∣∣∣∣∣θ

]+ E

[O∆,∆,Z(c2)|θ]

+E[O∆,∆,Z(c)|θ]

+ E[O∆,∆,Z(c2)|θ]

=∑

l,m

E

[Hlm(θ|Z)

∆m

∆j

∆l

∆i

∣∣∣∣∣θ

]+ O(c2) + 0 + O(c2)

=∑

l,m

E [Hlm(θ|Z)|θ] E

[∆m

∆j

]E

[∆l

∆i

]+ O(c2) + O(c2)

= E [Hij(θ|Z)|θ] + O(c2) + O(c2) (2.11)

Note that the ‘big-O’ terms, O(c2) and O(c2), satisfying |O(c2)/c2| < ∞ as c −→ 0

and |O(c2)/c2| < ∞ as c −→ 0 (by C.1 and C.3), are deterministic unlike the random

‘big-O’ terms in (2.10). The third equality follows from the fact that ∆, ∆ and Z are

statistically independent of each other. The fourth equality follows by the condition

C.1. In (2.11), E [Hij(θ|Z)|θ] has not been substituted by −Fij(θ) in order to keep

the derivation of (2.11) most general since (2.11) will be applicable as well for Jacobian

estimate. However, in the context of FIM, E [Hij(θ|Z)|θ] = −Fij(θ) by (Hessian-

based) definition using which E[H(L)ij |θ] is obtained as,

E[H(L)ij |θ] =

1

2

(E[J

(L)ij |θ] + E[J

(L)ji |θ]

)

=1

2(E [Hij(θ|Z)|θ] + E [Hji(θ|Z)|θ]) + O(c2) + O(c2)

= −Fij(θ) + O(c2) + O(c2). (2.12)

The last equality follows from the fact that the FIM, Fn(θ), is symmetric.

25

Page 34: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

The variance of H(L)ij is to be computed next. It is given by,

var[H(L)ij |θ] =

1

4var[(J

(L)ij + J

(L)ji )|θ]

=1

4

(var[J

(L)ij |θ] + var[J

(L)ji |θ] + 2cov[J

(L)ij , J

(L)ji |θ]

). (2.13)

The expression of a typical variance term, var[J(L)ij |θ], in (2.13) would now be deter-

mined followed by the deduction of the expression of covariance term, cov[J(L)ij , J

(L)ji |θ].

By the use of (2.10), var[J(L)ij |θ] is given by,

var[J(L)ij |θ] = var

[(∑

l,m

Hlm(θ|Z)∆m

∆j

∆l

∆i

+ O∆,∆,Z(c2) + O∆,∆,Z(c)

+ O∆,∆,Z(c2)

)∣∣∣∣θ]

= var

[∑

l,m

Hlm(θ|Z)∆m

∆j

∆l

∆i

∣∣∣∣∣θ

]+var

[O∆,∆,Z(c2)|θ]

+var[O∆,∆,Z(c)|θ]

+ var[O∆,∆,Z(c2)|θ]

+ 2cov[O∆,∆,Z(1), O∆,∆,Z(c2)|θ]

+ 2cov[O∆,∆,Z(1), O∆,∆,Z(c)|θ]

+ 2cov[O∆,∆,Z(1), O∆,∆,Z(c2)|θ]

+ 2cov[O∆,∆,Z(c2), O∆,∆,Z(c)|θ]

+2cov[O∆,∆,Z(c2), O∆,∆,Z(c2)|θ]

+ 2cov[O∆,∆,Z(c), O∆,∆,Z(c2)|θ]

, (2.14)

in which O∆,∆,Z(1) represents the term∑

l,m Hlm(θ|Z)(∆m/∆j)(∆l/∆i). Note that

the variance terms, var[O∆,∆,Z(c2)|θ]

, var[O∆,∆,Z(c)|θ]

and var[O∆,∆,Z(c2)|θ]

, re-

spectively, turn out to be O(c4), O(c2) and O(c4) by C.4. While the covariance terms,

2cov[O∆,∆,Z(1), O∆,∆,Z(c2)|θ]

, 2cov[O∆,∆,Z(1), O∆,∆,Z(c2)|θ]

and 2cov[O∆,∆,Z(c2),

O∆,∆,Z(c2)|θ], respectively, turn out to be O(c2), O(c2) and O(c2c2) by C.4, a

few other covariance terms — 2cov[O∆,∆,Z(1), O∆,∆,Z(c)|θ]

and 2cov[O∆,∆,Z(c2),

26

Page 35: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

O∆,∆,Z(c)|θ]— vanish since, for any given i, they involve terms like E[(∆l1∆m2∆l2)/

∆2i ], l1, l2,m2 = 1, · · · , p, all of which are zero by implication I and rest of the terms ap-

peared in O∆,∆,Z(c) and O∆,∆,Z(c2) do not depend on ∆ (as already mentioned below

(2.8) that θ±1 in O∆,∆,Z(c2) and θ±2 in O∆,∆,Z(c) depend only on θ and θ± c∆). Al-

though the remaining term, 2cov[O∆,∆,Z(c), O∆,∆,Z(c2)|θ]

, in (2.14) involves terms

like E[(∆m1∆l1∆l2∆m2∆n2)/∆2i ], l1,m1, n1, l2,m2, n2 = 1, · · · , p, which are zero by

implication I, the ‘big-O’ term, O∆,∆,Z(c2), contains θ that depends on θ and θ+ c∆

(see below (2.6)). Therefore, 2cov[O∆,∆,Z(c), O∆,∆,Z(c2)|θ]

is not zero and turns out

to be O(c3) by C.4. Then, by absorbing the term O(c4) into O(c2) and both O(c4)

and O(c3) into O(c2), the variance, var[J(L)ij |θ], turns out to be,

var[J(L)ij |θ] = var

[∑

l,m

Hlm(θ|Z)∆m

∆j

∆l

∆i

∣∣∣∣∣θ

]+ O(c2) + O(c2) + O(c2c2)

= E

(∑

l,m

Hlm(θ|Z)∆m

∆j

∆l

∆i

)2∣∣∣∣∣∣θ

(E

[∑

l,m

Hlm(θ|Z)∆m

∆j

∆l

∆i

∣∣∣∣∣ θ

])2

+ O(c2) + O(c2) + O(c2c2)

=∑

l1,m1,l2,m2

E [{Hl1m1(θ|Z)Hl2m2(θ|Z)}|θ] E

[∆m1∆m2

∆2j

]E

[∆l1∆l2

∆2i

]

− (E [Hij(θ|Z)|θ])2 + O(c2) + O(c2) + O(c2c2)

=∑

l,m

E[{Hlm(θ|Z)}2

∣∣ θ]E

[∆2

m

∆2j

]E

[∆2

l

∆2i

]− (E [Hij(θ|Z)|θ])2

+ O(c2) + O(c2) + O(c2c2)

=∑

l,m

{var [Hlm(θ|Z)|θ] + (E [Hlm(θ|Z)|θ])2} E

[∆2

m

∆2j

]E

[∆2

l

∆2i

]

− (E [Hij(θ|Z)|θ])2 + O(c2) + O(c2) + O(c2c2)

27

Page 36: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

=∑

l,m

alm(i, j)var [Hlm(θ|Z)|θ] +∑l,m

lm6=ij

alm(i, j) (E [Hlm(θ|Z)|θ])2

+ O(c2) + O(c2) + O(c2c2) (2.15)

in which alm(i, j) = E[∆2m/∆2

j ]E[∆2l /∆

2i ]. The first term after the third equality

follows from the fact that ∆, ∆ and Z are statistically independent of each other and

the second term has already been seen during the derivation of E[J(L)ij |θ] in (2.11).

The fourth equality follows by condition C.1 and implication I. Note that var[H(L)ii |θ]

follows straight from the expression of var[J(L)ij |θ] by replacing j with i in (2.15).

Next, the expression of cov[J(L)ij , J

(L)ji |θ], j 6= i, would be deduced. By using an

identical argument to the one that is mentioned right after (2.14), it can be readily

shown that cov[J(L)ij , J

(L)ji |θ] is given by,

cov[J(L)ij , J

(L)ji |θ] = cov

[(∑

l,m

Hlm(θ|Z)∆m

∆j

∆l

∆i

),

(∑

l,m

Hlm(θ|Z)∆m

∆i

∆l

∆j

)∣∣∣∣∣ θ

]

+ O(c2) + O(c2) + O(c2c2). (2.16)

Further simplification of the first term in the right-hand-side yields,

cov

[(∑

l,m

Hlm(θ|Z)∆m

∆j

∆l

∆i

),

(∑

l,m

Hlm(θ|Z)∆m

∆i

∆l

∆j

)∣∣∣∣∣ θ

]

= E

[(∑

l,m

Hlm(θ|Z)∆m

∆j

∆l

∆i

)(∑

l,m

Hlm(θ|Z)∆m

∆i

∆l

∆j

)∣∣∣∣∣ θ

]

−E [Hij(θ|Z)|θ] E [Hji(θ|Z)|θ]

=∑

l1,m1,l2,m2

E [{Hl1m1(θ|Z)Hl2m2(θ|Z)}|θ] E

[∆m1∆m2

∆i∆j

]E

[∆l1∆l2

∆i∆j

]− (Fij(θ))2

={

E [Hij(θ|Z)Hji(θ|Z)|θ] + E [Hii(θ|Z)Hjj(θ|Z)|θ] + E [Hjj(θ|Z)Hii(θ|Z)|θ]

+ E [Hji(θ|Z)Hij(θ|Z)|θ]}− (Fij(θ))2

28

Page 37: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

= 2E [Hij(θ|Z)Hji(θ|Z)|θ] + 2E [Hii(θ|Z)Hjj(θ|Z)|θ]− (Fij(θ))2

= 2E[{Hij(θ|Z)}2

∣∣ θ]+ 2E [Hii(θ|Z)Hjj(θ|Z)|θ]− (Fij(θ))2

= 2{var [(Hij(θ|Z))|θ] + (E [ (Hij(θ|Z))|θ])2} + 2E [Hii(θ|Z)Hjj(θ|Z)|θ]

− (Fij(θ))2 . (2.17)

Here, the first term after the second equality follows from the fact that ∆, ∆ and Z are

statistically independent of each other and the second term after the second equality

follows by the Hessian-based definition and symmetry of FIM. The third equality

follows by the use of the fact that j 6= i and by condition C.1 and implication I.

Therefore, cov[J(L)ij , J

(L)ji |θ] turns out to be,

cov[J(L)ij , J

(L)ji |θ]

= 2{var [(Hij(θ|Z))|θ] + (E [ (Hij(θ|Z))|θ])2} + 2E [Hii(θ|Z)Hjj(θ|Z)|θ]

− (Fij(θ))2 + O(c2) + O(c2) + O(c2c2), for j 6= i. (2.18)

Now, the variance of H(L)ij , var[H

(L)ij |θ], for j 6= i, can be readily obtained from

(2.13) by using (2.15) and (2.18). As already indicated, var[H(L)ii |θ] is same as

var[J(L)ii |θ] that can be directly obtained from (2.15) by replacing j with i. Since

the primary objective of the current work is to show that the variance of the new

estimators, Fij, of the elements, Fij(θ), of Fn(θ) are less than that of the current

estimators, Fij, it would suffice to show that var[H(L)ij |θ] is less than var[H

(L)ij |θ].

This later fact can be proved by comparing the contributions of the variance and

covariance terms to var[H(L)ij |θ] with the contributions of the respective variance and

29

Page 38: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

covariance terms (as appeared in (2.13)) to var[H(L)ij |θ]. This is the task of the next

section.

Mean and Variance of H(L)ij

We now derive the mean and variance of the element of H(L)ij for the purpose of

contrasting with the same quantities for H(L)ij . Consider the (i, j)-th element of Hk

associated with (2.3) that is given by,

H(L)ij = H

(L)ij − 1

2

l

m∈Il

[(−Flm(θ))

∆m

∆j

∆l

∆i

+ (−Flm(θ))∆m

∆i

∆l

∆j

], ∀j ∈ Ici

=1

2

(J

(L)ij + J

(L)ji

)− 1

2

l

m∈Il

[(−Flm(θ))

∆m

∆j

∆l

∆i

+ (−Flm(θ))∆m

∆i

∆l

∆j

], ∀j ∈ Ici

=1

2

(J

(L)ij + J

(L)ji

), ∀j ∈ Ici (2.19)

and

H(L)ij = −Fij(θ), ∀j ∈ Ii. (2.20)

In (2.19), J(L)ij is defined as,

J(L)ij = J

(L)ij −

l

m∈Il(−Flm(θ))

∆m

∆j

∆l

∆i

, ∀j ∈ Ici . (2.21)

Note that ∀j ∈ Ici in (2.21),

E

[∑

l

m∈Il(−Flm(θ))

∆m

∆j

∆l

∆i

∣∣∣∣∣θ

]= 0,

implying that E[J(L)ij ] = E[J

(L)ij ], ∀j ∈ Ici . By using this fact along with (2.19), (2.11)

and the symmetry of the FIM (for all j ∈ Ici) and by using (2.20) (for all j ∈ Ii), the

30

Page 39: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

following can be obtained readily, ∀i = 1, · · · , p,

E[H(L)ij |θ] =

−Fij(θ) + O(c2) + O(c2), ∀j ∈ Ici ,

−Fij(θ), ∀j ∈ Ii.(2.22)

Noticeably, the expressions of the ‘big-O’ terms both in (2.12) and in the first equation

of (2.22) are precisely same implying that E[H(L)ij |θ] = E[H

(L)ij |θ], ∀j ∈ Ici .

While var[H(L)ij |θ] = 0, ∀j ∈ Ii, clearly implying that var[H

(L)ij |θ] < var[H

(L)ij |θ],

∀j ∈ Ii, deduction of the expression of var[H(L)ij |θ], ∀j ∈ Ici , is the task that will be

considered now. In fact, this is the main result associated with the variance reduction

from prior information available in terms of the known elements of Fn(θ).

The first step in determining this variance is to note that the expression of J(L)ij

in (2.10) can be decomposed into two parts as shown below,

J(L)ij =

l

m∈IlHlm(θ|Z)

∆m

∆j

∆l

∆i

+∑

m∈IclHlm(θ|Z)

∆m

∆j

∆l

∆i

+ O∆,∆,Z(c2) + O∆,∆,Z(c) + O∆,∆,Z(c2). (2.23)

The elements, Hlm(θ|Z), of H(θ|Z) in the right-hand-side of (2.23) are not known.

However, since by (Hessian-based) definition E[Hlm(θ|Z)|θ] = −Flm(θ), approxima-

tion of the unknown elements of H(θ|Z) in the right-hand-side of (2.23), particularly

those elements that correspond to the elements of the FIM that are known a priori,

by the negative of those elements of Fn(θ) is the primary idea based on which the

modified resampling algorithm is developed. This approximation introduces an error

term, elm(θ|Z), that can be defined by,

Hlm(θ|Z) = −Flm(θ) + elm(θ|Z), ∀m ∈ Il, l = 1, · · · , p, (2.24)

31

Page 40: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

and this error term satisfies the following two conditions that directly follow from

(2.24), ∀m ∈ Il, l = 1, · · · , p,

E[elm(θ|Z)|θ] = 0, (2.25)

var[elm(θ|Z)|θ] = var[Hlm(θ|Z)|θ]. (2.26)

Substitution of (2.24) in (2.23) results in a known part in the right-hand-side of (2.23)

involving the analytically known elements of FIM. This known part is transferred

to the left-hand-side of (2.23) and, consequently, acts as a feedback to the current

resampling algorithm yielding, in the process, J(L)ij ,

J(L)ij ≡

(J

(L)ij −

l

m∈Il(−Flm(θ))

∆m

∆j

∆l

∆i

)

=∑

l

m∈Ilelm(θ|Z)

∆m

∆j

∆l

∆i

+∑

m∈IclHlm(θ|Z)

∆m

∆j

∆l

∆i

+ O∆,∆,Z(c2) + O∆,∆,Z(c) + O∆,∆,Z(c2), ∀j ∈ Ici . (2.27)

This expression of J(L)ij can be employed to show that the new estimator, H

(L)ij , in

(2.19) has less variance than that of the current estimator, H(L)ij . This is achieved in

the following.

Introduce Xlm, l = 1, · · · , p, as defined below,

Xlm(θ|Z) =

elm(θ|Z), if m ∈ Il,

Hlm(θ|Z), if m ∈ Icl .(2.28)

Then, (2.27) can be compactly written as,

J(L)ij =

l,m

Xlm(θ|Z)∆m

∆j

∆l

∆i

+O∆,∆,Z(c2)+O∆,∆,Z(c)+O∆,∆,Z(c2), ∀j ∈ Ici . (2.29)

32

Page 41: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

On the other hand, variance of H(L)ij in (2.19) is given by,

var[H(L)ij |θ] =

1

4var[(J

(L)ij + J

(L)ji )|θ]

=1

4

(var[J

(L)ij |θ] + var[J

(L)ji |θ] + 2cov[J

(L)ij , J

(L)ji |θ]

). (2.30)

Therefore, var[J(L)ij |θ] and cov[J

(L)ij , J

(L)ji |θ] are required to be computed next in order

to compare them, respectively, with var[J(L)ij |θ] in (2.15) and cov[J

(L)ij , J

(L)ji |θ] in (2.18)

and, consequently, facilitating the comparison between var[H(L)ij |θ] and var[H

(L)ij |θ],

∀j ∈ Ici .

The variance of J(L)ij , ∀j ∈ Ici , can be computed by considering the right-hand-side

of (2.29) in an identical way as described for J(L)ij in the previous subsection. The

expression for var[J(L)ij |θ], ∀j ∈ Ici , follows readily from (2.15) by replacing Hlm with

Xlm because of the similarity between (2.10) and (2.29) as shown below, ∀j ∈ Ici ,

i = 1, · · · , p,

var[J(L)ij |θ] =

l,m

alm(i, j)var [Xlm(θ|Z)|θ] +∑l,m

lm6=ij

alm(i, j) (E [Xlm(θ|Z)|θ])2

+ O(c2) + O(c2) + O(c2c2)

=∑

l,m

alm(i, j)var [Hlm(θ|Z)|θ] +∑

l

∑m∈Ic

llm6=ij

alm(i, j) (E [Hlm(θ|Z)|θ])2

+ O(c2) + O(c2) + O(c2c2). (2.31)

The second equality follows by the use of (2.28) and a subsequent use of (2.25)-(2.26)

on the resulting expression.

33

Page 42: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Subtracting (2.31) from (2.15), it can be readily seen that ∀j ∈ Ici ,

var[J(L)ij |θ]− var[J

(L)ij |θ] =

l

m∈Ilalm(i, j) (E [Hlm(θ|Z)|θ])2

+ O(c2) + O(c2) + O(c2c2)

=∑

l

m∈Ilalm(i, j) (Flm(θ))2

+ O(c2) + O(c2) + O(c2c2) > 0. (2.32)

Here, the second equality follows by the Hessian-based definition of FIM, E[Hlm(θ|Z)

|θ] = −Flm(θ), and the inequality follows from the fact that alm(i, j) = (E[∆2m/∆2

j ]

E[∆2l /∆

2i ]) > 0, l, m = 1, · · · , p, for any given (i, j) and assuming that at least one of

the known elements, Flm(θ), in (2.32) is not equal to zero. It must be remarked that

the bias terms, O(c2), O(c2) and O(c2c2), can be made negligibly small by selecting

c and c small enough that are primarily controlled by users. Note that if ∆1, · · · , ∆p

and ∆1, · · · , ∆p are both assumed to be Bernoulli ±1 i.i.d. random variables, then

alm(i, j) turns out to be unity.

At this point it is already clear that var[H(L)ii |θ] < var[H

(L)ii |θ] if j = i ∈ Ici ,

since var[H(L)ii |θ] and var[H

(L)ii |θ] are exactly given by var[J

(L)ii |θ] and var[J

(L)ii |θ],

respectively, and var[J(L)ii |θ] < var[J

(L)ii |θ] by (2.32).

Next step is to compare cov[J(L)ij , J

(L)ji |θ] to cov[J

(L)ij , J

(L)ji |θ], j 6= i, ∀j ∈ Ici , in

order to conclude that var[H(L)ij |θ] < var[H

(L)ij |θ]. As var[J

(L)ij |θ] is deduced from

var[J(L)ij |θ] by the similarity of the expressions between (2.10) and (2.29), the ex-

pression of cov[J(L)ij , J

(L)ji |θ] also follows readily by replacing Hlm with Xlm and using

34

Page 43: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

identical arguments that are used in deducing (2.16) and (2.17). The expression of

cov[J(L)ij , J

(L)ji |θ], therefore, turns out to be, for j 6= i, ∀j ∈ Ici , i = 1, · · · , p,

cov[J(L)ij , J

(L)ji |θ] = 2

{var [(Xij(θ|Z))|θ] + (E [ (Xij(θ|Z))|θ])2}

+ 2E [Xii(θ|Z)Xjj(θ|Z)|θ]−(Fij(θ))2+O(c2)+O(c2)+O(c2c2)

= 2{var [(Hij(θ|Z))|θ] + (E [ (Hij(θ|Z))|θ])2}

+ 2E [Xii(θ|Z)Xjj(θ|Z)|θ]−(Fij(θ))2 + O(c2)+O(c2)+O(c2c2).

(2.33)

Since j ∈ Ici , the second equality follows from the definition of Xij in (2.28). In (2.33),

Xii and Xjj must take the form from one of following four possibilities: (1) eii and

ejj, (2) eii and Hjj, (3) Hii and ejj, (4) Hii and Hjj. Subtracting (2.33) from (2.18),

the difference between cov[J(L)ij , J

(L)ji |θ] and cov[J

(L)ij , J

(L)ji |θ] is given by,

cov[J(L)ij , J

(L)ji |θ]− cov[J

(L)ij , J

(L)ji |θ]

= 2 (E [Hii(θ|Z)Hjj(θ|Z)|θ]− E [Xii(θ|Z)Xjj(θ|Z)|θ])

+O(c2) + O(c2) + O(c2c2), for j 6= i,∀j ∈ Ici . (2.34)

For the first possibility, it is shown below that E[Hii(θ|Z)Hjj(θ|Z)|θ] − E[Xii(θ|Z)

Xjj(θ|Z)|θ] is given by Fii(θ)Fjj(θ) that is greater than 0 since Fn(θ) is positive

definite matrix and a positive definite matrix always has non-zero positive diagonal

entries.

E [Hii(θ|Z)Hjj(θ|Z)|θ]− E [Xii(θ|Z)Xjj(θ|Z)|θ]

35

Page 44: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

= E [Hii(θ|Z)Hjj(θ|Z)|θ]− E [eii(θ|Z)ejj(θ|Z)|θ]

= E [Hii(θ|Z)Hjj(θ|Z)|θ]− E [ (Hii(θ|Z) + Fii(θ)) (Hjj(θ|Z) + Fjj(θ))|θ]

= Fii(θ)Fjj(θ)

> 0.

Here, the second equality follows by (2.24) and the third equality follows by the

Hessian-based definition of FIM, E[Hlm(θ|Z)|θ] = −Flm(θ). While the above dif-

ference is 0 for the fourth possibility (Xii and Xjj, respectively, being same as Hii

and Hjj), it can be shown easily in a similar way as determined above for the first

possibility that the difference again becomes Fii(θ)Fjj(θ) > 0 for the second (Xii and

Xjj, respectively, being same eii and Hjj) and third (Xii and Xjj, respectively, being

same as Hii and ejj) possibilities.

Therefore, using (2.32) and the fact that E[Hii(θ|Z)Hjj(θ|Z)|θ] − E[Xii(θ|Z)

Xjj(θ|Z)|θ] ≥ 0, that appeared in (2.34), and also noting that the bias terms, O(c2),

O(c2) and O(c2c2), can be made negligibly small by selecting c and c small enough,

the following can be concluded immediately from (2.13) and (2.30),

var[H(L)ij |θ] < var[H

(L)ij |θ], i, j = 1, · · · , p.

Here, the ranges of i and j are mentioned as {1, · · · , p} not just j ∈ Ici since for the

other case, j ∈ Ii, it has already been noted that var[H(L)ij |θ] = 0, by (2.20), that is

less than var[H(L)ij |θ].

The objective of the next section is to show that the variance of H(g)ij is less than

36

Page 45: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

the variance of H(g)ij .

2.3.2 Case 2: measurements of g are available

The mean and variance of H(g)ij are computed first followed by the mean and

variance of H(g)ij .

Mean and Variance of H(g)ij

In this section, it is assumed that the measurements of exact gradient vector, g, are

available. Therefore, the (i, j)-th element, H(g)ij , (the dependence on k is suppressed)

of H(g)k is directly given by,

H(g)ij =

1

2

[gi(θ + c∆|Z)− gi(θ − c∆|Z)

2 c ∆j

+gj(θ + c∆|Z)− gj(θ − c∆|Z)

2 c ∆i

]. (2.35)

As earlier, consider a typical term (say, the first term in the right-hand-side of (2.35))

and denote it by J(g)ij which, based on a Taylor expansion, turns out to be as follows,

J(g)ij ≡ gi(θ + c∆|Z)− gi(θ − c∆|Z)

2 c ∆j

=1

2 c ∆j

[(gi(θ|Z) + c

l

Hil(θ|Z)∆l +1

2c2

l,m

∂Hil(θ|Z)

∂θm

∆m∆l

+1

6c3

l,m,s

∂2Hil(θ+1 |Z)

∂θs∂θm

∆s∆m∆l

)−

(gi(θ|Z) + (−c)

l

Hil(θ|Z)∆l

+1

2(−c)2

l,m

∂Hil(θ|Z)

∂θm

∆m∆l +1

6(−c)3

l,m,s

∂2Hil(θ−1 |Z)

∂θs∂θm

∆s∆m∆l

)]

=∑

l

Hil(θ|Z)∆l

∆j

+1

12c2

l,m,s

(∂2Hil(θ

+1 |Z)

∂θs∂θm

+∂2Hil(θ

−1 |Z)

∂θs∂θm

)∆s∆m∆l

∆j

︸ ︷︷ ︸O∆,Z(c2)

. (2.36)

37

Page 46: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Here, the point of evaluation, θ, is suppressed in the random ‘big-O’ term, O∆,Z(c2)

as done in previous section, θ±1 denote some points on the line segments between θ

and θ ± c∆ and, the superscript, ± sign, in θ±1 , corresponds to whether θ + c∆ or

θ − c∆ is being considered for the argument of gi(·) above. The expectation of J(g)ij

follows straight from (2.36) by using condition C.1,

E[J(g)ij |θ] = E [Hij(θ|Z)|θ] + O(c2), (2.37)

yielding the expectation of H(g)ij as,

E[H(g)ij |θ] =

1

2

(E[J

(g)ij |θ] + E[J

(g)ji |θ]

)

= −Fij(θ) + O(c2), (2.38)

in which the Hessian-based definition, E [Hij(θ|Z)|θ] = −Fij(θ), and the symmetry,

Fij(θ) = Fji(θ), of FIM are used.

Next, the variance of H(g)ij is considered and it is given by,

var[H(g)ij |θ] =

1

4var[(J

(g)ij + J

(g)ji )|θ]

=1

4

(var[J

(g)ij |θ] + var[J

(g)ji |θ] + 2cov[J

(g)ij , J

(g)ji |θ]

). (2.39)

The expression of the variance term and the covariance term in (2.39) will now be

determined in succession.

By the use of (2.36), var[J(g)ij |θ] in (2.39) is given by,

var[J(g)ij |θ] = var

[( ∑

l

Hil(θ|Z)∆l

∆j

+ O∆,Z(c2)

)∣∣∣∣θ]

= var

[∑

l

Hil(θ|Z)∆l

∆j

∣∣∣∣∣ θ

]+ O(c2), (2.40)

38

Page 47: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

in which var[O∆,Z(c2)|θ], which is a O(c4) term by C.4, is absorbed in the 2cov[∑

l Hil(

θ|Z)(∆l/∆j), O∆,Z(c2)|θ] that is a O(c2) term (by C.4). Further, var[J(g)ij |θ] can be

written as,

var[J(g)ij |θ] = E

(∑

l

Hil(θ|Z)∆l

∆j

)2∣∣∣∣∣∣θ

(E

[∑

l

Hil(θ|Z)∆l

∆j

∣∣∣∣∣ θ

])2

+O(c2)

=∑

l1,l2

E [{Hil1(θ|Z)Hil2(θ|Z)}|θ] E

[∆l1∆l2

∆2j

]− (E [Hij(θ|Z)|θ])2

+O(c2)

=∑

l

E[{Hil(θ|Z)}2

∣∣ θ]E

[∆2

l

∆2j

]− (E [Hij(θ|Z)|θ])2 + O(c2)

=∑

l

{var [Hil(θ|Z)|θ] + (E [Hil(θ|Z)|θ])2}E

[∆2

l

∆2j

]

− (E [Hij(θ|Z)|θ])2 + O(c2)

=∑

l

bl(j)var [Hil(θ|Z)|θ]+∑

l 6=j

bl(j) (E [Hil(θ|Z)|θ])2 + O(c2),(2.41)

in which bl(j) = E[∆2l /∆

2j ]. The first term after the second equality follows from the

fact that ∆ and Z are statistically independent of each other and the second term has

already been seen in the derivation of E[J(g)ij |θ] in (2.37). The third equality follows

by C.1 and I.

By using an identical argument to the one that is mentioned right after (2.40),

the expression of cov[J(g)ij , J

(g)ji |θ], j 6= i, can be readily shown to be given by,

cov[J(g)ij , J

(g)ji |θ] = cov

[(∑

l

Hil(θ|Z)∆l

∆j

),

(∑

l

Hjl(θ|Z)∆l

∆i

)∣∣∣∣∣ θ

]+ O(c2). (2.42)

39

Page 48: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Simplification of the first term in the right-hand-side of (2.42) results in,

cov

[(∑

l

Hil(θ|Z)∆l

∆j

),

(∑

l

Hjl(θ|Z)∆l

∆i

)∣∣∣∣∣ θ

]

= E

[(∑

l

Hil(θ|Z)∆l

∆j

)(∑

l

Hjl(θ|Z)∆l

∆i

)∣∣∣∣∣ θ

]− E [Hij(θ|Z)|θ] E [Hji(θ|Z)|θ]

=∑

l1,l2

E [{Hil1(θ|Z)Hjl2(θ|Z)}|θ] E

[∆l1∆l2

∆j∆i

]− (Fij(θ))2

= E [Hij(θ|Z)Hji(θ|Z)|θ] + E [Hii(θ|Z)Hjj(θ|Z)|θ]− (Fij(θ))2

= E[{Hij(θ|Z)}2

∣∣ θ]+ E [Hii(θ|Z)Hjj(θ|Z)|θ]− (Fij(θ))2

={var [Hij(θ|Z)|θ] + (E [Hij(θ|Z)|θ])2} + E [Hii(θ|Z)Hjj(θ|Z)|θ]

− (Fij(θ))2 (2.43)

Here, the first term after the second equality follows from the fact that ∆ and Z are

statistically independent of each other and the second term after the second equality

follows from the fact that the Hessian matrix is symmetric in the context of FIM and

by the Hessian-based definition of FIM. The third equality follows by the use of the

fact that j 6= i and by C.1 and I. Therefore, cov[J(g)ij , J

(g)ji |θ] turns out to be,

cov[J(g)ij , J

(g)ji |θ] =

{var [(Hij(θ|Z))|θ] + (E [ (Hij(θ|Z))|θ])2}

+ E [Hii(θ|Z)Hjj(θ|Z)|θ]− (Fij(θ))2 + O(c2), j 6= i. (2.44)

Now, the variance of H(g)ij , var[H

(g)ij |θ], for j 6= i, can be readily obtained from (2.39)

by using (2.41) and (2.44). Note that var[H(g)ii |θ] is same as var[J

(g)ii |θ] that can

be directly obtained from (2.41) by replacing j with i. In the next section, the

contributions of the variance and covariance terms to var[H(g)ij |θ] would be compared

40

Page 49: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

with the contributions of the respective variance and covariance terms (as appeared

in (2.39)) to var[H(g)ij |θ] to prove that var[H

(g)ij |θ] is less than var[H

(g)ij |θ].

Mean and Variance of H(g)ij

It is convenient to express J(g)ij as below,

J(g)ij = J

(g)ij −

l∈Ii(−Fil(θ))

∆l

∆j

, ∀j ∈ Ici , (2.45)

where it can be readily shown that the (i, j)-th element of Hk associated with (2.4)

is given by,

H(g)ij = H

(g)ij − 1

2

l∈Ii(−Fil(θ))

∆l

∆j

+∑

l∈Ij(−Fjl(θ))

∆l

∆i

, ∀j ∈ Ici

=1

2(J

(g)ij + J

(g)ji )− 1

2

l∈Ii(−Fil(θ))

∆l

∆j

+∑

l∈Ij(−Fjl(θ))

∆l

∆i

, ∀j ∈ Ici

=1

2(J

(g)ij + J

(g)ji ), ∀j ∈ Ici (2.46)

and

H(g)ij = −Fij(θ), ∀j ∈ Ii. (2.47)

As shown for H(L)ij earlier in Case 1, it can shown in a similar way that ∀i =

1, · · · , p,

E[H(g)ij |θ] =

−Fij(θ) + O(c2), ∀j ∈ Ici ,

−Fij(θ), ∀j ∈ Ii.(2.48)

Again, the expressions of the ‘big-O’ terms both in (2.38) and in the first equation of

(2.48) are precisely same implying that E[H(g)ij |θ] = E[H

(g)ij |θ], ∀j ∈ Ici .

41

Page 50: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Next to compute var[H(g)ij |θ], ∀j ∈ Ici , J

(g)ij in (2.36) is decomposed into two parts

in a similar fashion as J(L)ij was decomposed for Case 1. Subsequent approximation

of the unknown elements of H(θ|Z) (particularly the ones that correspond to the

elements of the FIM that are known a priori) in the resulting expression of J(g)ij by

the negative of those elements of the FIM yields the following,

J(g)ij ≡ J

(g)ij −

l∈Ii(−Fil(θ))

∆l

∆j

=∑

l

Xil(θ|Z)∆l

∆j

+ O∆,Z(c2) ∀j ∈ Ici , (2.49)

in which the equality follows, as just described, by the use of (2.24) and (2.28).

Therefore, by the similarity between (2.36) and (2.49) and by the use of (2.28) and

(2.25)-(2.26), var[J(g)ij |θ], ∀j ∈ Ici , ∀i = 1, · · · , p, follows readily from (2.41) by replac-

ing Hlm with Xlm as reported below,

var[J(g)ij |θ] =

l

bl(j)var [Hil(θ|Z)|θ]+∑l∈Ic

il6=j

bl(j) (E [Hil(θ|Z)|θ])2 + O(c2). (2.50)

Subtracting (2.50) from (2.41), it can be readily seen that ∀j ∈ Ici ,

var[J(g)ij |θ]− var[J

(g)ij |θ] =

l∈Iibl(j) (E [Hil(θ|Z)|θ])2 + O(c2)

=∑

l∈Iibl(j) (Fil(θ))2 + O(c2) > 0. (2.51)

Here, the last inequality follows from the fact that bl(j) = E[∆2l /∆

2j ] > 0, l = 1, · · · , p,

for any given (i, j) and assuming that at least one of the known elements, Fil(θ), in

(2.51) is not equal to zero. Note that bl(j) turns out to be unity if ∆1, · · · , ∆p is

assumed to be Bernoulli ±1 i.i.d. random variables and, as always, the bias term,

42

Page 51: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

O(c2), can be made negligibly small by selecting the user-controlled variable c small

enough.

As remarked earlier in the context of Case 1 (p. 34), it can be noted that var[H(g)ii |θ]

< var[H(g)ii |θ] if j = i ∈ Ici , since var[H

(g)ii |θ] and var[H

(g)ii |θ] are precisely and respec-

tively given by var[J(g)ii |θ] and var[J

(g)ii |θ] and, var[J

(g)ii |θ] < var[J

(g)ii |θ] by (2.51).

Now, since variance of H(g)ij in (2.46) is given by,

var[H(g)ij |θ] =

1

4

(var[J

(g)ij |θ] + var[J

(g)ji |θ] + 2cov[J

(g)ij , J

(g)ji |θ]

), (2.52)

the covariance term, cov[J(g)ij , J

(g)ji |θ], j 6= i, ∀j ∈ Ici , is computed and compared

to cov[J(g)ij , J

(g)ji |θ] next to conclude that var[H

(g)ij |θ] < var[H

(g)ij |θ], ∀j ∈ Ici . The

covariance term, cov[J(g)ij , J

(g)ji |θ], j 6= i, ∀j ∈ Ici , follows readily by substituting Hlm

with Xlm and using the identical arguments employed in deducing (2.42) and (2.43)

yielding a similar expression to (2.44), for j 6= i, ∀j ∈ Ici , i = 1, · · · , p,

cov[J(g)ij , J

(g)ji |θ] =

{var [(Xij(θ|Z))|θ] + (E [ (Xij(θ|Z))|θ])2}

+ E [Xii(θ|Z)Xjj(θ|Z)|θ]− (Fij(θ))2 + O(c2)

={var [(Hij(θ|Z))|θ] + (E [ (Hij(θ|Z))|θ])2}

+ E [Xii(θ|Z)Xjj(θ|Z)|θ]− (Fij(θ))2 + O(c2). (2.53)

The second equality follows from the definition of Xij in (2.28) by particularly noting

the fact that j ∈ Ici . As already indicated in the context of cov[J(L)ij , J

(L)ji |θ] for Case 1

(see after (2.33), p. 35), Xii and Xjj must take the form from one of following four

possibilities: (1) eii and ejj, (2) eii and Hjj, (3) Hii and ejj, (4) Hii and Hjj. It is also

43

Page 52: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

shown there that E [Hii(θ|Z)Hjj(θ|Z)|θ]−E [Xii(θ|Z)Xjj(θ|Z)|θ] is Fii(θ)Fjj(θ) >

0 for the first, second and third possibilities and 0 for the fourth possibility. Since the

difference between cov[J(g)ij , J

(g)ji |θ] in (2.44) and cov[J

(g)ij , J

(g)ji |θ] in (2.53) is given by,

cov[J(g)ij , J

(g)ji |θ]− cov[J

(g)ij , J

(g)ji |θ]

= E [Hii(θ|Z)Hjj(θ|Z)|θ]−E [Xii(θ|Z)Xjj(θ|Z)|θ]+O(c2), for j 6= i, ∀j ∈ Ici ,(2.54)

it can be immediately concluded from (2.39) and (2.52) that

var[H(g)ij |θ] < var[H

(g)ij |θ], i, j = 1, · · · , p,

by using (2.51) and the fact that E[Hii(θ|Z)Hjj(θ|Z)|θ]−E[Xii(θ|Z) Xjj(θ|Z)|θ] ≥ 0,

that appeared in (2.54), and also noting that the bias term, O(c2), can be made

negligibly small by selecting c small enough. Note that the ranges of i and j are

mentioned here as {1, · · · , p} not just j ∈ Ici . This is simply because for the other

case, j ∈ Ii, var[H(g)ij |θ] = 0, by (2.47), and therefore, var[H

(g)ij |θ] < var[H

(g)ij |θ],

j ∈ Ii.

As a special remark it should be observed here that the several derivations for

Case 2 — the case when direct measurements of g are available — as described in

this section remain still valid even if the condition on Hlm,s in C.4 is relaxed.

Finally, since Hk, k = 1, · · · , N , are statistically independent of each other and

Hk, k = 1, · · · , N , are also statistically independent of each other, it can be concluded

straightway that (i, j)-th element of Fn =−(1/N)∑N

k=1Hk and Fn =−(1/N)∑N

k=1Hk

44

Page 53: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

(with appropriate superscript, (L) or (g)) satisfy the following relation,

var[Fij|θ] =var[Hij|θ]

N< var[Fij|θ] =

var[Hij|θ]

N, i, j = 1, · · · , p. (2.55)

Therefore, we conclude this section by stating that the better estimator, Fn, (vis-

a-vis the current estimator, Fn) as determined by using the modified resampling

algorithm would preserve the exact analytically known elements of Fn(θ) as well as

reduce the variances of the estimators of the unknown elements of Fn(θ).

Next chapter presents a few examples illustrating the effectiveness of the modified

resampling algorithm.

45

Page 54: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Chapter 3

Numerical Illustrations and

Discussions

In this chapter, two example problems are presented. A problem with a scalar-

valued random variable is considered in the first example and its multivariate version

is considered in the second example.

3.1 Example 1

Consider independently distributed scalar-valued random data zi with zi ∼ N(µ, σ2

+ciα), i = 1, · · · , n, in which µ and (σ2 + ciα) are, respectively, mean and variance of

zi with ci being some known nonnegative constants and α > 0. Here, θ is considered

as θ = [µ, σ2, α]T . This is a simple extension of an example problem already con-

sidered in literature [12, Example 13.7]. The analytical FIM, Fn(θ), can be readily

46

Page 55: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

determined for this case so that the MC resampling-based estimates of Fn(θ) can be

verified with the analytical FIM. It can be shown that the analytical FIM is given by,

Fn(θ) =

F11 0 0

0 F22 F33

0 F33 F33

,

in which F11 =∑n

i=1(σ2+ciα)−1, F22 = (1/2)

∑ni=1(σ

2+ciα)−2 and F33 = (1/2)∑n

i=1 ci

(σ2 + ciα)−2. Here, the value of θ, that is used to generate the pseudo data vector

(as a proxy for Zn = [z1, · · · , zn]T ) and to evaluate Fn(θ), is assumed to correspond

to µ = 0, σ2 = 1 and α = 1. The values of ci across i are chosen between 0 and

1 which are generated by using MATLAB uniform random number generator, rand,

with a given seed (rand(‘state’,0)). Based on n = 30 yields a positive definite

Fn(θ) whose eigenvalues are given by 0.5696, 8.6925 and 20.7496.

To illustrate the effectiveness of the modified MC based resampling algorithm, it

is assumed here that only the upper-left 2 × 2 block of the analytical FIM is known

a priori. Using this known information, both the existing [13] and the modified

resampling algorithm as proposed in this work are employed to estimate the FIM.

The number of pseudo data vectors generated for both the algorithms is considered

as N = 2000. For Hessian estimation per (1.2), c is considered as 0.0001 and, for

gradient-approximation per (1.3), c is considered as 0.00011. Bernoulli ±1 random

variable components are considered to generate both the perturbation vectors, ∆k

and ∆k.

47

Page 56: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Discussions on results are presented after the description of Example 2.

3.2 Example 2

Consider independently distributed random data vector Z i with Z i ∼ N(µ,Σ +

Pi), i = 1, · · · , n, in which µ and (Σ + Pi) are, respectively, mean vector and covari-

ance matrix of Z i with Pi being a known matrix such that (Σ+Pi) is a positive def-

inite matrix. Here, θ is considered as the elements of µ and the functionally indepen-

dent elements of Σ (i.e., elements of the upper or lower part of the symmetric matrix Σ

including the diagonal elements). Hence, if Z i is a M -dimensional random data vector

and, µ is given by µ = [µ1, · · · , µM ]T and (i, j)-th element of Σ is given by Σij, i, j =

1, · · · ,M , then θ is considered as θ = [µ1, · · · , µM , Σ11, Σ12, Σ22, · · · , Σ1M , · · · , ΣMM ]T

with p = dim(θ) = M + M(M + 1)/2. This is a multivariate version of the previous

example and have already been considered in literature [13].

It can be readily shown that the log-likelihood function, L, for this example is

given by,

L(θ|Zn) = −nM

2ln(2π)− 1

2

n∑i

ln det (Σ + Pi)

− 1

2

n∑i=1

(Z i − µ)T (Σ + Pi)−1 (Z i − µ) .

Now, using a simplified notation for derivative, let us denote the derivative of a matrix

48

Page 57: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

C(θ), whose elements depend on θ, with respect to the l-th element, θl, of θ by,

C(θ),l ≡ ∂C(θ)

∂θl

.

Then, the l-th element of the gradient vector, g, is given by [15], l = 1, · · · , p,

gl(θ|Zn) = −1

2

n∑i=1

tr[(Σ + Pi)

−1 (Σ + Pi),l − (Σ + Pi)−1 (Σ + Pi),l (Σ + Pi)

−1

{(Z i − µ) (Z i − µ)T

}+ (Σ + Pi)

−1{

(Z i − µ) (Z i − µ)T}

,l

],

in which tr represents the trace of a matrix. Finally, the p × p analytical FIM,

Fn(θ), can also be determined and can be compared to the MC resampling-based

estimates of Fn(θ). The (l,m)-th element of the analytical FIM, Fn(θ), is given by

[15], l, m = 1, · · · , p,

Flm(θ) =1

2

n∑i=1

tr[A

(i)l A(i)

m + (Σ + Pi)−1 Mlm

], (3.1)

in which A(i)l ≡ (Σ + Pi)

−1 (Σ + Pi),l and Mlm ≡ µ,lµT,m + µ,mµT

,l .

In the current example, the dimension of the random data vector, Z i, is considered

to be M = 4 so that p = 4+4(4+1)/2 = 14. The value of θ, that is used to generate

the pseudo data vector (as a proxy for Zn = [Z1, · · · ,Zn]T ) and to evaluate Fn(θ),

corresponds to µ = [0, 0, 0, 0]T , Σ being a 4 × 4 matrix with 1 on the diagonal and

0.5 on the off-diagonals and Pi being a 4× 4 matrix of zero elements for all i. Using

n = 30, (3.1) then yields a positive definite Fn(θ) with smallest eigenvalue 3.8207

and largest eigenvalue 120.0000.

To illustrate the effectiveness of the modified MC based resampling algorithm for

this example, it is assumed here that only the upper-left 4 × 4 block of the FIM

49

Page 58: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

associated with the elements of µ and, all the diagonal elements of the rest of FIM

(associated with the functionally independent elements of Σ, i.e., associated with

the elements of the upper triangular part of Σ) are known a priori. The struc-

ture of the FIM in terms of the assumed known elements is shown schematically in

Figure 3.1. Using this known information, both the existing [13] and the modified

1 5 10 14

1

5

10

14

Known elements = 26

Figure 3.1: F(given)30 for Example 2 with known elements as marked; void part consists

of unknown elements to be estimated by using MC based resampling algorithms.

resampling algorithm as proposed in this work are employed to estimate the FIM.

The number of pseudo data vectors generated for both the algorithms is considered

as N = 40000. For Hessian estimation per (1.2), c is considered as 0.0001 and, for

gradient-approximation per (1.3), c is considered as 0.00011. Bernoulli ±1 random

variable components are considered to generate both the perturbation vectors, ∆k

and ∆k.

It should be remarked here that the benefit of the recursive representation of

sample mean of the Hessian estimates as mentioned at p. 9 could not be realized for

50

Page 59: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Example 1 since the computation cost involved in its FIM estimation was negligible.

However, use of the standard recursive representation of sample mean of the (random)

elements of Hk (with appropriate superscripts, (L) or (g)) have enormously reduced

the computation time for this example problem. Though in practice computation of

sample variance is usually not required, sample variance of the (random) elements of

Hk are computed to compare with that of Hk in order to investigate the effectiveness

of the modified resampling algorithm. If var[H(θ|Zn)|θ] represents a matrix whose

(l,m)-th element is given by var[Hlm(θ|Zn)|θ], then the recursive representation of

estimate of var[H(θ|Zn)|θ] is given by,

[varHk] =

k − 1

k[varHk−1

] +1

k − 1

{(Hk − Hk

)◦(Hk − Hk

)}, k = 2, · · · , N,

(3.2)

in which at the first step (k = 1), [varH1 ] in the left-hand-side is set to a matrix

of zeros (i.e., the variance computation essentially starts from k = 2), the symbol

◦ represents the element-wise product operator (also known as Hadamard product

operator) and Hk is computed by using (1.4). At the end of the N -th step, [varHN] is

normalized by (N−1) by setting [varHN] = (N/(N−1))[varHN

] to obtain the unbiased

estimator of var[H(θ|Zn)|θ]. An identical expression is also used for Hk. Note that

the (l,m)-th element of [varHk] and [varHk

] are, respectively, same as var[Hlm|θ] and

var[Hlm|θ] (see (2.55)).

Below the discussions on results for both the examples as described above are

presented.

51

Page 60: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

3.3 Discussions on results

The FIM is estimated based on both the log-likelihood measurements (Case 1)

and the gradient vector measurements (Case 2). The results for Example 1 and

Example 2 are, respectively, tabulated in Table 3.1 and Table 3.3. The absolute

mean-squared error (MSE) of Fn and Fn as presented in the third and fifth columns

of these tables are computed, for example, in the case of Fn in the third column,

as MSE(Fn) =∑ij

(Fij − Fij(θ))2. The relative MSE as presented in the second

and fourth columns are computed, for example, in the case of Fn, as relMSE(Fn) =

100×MSE(Fn)/∑ij

(Fij(θ))2. The effectiveness of the modified resampling algorithm

can be clearly seen from the sixth column of the these tables that show substantial

MSE reduction. The relative MSE reduction in these tables are computed as 100 ×

(MSE(Fn) −MSE(Fn))/MSE(Fn). In these columns also shown within parentheses

are variance reduction. The relative variance reduction are computed as 100× (A−

B)/A, in which A =∑ij

var[Fij|θ] and B =∑ij

var[Fij|θ].

It would also be interesting to investigate the effect of the modified resampling

algorithm on the MSE reduction in the estimators of the unknown elements of the

FIM in contrast to a rather ‘naive approach’ in which the estimates of the unknown

elements are simply extracted from Fn. To see the improvement in terms of MSE

reduction of the estimators of the unknown elements of the FIM, the elements corre-

sponding to the upper-left 2 × 2 block for Example 1 and the elements as shown in

Figure 3.1 for Example 2 of Fn obtained from the current resampling algorithm are

52

Page 61: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

replaced by the corresponding known analytical elements of the FIM, Fn(θ). Clearly,

the contribution of MSE of the elements of Fn corresponding to the known elements

of the FIM are now zero to MSE(Fn) like in the context of Fn in which the contribu-

tion of MSE of the elements of Fn corresponding to the known elements of the FIM

are zero to MSE(Fn). Therefore, the results as shown in Table 3.2 and Table 3.4 only

display the contributions of the MSE from the estimators of the unknown elements

of FIM. The relative MSE reduction in the last columns of these tables are reported

as earlier by showing 100 × (MSE(Fn) −MSE(Fn))/MSE(Fn). These tables clearly

reflect the superiority of the modified resampling algorithm as presented in this work

over the current resampling algorithm. In these tables, similar results on variance are

also reported within parentheses.

Let us now focus more carefully on the numerical values shown in these tables.

For example, consider Case 2 in Table 3.2 in which the MSE reduction of 0.7930 % is

much less than the variance reduction of 94.4222 %. It is known [12, Section 13.1.1]

that the MSE can be decomposed into two parts – one due to the variance and the

other being contributed from bias. The fact that the 94.4222 % variance reduction

is associated with a very small MSE reduction of 0.7930 % (from 0.0885 to 0.0878

only) clearly indicates that the effect of bias to the MSE is relatively much more

than the effect of variance to the MSE. Notwithstanding the fact that this significant

observation is against the backdrop of very small MSE relative to the true FIM with

relMSE(Fn) = 0.0175 % and relMSE(Fn) = 0.0173 %, it worths a mention. The

53

Page 62: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Error in FIM estimates MSE(variance)

Cases relMSE(Fn) MSE(Fn) relMSE(Fn) MSE(Fn) reduction

Case 1 0.3815 % 1.9318 0.0033 % 0.0169 99.1239 %(97.7817 %)

Case 2 0.0533 % 0.2703 0.0198 % 0.1005 62.8420 %(97.5856 %)

Table 3.1: Example 1 – MSE and MSE reduction of FIM estimates based on N = 2000pseudo data (results on variance are reported within parentheses).

other numbers in Table 3.2 and Table 3.4 can be interpreted in a similar fashion.

Table 3.1 to Table 3.4 essentially highlight the substantial improvement of the

results (in the sense of MSE reduction as well as variance reduction) of the modified

MC based resampling algorithm over the results of the current MC based resampling

algorithm. Of course, this degree of improvement is controlled by the values of the

known elements of the analytical FIM; see (2.32) and (2.34) for Case 1 and (2.51)

and (2.54) for Case 2. Finally, to validate the correctness of simulation results as

presented above, theoretical difference between the variance of estimators of the un-

known elements of the FIM for Example 1 as computed by using (2.32), (2.34), (2.51),

(2.54), (2.13), (2.30), (2.39), (2.52), and (2.55), are presented in Table 3.5 against the

same quantities obtained from simulation.

54

Page 63: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

MSE(Fn) MSE(Fn) MSE(and A) (and B) (variance)

Cases – naive approach reduction

Case 1 0.1288 0.1021 20.7235 %(0.0159) (0.0006) (95.9179 %)

Case 2 0.0885 0.0878 0.7930 %(0.0030) (0.0002) (94.4222 %)

Table 3.2: Example 1 – MSE comparison for Fn and Fn for only the unknown elementsof Fn(θ) according to F

(given)n based on N = 100000 pseudo data (similar results

on variance are reported within parentheses, A =∑p

i=1

∑j∈Ici var[Fij|θ] and B =∑p

i=1

∑j∈Ici var[Fij|θ]).

Error in FIM estimates MSE(variance)

Cases relMSE(Fn) MSE(Fn) relMSE(Fn) MSE(Fn) reduction

Case 1 0.5991 % 430.1066 0.1516 % 108.8217 74.6989 %(84.5615 %)

Case 2 0.3860 % 277.1268 0.0966 % 69.3549 74.9736 %(88.7458 %)

Table 3.3: Example 2 – MSE and MSE reduction of FIM estimates based on N =40000 pseudo data (results on variance are reported within parentheses).

MSE(Fn) MSE(Fn) MSE(and A) (and B) (variance)

Cases – naive approach reduction

Case 1 233.3016 108.8217 53.3558 %(150.0580) (28.0297) (81.3208 %)

Case 2 78.8887 69.3549 12.0851 %(16.7351) (2.0077) (88.0030 %)

Table 3.4: Example 2 – MSE comparison for Fn and Fn for only the unknown el-ements of Fn(θ) according to F

(given)n based on N = 40000 pseudo data (similar

results on variance are reported within parentheses A =∑p

i=1

∑j∈Ici var[Fij|θ] and

B =∑p

i=1

∑j∈Ici var[Fij|θ]).

55

Page 64: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

var[F13|θ]− var[F13|θ] var[F23|θ]− var[F23|θ] var[F33|θ]− var[F33|θ]Cases simulation theory simulation theory simulation theory

Case 1 0.1417 0.1399 0.1320 0.1283 0.2438 0.2434Case 2 0.0629 0.0629 0.0093 0.0103 0 0

Table 3.5: Example 1 – Variance comparison between theoretical results (correct upto a bias term as discussed in Section 2.3) and simulation results (with N = 2000).

56

Page 65: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Chapter 4

Conclusions and Future Research

Analytical evaluation of the Fisher information matrix (FIM), in the general case,

is difficult or next to impossible since it involves computation of first or second order

derivative of log-likelihood function and expectation of a resulting multivariate and

non-linear function. To avoid this formidable task, a simulation-based technique,

resampling algorithm [12, Section 13.3.5], [13], is usually recommended for estimation

of the FIM. In some practical applications [4, 5], there arise cases in which some of

the elements of the FIM are known a priori and the remaining elements unknown

(and need to be estimated). In such situations, a naive approach is usually followed

— by simply extracting the estimates of the unknown elements from the full estimate

(outcome of the existing resampling algorithm) of the FIM — whereby the information

contained in the known elements is not exploited to enhance the statistical qualities

of the estimators of the unknown elements.

57

Page 66: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

4.1 Summary of Contributions Made

In the current work, it is assumed that the measurements of the log-likelihood

function or the gradient vector of log-likelihood function are available. The present

work re-visits the resampling algorithm and computes the variance of the estimator of

an arbitrary element of the FIM. A modification in the existing resampling algorithm

is proposed simultaneously preserving the known elements of the FIM and improving

the statistical characteristics of the estimators of the unknown elements (in the sense

of variance reduction) by utilizing the information available from the known elements.

Numerical examples presented in Chapter 3 showed significant improvement of the

results (in the sense of MSE reduction as well as variance reduction) of the modified

MC based resampling algorithm over the results of the current MC based resampling

algorithm.

4.2 Suggestions for Future Research

Based on the study carried out in this work, the following are a few suggestions

for further research:

1. A way to employ the running feedback (similar to the work presented in [14])

to the averaging process of the resampling algorithm to further enhance the

statistical characteristics of the estimators of the unknown elements of the FIM

may be investigated.

58

Page 67: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

2. Though the prior information available in terms of the known elements of the

FIM are used in the current work to enhance the statistical characteristics

of the estimators of the unknown elements of the FIM, the estimates of the

known elements resulted from the resampling algorithm are not used since these

estimates of the known elements are simply replaced by the analytically known

elements. The issue that need further attention is to see if there is any way to

consider these “wasted” information as well to further enhance the statistical

characteristics of the estimators of the unknown elements of the FIM.

59

Page 68: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Bibliography

[1] P J. Bickel and K A. Doksum. Mathematical Statistics: Basic Ideas and Selected

Topics. Prentice Hall, 2nd edition, 2001.

[2] G. Chryssolouris, M. Lee, and A. Ramsey. Confidence interval prediction for

neural networks models. IEEE Transactions on Neural Networks, 7(1):229–232,

January 1996.

[3] T C. Cover and J A. Thomas. Elements of Information Theory. Wiley-

Interscience, 2006.

[4] S. Das, R. Ghanem, and J C. Spall. Asymptotic sampling distribution for polyno-

mial chaos representation of data: A maximum-entropy and Fisher information

approach. SIAM Journal on Scientific Computing, 2006. submitted.

[5] S. Das, R. Ghanem, and J C. Spall. Asymptotic Sampling Distribution for

Polynomial Chaos Representation of Data: A Maximum Entropy and Fisher

information approach. In Proc. of the 45th IEEE Conference on Decision and

Control, San Diego, CA, USA, Dec 13-15, 2006. CD rom.

60

Page 69: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

[6] B. Efron and D V. Hinkley. Assessing the accuracy of the maximum likelihood

estimator: Observed versus expected fisher information. Biometrika, 65(3):457–

487 (with discussion), December 1978.

[7] J.T.G. Hwang and A.A. Ding. Prediction intervals for artificial neural networks.

J. Amer. Statist. Assoc., 92(438):748–757, June 1997.

[8] H. Jeffreys. An invariant form for the prior probability in estimation problems.

Proceedings of the Royal Society of London. Series A, Mathematical and Physical

Sciences, 186(1007):453–461, September 1946.

[9] D C. Kammer. Sensor set expansion for modal vibration testing. Mechanical

Systems and Signal Processing, 19:700–713, 2005.

[10] J C. Spall. Multivariate stochastic approximation using a simultaneous pertur-

bation gradient approximation. IEEE Trans. Automat. Contr., 37(3):332–341,

1992.

[11] J C. Spall. Adaptive stochastic approximation by the simultaneous perturbation

method. IEEE Trans. Automat. Contr., 45(10):1839–1853, 2000.

[12] J C. Spall. Introduction to Stochastic Search and Optimization: Estimation,

Simulation and Control. Wiley-Interscience, 2003.

[13] J C. Spall. Monte Carlo computation of the Fisher information matrix in non-

61

Page 70: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

standard settings. Journal of Computational and Graphical Statistics, 14(4):889–

909, 2005.

[14] J C. Spall. Feedback and weighting mechanisms for improving Jacobian (Hessian)

estimates in the adaptive simultaneous perturbation algorithm. In Proc. of the

2006 American Control Conference, pages 3086–3091, Minneapolis, Minnesota,

USA, June 14-16, 2006.

[15] M. Tegmark, A N. Taylor, and A.F. Heavens. Karhunen-Loeve eigenvalue prob-

lems in cosmology: How should we tackle large data sets? The Astrophysical

Journal, 480(1):22–35, 1997.

[16] M. Simandl, J. Kralovec, and P. Tichavsky. Filtering, predictive, and smoothing

Cramer-Rao bounds for discrete-time nonlinear dynamic systems. Automatica,

37(11):1703–1716, 2001.

62

Page 71: AN EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX ...€¦ · EFFICIENT CALCULATION OF FISHER INFORMATION MATRIX: MONTE CARLO APPROACH USING PRIOR INFORMATION by Sonjoy Das A

Vita

Sonjoy Das was born on August 22, 1972 in West Bengal, India. He earned a

Bachelor of Engineering degree in Civil Engineering from Bengal Engineering College

(Deemed University), West Bengal, India in 1995 and a Masters of Science (Engi-

neering) degree in Civil (Structure) Engineering from Indian Institute of Science,

Bangalore, India in 2001. He has a total of almost four years of experience in Civil

Engineering industry as an assistant engineer and in Indian Space Research Organi-

sation as a scientist. He started his Doctor of Philosophy program in the Department

of Civil Engineering at The Johns Hopkins University, Baltimore, USA in Septem-

ber, 2002 and, later transferred to the University of Southern California, Los Angeles,

USA in August, 2005 to continue his work. Currently, he is a candidate for the degree

of Doctor of Philosophy in the Department of Civil and Environmental Engineering

at the University of Southern California and, also concurrently pursuing his Masters

of Science in Engineering program in the Department of Applied Mathematics and

Statistics at The Johns Hopkins University.

63