32
AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER ACCURACY FOR STOCHASTIC DIFFERENTIAL EQUATIONS by SALLY THERESA GOODLETT, B.S. A THESIS IN MATHEMATICS Submitted to the Graduate Faculty of Texas Tech University in Partial Fulfillment of the Requirements for the Degree of MASTER OF SCIENCE Approved Accepted May, 1992

AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

  • Upload
    others

  • View
    6

  • Download
    0

Embed Size (px)

Citation preview

Page 1: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER ACCURACY

FOR STOCHASTIC DIFFERENTIAL EQUATIONS

by

SALLY THERESA GOODLETT, B.S.

A THESIS

IN

MATHEMATICS

Submitted to the Graduate Faculty of Texas Tech University in

Partial Fulfillment of the Requirements for

the Degree of

MASTER OF SCIENCE

Approved

Accepted

May, 1992

Page 2: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

ACKNOWLEDGMENTS

I would like to thank the committee members of my thesis for their

assistance. And I would like to thank especially my chairman, Prof. Edward

Allen, for placing his knowledge and expertise at my disposal. His guidance

has been invaluable.

I would also like to thank my family for their assistance and

understanding throughout my years of study at Texas Tech University. Most of

all, I would like to thank my husband, Sean, for his encouragement and support.

n

Page 3: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

TABLE OF CONTENTS

ACKNOWLEDGMENTS ................................................................................................. ii

ABSTRACT ..................................................................................................................... .iv

LIST OF TABLES ............................................................................................................. v

CHAPTER

I. INTRODUCTION ...................................................................................... 1

II. THEORETICAL ANALYSIS ................................................................... 3

Ill. VARIANCE REDUCTION METHODS ................................................ 15

IV. NUMERICAL RESULTS ...................................................................... 19

v. CONCLUSION ........................................................................................ 26

BIBLIOGRAPHY ............................................................................................................. 27

lll

Page 4: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

ABSTRACT

An extrapolated Euler method is developed for numerical solutions of

stochastic differential equations. It is proven that expectations of functions of the

stochastic process and expectations of solutions of systems of stochastic

differential equations are approximated to second-order accuracy using the

extrapolated Euler method. Numerical results support the theoretical analysis.

In addition, a new variance reduction procedure is easily implemented with

Euler's method and is described and tested.

IV

Page 5: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

LIST OF TABLES

4.1 Euler's Method for a Scalar Linear Equation ............................................... 19

4.2 Richardson Extrapolation for a Scalar Linear Equation ............................ 20

4.3 Euler's Method for a Scalar Nonlinear Equation ........................................ 20

4.4 Richardson Extrapolation for a Scalar Nonlinear Equation ...................... 20

4.5 Euler's Method for a Function of a Scalar Linear Equation ...................... 21

4.6 Richardson Extrapolation for a Function of a Scalar Linear Equation .... 21

4. 7 Euler's Method for a Function of a Scalar Nonlinear Equation ................ 22

4.8 Richardson Extrapolation for a Function of a Scalar Nonlinear Equation ............................................................................................................... 22

4.9 Euler's Method for a System of Linear Equations ....................................... 23

4.10 Richardson Extrapolation for a System of Linear Equations .................... 23

4.11 Euler's Method for a System of Nonlinear Equations - Solution of y1 ..................................................................................................................... 24

4.12 Richardson Extrapolation for a System of Nonlinear Equations -Solution of Y1 ..................................................................................................... 24

4.13 Euler's Method for a System of Nonlinear Equations - Solution of Y2 ..................................................................................................................... 25

4.14 Richardson Extrapolation for a System of Nonlinear Equations -Solution of Y2 ..................................................................................................... 25

v

Page 6: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

CHAPTER I

INTRODUCTION

Considered in this thesis are stochastic differential equations of the form

{dy(t) = f(t,y(t))dt + g(t,y(t))dW(t)

y(to) = YO (1·1)

where y(t) is a random variable, f(t,y(t)) and g(t,y(t)) are functions of timet and

the random process y(t), and W(t) is the Wiener process. It is assumed that

functions f and g satisfy the following conditions given by Gard [2] such that (1.1)

has a unique solution: the functions f and g are measurable with respect tot

and x, forte [O,T] and x e 1R; the functions f and g are also be Lipschitz

continuous and exhibit linear growth in x.

It is often the case that one must find numerical solutions of expectations

of functions of y(t). Most numerical methods for solving (1.1) yield

approximations to expectations of functions of only first order. However,

Milshtein [4,5], Klauder-Petersen [3] and Talay [6] have developed methods of

O(h2) where h is the time interval in the numerical scheme. Unfortunately, these

methods are complicated and may involve first and second derivatives of the

functions f and g. Therefore, for some practical problems, these methods may

be very difficult to implement. It should be noted that Wagner's scheme [7] for

avoiding systematic error due to time discretization based on unbiased

estimation of the transition density of the solution process was not considered in

this investigation.

In the present investigation, an extrapolated Euler method is used to

approximate expectations of functions. This method is very simple to

implement, because it involves only the application of Richardson extrapolation

1

Page 7: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

to two approximations (with different step sizes) of Euler's method. In the scalar

case, it is proven that expectations of functions of the random process are

approximated with second-order accuracy using this method. It is also proven

that the extrapolated Euler method approximates expectations of solutions to

systems of stochastic differential equations with second-order accuracy.

Several numerical examples are given which support the theoretical analysis.

Theoretical work concerning the expectations of functions of systems of

stochastic differential equations remains for future work. However, numerical

examples indicate that the accuracy may be O(h2) for expectations of functions

for systems and no counter-examples have been found in this investigation to

indicate order less than two.

To be useful in practical problems, a variance reduction procedure is

implemented to reduce statistical errors. A variance reduction method

developed by Chang [1] appeared to be particularly suited to the incorporation

into the extrapolated Euler method. Chang's method was extended in the

present investigation to reduce further statistical errors. Since this new method

is simple to implement, it is suitable for many problems. Numerical tests

indicate that the statistical variation can be dramatically reduced using the new

variance reduction procedure.

2

Page 8: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

CHAPTER II

THEORETICAL ANALYSIS

In this chapter, theoretical analysis prove that the extrapolated Euler

method gives second-order accurate approximations concerning both scalar

stochastic differential equations and systems of stochastic differential equations.

In the scalar case, this method is shown to provide approximations of

expectations of functions of the random variable y(t) which are accurate to

O(h2). The extrapolated Euler method is also proven to be accurate to O(h2) in

approximating the expectations of solutions to systems. In the following

lemmas and theorems, expectations of stochastic differential equations will be

taken. Thus, it is useful to first evaluate the expectation of the following

stochastic differential equation:

dy = f(t,y(t))clt + g(t,y(t))dW(t). (2.1)

Taking the integral of both sides of (2.1) over the interval [O,h], we obtain

h h h

Jdy = Jt(t,y(t))clt + Jg(t,y(t))dW(t). (2.2) 0 0 0

h

From Gard [2], we know the expectation of the integral Jg(t,y(t))dW(t) is zero. 0

Thus,

E{jdy} = ~;(t,y(t))~} and E{y(h)- y(O)} = E{F(h,y(h))- F(O,y(O)} (2.3)

where F'(t,y(t)) = t(t,y(t)). Dividing (2.3) by h and taking the limit as h approaches

zero, we obtain

lim Ey(h) - Ey(O) lim EF(h,y(h)) - EF(O,y(O)) h~O h = h~O h

which by definition is dEy= Ef(t,y(t))dt.

3

Page 9: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

Now we consider the scalar case. We require the following lemma on local

error.

Lemma 1

Let y and y be one-step and two-step Euler approximations, respectively,

to the exact value y(h). Then, assuming functions f(y(t)) and g(y(t)) are

sufficiently smooth,

a. E[y(h)- (2Y1 -y1)] = O(h3)

b.

c.

d.

Proof:

E[y2(h) - (2~- y~)] = O(h3)

E[y3(h) - (2~- y~)] = O(h3)

E[y4(h) - (2y~- y~)] = O(h3).

First, consider the following scalar equations which are derived using

Ito's formula.

d[y(t)- YO] = f(y(t))dt + g(y(t))dW

d[(y(t)- YO)"]= [n(y(t)- YO)n-1f(y(t)) +~ n(n-1)(y(t)- YO)n-2g2(y(t))]dt +

n(y(t)- YO)n-1g(y(t))dW

Hence,

dE[y(t) -YO] = Ef(y(t))dt and (2.4)

dE[(y(t)- YO)"]= E[n(y(t)- YO)n-1f(y(t)) + ~ n(n- 1 )(y(t)- YO)n-2g2(y(t))]dt. (2.5)

By expanding equations (2.4) and (2.5) into Taylor series about yo, we obtain

dE[y(t) - YO] = E[f(YO) + (y(t) - YO)f'(YO) + ~ (y(t) - Y0)2f"(YO) + ... ]dt ( 2. 6)

dE[(y(t) -YO)"] = E[n(y(t)- YO)n-1 f(YO) + n(y(t)- YO)n f'(YO) +

~ (y(t) - YO)n+ 1 f"(YO) + ~ (n- 1 )(y(t) - YO)n-2 g2(yo) + (2. 7)

~ (n- 1 )(y(t)- yo)"-1 (g2(yo))' + ~ (n- 1 )((y(t)- yo)"(g2(yo))" + ... ]dt.

4

Page 10: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

Let F 1 = E(y(t) - YO), F2 = E(y(t) - Y0)2, F3 = E(y(t) - Y0)3, and F 4 = E(y(t) - Y0)4.

Substituting F1 , F2 , F3 , and F4 into equations (2.6) and (2.7), we obtain

.... dF .... .... .... .... dt (t) = AF(t) + b + O(tt3) with F(O) = 0

where b = [ f 92 o o ]T, F(t) = [ F1 (t) F2(t) F3(t) F4(t) ]T, and A is the matrix

f' 1 f" 2

1 f"' 6

_1 f"" 24

2f + (g2)' 2f' + 1.. (g2)" 2

f" + 1.. (g2)'" 6

1 f"' + _1 (g2)"" 3 24

3g2 3f + 3(g2)' 3f' + ~ (g2)" 2

~ f" + 1.. (g2)"' 2 2

0 6g2 4f + 6(g2)' 4f' + 3(g2)" .... ....

Note that in b and A, all functions are evaluated at t = 0. Now, expanding F(h)

in a Taylor's series about t = 0, we obtain

.... .... dF h2d2'F F(h) = F(O) + hdt (0) +2 dt2 (0) + O(h3).

.... 2 .... Note that ~~ (O) = b, and ~t: (0) =A b. Thus, clearly

1f f' + !...f"g2 2 4

. 1 1 f2 + fgg' + f'g2 + 2 g3g" + 2 g2(g')2 (2.8)

3fg2 + 3g3g• 3g4

By equation (2.8), the moments, F1 , F2 , F3 , and F4, are given to an accuracy of

O(h3). Now, consider the one-step and two-step Euler methods where, for

convenience, Yo = 0.

y 1 = h f(YO) + ...Jh Q(Y0)1lO 1\ h - Til_ Y1 /2 = 2 f(YO) + 'J 2 Q(Y0)1l1

1\ 1\ hA _Th_A Y1 = Y1/2 +2f(Y1/2) + 'J 29(Y1/2)1l2·

(2.9)

(2.1Qa)

(2.10b)

Note that 110. 111, 112 are independent normal random variables with unit

variance and E[11oJ = E[111J = E[112J = o. Now, by expanding functions f(Y1/2) and

5

Page 11: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

Q(Y1/2) in Taylor series about yo, we obtain from (2.1 0)

Ey1 = h f(YO) +f h2 f(YO) f'(YO) +i h2 f"(YO) g2(yo) + O(h3).

By applying Richardson extrapolation, we obtain

E(2Y1- Y1) = h f +i h2f f' +f h2 f"g2 + 0(h3). (2.11)

Similarly, we have

E(2~- y~) = hg2 + h2t2 + h2fgg' +i h2g2(g')2 + ~ h2g3g" + O(h3) (2.12)

E(2¢;- y~) = 3h2fg + ~2g3g• + O(h3) (2.13)

1\4 -4 2 4 3 E(2y1-y1)=3h g +O(h ). (2.14)

In equations (2.11 ), (2.12), (2.13) and (2.14), all the functions are evaluated at

t = 0. From equations (2.8) and (2.11) where t = h and Yo = 0, we see

E[y(h) - (2Y1 - y 1)] = O(h3).

Similarly, we have

E[y2(h) - (2~- y~)] = O(h3)

3 1'3 -3 3 E[y (h)- (2y1- y 1)] = O(h )

.A 1\4 -4 3 E[y ·(h)- (2y1- y 1)] = O(h ).

This completes the proof. Notice that it is only necessary to consider the first

four moments as E[y(t)- yo]n is inherently O(h3) for n > 4. Also, it is

straightforward to show using the same procedure that E[y"(h)- y~] = O(h2)

for n = 1 ,2,3, 4 where y is the one-step Euler method. We are now ready to

show that the local error in approximating EF(y(h)) is of O(h3).

Theorem 1

Let y(t) and y(t) be one-step and two-step Euler approximations,

respectively, to y(t). Then, for a function F(y(t)) where t = h,

E[F(y(h)) - (2 F(y(h))- F(y(h))] = O(h3).

6

Page 12: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

Proof:

First, by expanding function F(y(h)), F(y(h)) and F(y(h)) about yo, we obtain

F(y(h)) = F(yo) + (y(h) - YO) P(YO) + i (y(h) - Y0)2F"(YO) + ...

A A 1 A 2 F(y(h)) = F(yo) + (y(h)- YO) F'(YO) +2 (Y(h)- YO) F"(YO) + ···

F(y(h)) = F(YO) + (y(h)- YO) F'(YO) +i (y(h)- Y0)2P'(YO) + ....

Thus, E[F(y(h))- (2 F(y(h))- F(y(h)) 1 = E[y(h)- (2Y1 - Y1)] F'(YO) +

~ E[(y(h)- Y0)2 - (2(y(h)- Y0)2 - (y(h)- Y0)2)] P'(YO) +

i E[(y(h) - Y0)3 - (2(y(h)- Y0)3 - (y(h) - Y0)3)] P"(YO) +

214 E[(y(h)- Y0)4 - (2(y(h)- Y0)4 - (y(h)- Y0)4)] F"'(YO) + O(h3).

Clearly, by Lemma 1 we know that

E[y(h) - (2y(h)- y(h))] = O(h3)

and E[(y(h) - Y0)2 - (2(y(h)- Y0)2- (y(h) - Y0)2)] = E[(y2(h)- (2~- y~))- 2yo(y(h)- (2y(h)- y(h))] = O(h3).

Similarly,

E[(y(h) - Y0)3 - (2(y(h) - Y0)3 - (y(h) - Y0)3)] = O(h3)

and E[(y(h)- Yo)4 - (2(y(h)- Y0)4 - (y(h)- Y0)4)] = O(h3).

Thus, E[F(y(h)) - (2 F(y(h))- F(y(h))] = O(h3).

This completes the proof. The following lemma is required to show global error.

Lemma 2

Let y andy be one-step and two-step Euler approximations, respectively,

to the exact value y(t) for any t = kh. Then, assuming functions f(y(t)) and g(y(t))

are sufficiently smooth,

a. E[y(t)- (2y(t)- y(t))] = O(h2)

b. E[y2(t)- (2y2(t)- y2(t))] = 0(~)

c. E[y3(t) - (2y3(t)- y3(t))] = O(h2)

d. E[y4(t) - (294(t)- y4(t))] = O(h2).

7

Page 13: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

Proof:

As in Lemma 1, let F (t) be a vector of length four with components ~

Fn(t) = E[y(t)- Yo]" for n = 1 ,2,3,4. ~

Let Gi be the vector with components

~ A -(Gi)n = E[2(Yi- Yo)"- (Yi- Yo)"] for n = 1 ,2,3,4.

Let II • II be the maximum norm, i.e., II vII= 1 ~~ 4 I vi I.

It is known from Lemma 1 that the local error satisfies II F(h) -G1 II = O(h3). The

global error is now considered. It is straightforward from (2.4)- (2.7) to find a ~

function <I> such that

F(ti+1) = F(tj) + h<i>(h.F(tj)) + o(~).

Similarly,

~+ 1 = Gj + h'i'(h,Gj) + O(h3).

Furthermore, by the local error result Lemma 1

II <i>(h,z) - i¥(h,z) II = O(~.

Defining, the local error ei =II F(ti)-Gi II. we have

ei+1 = ei + hll ~h.F(ti))- 'i'(h,Gi) II+ O(h3)

~ ej +hi I ~h.F(tj)) _ljl(h.F(tj)) II+ hllljl(h.F(tj)) _ljl(h,Gj) II+ O(~). ~

Thus, ei+1 ~ ei + hLei + O(h3), assuming the'¥ is Lipschitz continuous in the second

variable with Lipschitz constant L. (This is guaranteed provided that f and g are

sufficiently smooth.) Thus, the above inequality implies that

ei+ 1 ~ O(h2) for any i. We are now ready to prove the extrapolated Euler method

approximates expectations of functions of the random variable y(t) with an

accuracy of O(h2).

8

Page 14: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

Theorem 2

Given the local error, E[F{y{h))- {2F{y{h))- F{y{h))], has an accuracy of O(h3)

where y(h) and y(h) are the one-step and two-step Euler's methods,

respectively, then the global error satisfies

E[F(y(t))- {2F(y{t))- F(y(t))] = O(h2)

where t = hm is fixed time.

Proof:

Define the e(t) as

" -e{t) = E[F(y{t))- (2F(y(t))- F(y(t))].

By expanding the function F(y(t)) in a Taylor series about yo, we obtain

" -e(t) = E[F'(Yo)[y(t) - (2y(t)- y(t))) +

~ F"(Yo)[(y(t) - Yo)2 - (2(y(t) - Yo)2- (y(t) - Yo)2))] + ...

1 1 1 2 = P(yo)e1 (t) +2 F"(yo)e2(t) + 6 F"'(yo)e3(t) + 24 F'"'(Yo)e4(t) + O(h )

where en(t) = E[(y(t)- yo)n- (2(y(t)- y0)n- (y(t)- y0)n)] and the O(h2) term arises due to

errors in moments greater than four which have inherent local errors of O(h3).

The proof is now completed by applying Lemma 2. Note that Theorem 2

applies only to scalar equations. We now prove that the extrapolated Euler

method approximates expectations of solutions of systems of stochastic

differential equations with an accuracy of O(h2). First the following lemma on

local error is required.

Lemma 3 1\ -Consider a system of stochastic differential equations. Let ya(t) and ya(t)

be one-step and two-step Euler approximations, respectively, to ya(t) where

9

Page 15: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

1 ::;; a and J3::;; K, subject to some initial condition . Then, assuming functions

fa(y(t)) and gp(Y(t)) are sufficiently smooth,

a. E[ya(h} - (2ya(h} - ya(h))) = O(h3)

b. E[ya(h)y~(h)- (2Rh)y~(h)- ya(h)y~(h))] = 0(~).

Proof:

The integral of the stochastic differential equation

dya(t) = fa(y(t))dt + gp(y(t))dW~(t)

has the formal solution, fort = h, h h

ya(h) = yg + Jta(y'Y(s))ds + J 9p(Y'Y(s))T)~(s)ds. 0 0

Klauder-Petersen [3] show this solution has deterministic part, D(ya(h)), and a

stochastic part, S(ya(h)) where D(ya(h)) and ES(ya(h)) are

D(y~h)) = yg + h fa+~ h2t ~.~ + 0(~) (2.15)

where t.~ = ~ta(y'Y)I~y~ I y 'Y = yJ

d (l 1 2 Jl v (l 0 3 an ES(y (h))= 4" h gtgtf,Jlv + (h ). (2.16)

Klauder-Petersen [3] also show that the covariance of the stochastic part of the

formal solution has the form

E[S(ya(h)S(y~(h)] = hgpg~ +~ h2~~f~ +~ h2g~~g~~ +

(2.17)

It is now shown that the Euler approximations have the identical deterministic

and stochastic parts (to O(h3)) hence implying (a) and (b). Expanding ya(h) and

10

Page 16: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

ya(h) in Taylor series about yg we obtain

/\-.. a 1 2 R.a ~a R ~ a R h R a II 't yu(h) = y + h fa +- h f 1-'T r:t + - g 11'"' + - a.; 11'"' +- g'"' g R 11,..11 + 0 4 .... 2 p 0 2 -p 1 2 ll 't,p 0 1

_1_ h3/2gPf ~n! + _1_ h3/2fP~ R'Jl 't + _1_ ~/2gP gil Q; r:t n!Tl all 't + 2...J2 'Y •1-' 'U 2...[2 •1-' 1 4...[2 y a •I-'ll \J 0 1

1 2 pva 'Y.A 1 2 pllAa "(.acj>'t 8 h g'YgAf,pyllollo + 24 h 9y9a9cp9't,PilA 11Q1lo1lo1l1 +

~ h2 g~Ag~pATlcill~ + 0(~)

and ya(h) = yg + h fa+ ...ft. Q~Tl~ + 0(~)

where E[TlO] = E[111J = E[1121 = o.

Thus,

(2.18)

(2.19)

(2.20)

and E[2ya(h)yA(h)- ya(h)yA(h)] = D*(y(h),y(h),a,A) + S*(y(h),y(h),a,A) + O(h3) (2.21)

where .. ,.. - a A A a + a A 1 2 A r:t.a 1 2 A p v a

D (y(h),y(h),a,A)= Yo y0 + h y0 f hy0 f + 2 h y0f 1-'l,p +4 h y0g'Yg'Yf ,pv +

1 2 a r:t.A 1 2 a p v a ._~ 2 h Yo f 1-'l,p +4 h Yo 9y9yf ,pv + O(n-)

d - * ,.. - a A 1 2 A p a 1 2, . .A.p a an ES (y(h),y(h),a,A) = hQp9p +2 h 9y9yf ,p +2 h ~ 9y,p +

1 2 paa A 1 2 Aalla 1 2 apA 2 h 9iJ y9v,p9v,a + 4 h 9y9 p9p9y,CJJJ. + 2 h 9y 9/ ,p +

1 2apA 1 2aallA 3 2 h 9y f 9y,p + 4 h 9y 9p9p9y,CJJJ. + O(h ).

Thus from (2.20) and letting ya(h) = 2ya(h)- ya(h), we obtain

D(ya(h)) = yg + h fa+~ h2f ~.~ + 0(~)

and

1 1

(2.22)

(2.23)

Page 17: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

From equations (2.15), (2.16), (2.17), (2.20), (2.21 ),(2.22) and (2.23) we obtain

D(y~h))- D(ya(h)) = O(h3),

ES(ya(h)) - ES(ya(h)) = O(h3),

and E[S(ya(h)S(y~(h)) - ES*(y(h),y(h),a,A.) = O(h3).

Hence, for ea(t) = ya(t)- ya(t),

Eea(t+h) = ea(t) + hE(D(ya(t))- D(ya(t))) + 0(h3).

This completes the proof. We are now ready to show the global error in

approximating expectations of solutions of systems of stochastic differential

equations is O(h2).

Theorem 3 A ...

Consider a system of stochastic differential equations. Let y<X(t) and ya(t)

be one-step and two-step Euler approximations, respectively, to ya(t) where

1 ~a and~~ K, subject to some initial condition. Then, assuming functions

fa(y(t)) and g~(y(t)) are sufficiently smooth, the global error satisfies

E[ya(t) - (2ya(t)- ya(t))) = 0(h2)

where t = km is a fixed time.

Proof:

First, we will define the notation for the cumulants. Let

« ya ,, = E[ya]

« yay A. , = E[yayA.].

Furthermore, let

za(ti+ 1) = 2ya(ti+ 1) - ya(ti+ 1)

vaA.(ti+ 1) = 2ya(ti+ 1 )yA(t;+ 1) - ya(ti+ 1 )yA(ti+ 1).

12

Page 18: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

For the analytic solution, we can find functions .11 and .12 such that

a a « Yi+ 1 ,, = cc Yj ,, + hA1 (h,yi)

(2.24)

Similarly, for the approximate solution

a a 11 -« zi+ 1 ,, = « ~ ,, + hel>1 (h,yi.Yi)

aA. a A. 11 -« vi+ 1 ,, = « ~ zi ,, + hel»2(h,yj,yj)

(2.25)

where by Lemma 3 I Cl>1 (h,yj,yj)- .11 (h,yj) I ~ C1 h2,

(2.26)

Let the errors be denoted by e1 and e2. where e1 = cc Zj~1 ,, - '' y~1 ,, and

aA. a A. e2 = « vi+1 ,, - cc Yi+1Yi+1 ». Thus, by subtracting (2.24) from (2.25), we have

(2.27)

e2(ti+ 1) = e2(ti) + h{ Cl>2(h,yi.Yi)- C1>2(h,yj,yi) + Cl>2(h,yj,yi)- .12(h,yi) }.

Analogous to Klauder-Petersen [3], e1>1 and C1>2 are expanded to the second

order in the cumulants and thus obtaining

"" 11 - "" a a aA. a A. "'1 (h,yi.Yi) - "'1 (h,yj,yi) = a1 { cc ~+ 1 , - cc Yi+ 1 , } + a2{ cc '1+ 1, - cc Yi+ 1 Yi+ 1 , } (2.28)

where a1, a2. b1 and b2 are the expansion coefficients. Thus, to this order in the

cumulants from (2.26), (2.27) and (2.28), we find that

le1(ti+1)l ~ le1(ti)l+hla1lle1(ti)l+hla2lle2(ti)l + c1h3,

I e2(ti+1) I ~ I e2(ti) I+ hI b1 II e1 (ti) I+ hI b2ll e2(ti) I + c2h3.

Let II • II be the norm defined as II e II= I e1 I+ I e2l. Letting L be the constant

L = max{ I a1 I+ I b1 I. I a2l +I b2l } and c = c1 + c2. we find

13

Page 19: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

II e(ti+ 1 > II = II e(ti) II + hLII e(ti) II + ctr3.

Equation (2.29) is easily solved for a fixed time t = kh, giving k-1

II e(tk) II~ (1 + hL)II e(O) II+ L(1 + hL)jch3. j=O

Because the initial error, II e(O) 11. is zero, we have

II e(tk) II ~ ( 1 + ~t>k - 1 ch3 ~ eLh~ - 1 ch2.

Thus, the proof is complete.

14

(2.29)

Page 20: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

CHAPTER Ill

VARIANCE REDUCTION METHODS

Through the application of Richardson Extrapolation in conjunction with

Euler's method, we have reduced the deterministic error involved in

approximating stochastic differential equations. The stochastic error can also

be reduced by implementing a variance reduction technique. In this chapter,

we consider different variance reduction techniques for both scalar stochastic

differential equations and systems of stochastic differential equations. We

consider two variance reduction techniques, namely Chang's method [1] and a

new method, applied in unison with Euler's method in order to evaluate the

expectations of functions of scalar stochastic differential equations. We

also show that both methods are easily modified to evaluate expectations of

solutions of systems of stochastic differential equations. Although not

considered here, the methods can be modified to reduce the stochastic error in

evaluating functions of solutions of systems of stochastic differential equations.

First, we describe Chang's variance reduction method.

Chang's Variance Reduction Method

Consider the scalar stochastic differential equation

dy = f(t,y)dt + g(t,y)dW(t).

Recall that Euler's method for solving (3.1) is

Yk+ 1 = Yk + hfk + ..JhgkZk

where E[zk] = o and Var{zk] = 1. Let k = 0. Thus Euler's method for Y1 is

Y1 = YO + hfo + ..Jhgozo.

(3.1)

Define F(y1) to be a function of the stochastic variable Y1· By expanding F(y1) in

15

Page 21: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

a Taylor series about yo, we obtain

F(Y1) = F(yo) + F'(YO)(hfO + ..Jhgozo) + ....

Define F* (Y1) as

F•(Y1) = F(YO) + F'(YO)[hfQ) + ... = F(Y1)- F'(yo)[..Jhgozo).

Notice that E[F.(Y1)] = E[F(y1)] and it is likely that Var[F.(y1)] s Var[F(y1)].

Thus,

F• (Y1) = F(YO) + [F(Y1) - F(yQ) - ..JhF'(YO)QQZQ].

Now consider k = 1. By Euler's method, we obtain

Y2 = Y1 + hf1 + ..Jhg1 Z1

and by expanding F(y2) in a Taylor series about Y1. we obtain

F(Y2) = F(Y1) + F'(Y1)[hf1 + ..Jhg1z1] + ....

Thus,

F•(Y2) = F•(Y1) + [F(y2)- F(Y1)- -ftlF'(Y1)Q1Z1]·

Again, we see that E[F. (Y2H = E[F(y2)] and it is likely that Var[F. (Y2)] s Var[F(Y2H·

Hence, F.(Yk) can be defined recursively as

F• (y~~ 1) = F• (y~)) + [F(y~~ 1)- F(y~))- ..JhF'(y~))QkZk]

fork= 0,1,2, ... and i = 1, ... , N where N is the number of time intervals. Note that n

EF(y(t)) =~ :2l·(y~ where Mh = t.

i=1

Chang's method can easily be extended to solutions of systems of stochastic

differential equations. Consider the system .. ...llrro.... ..a.. ...

dy = f (Y (t))clt + g(y(t))dW(t)

where y(t) = [Y1 (t), Y2(t), ... , Yn(t)]T, f (y) = [f1 (y),f2(y), ... ,fn(Y)]T and g is an n x n

matrix of functions of y (t). Thus, Chang's variance reduction technique for

systems is

16

Page 22: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

I

where y k = [1, 1, .... , 1 ]T fork= 0,1 ,2, ... and i = 1 , ... , N where N is number of time

intervals.

Now, by modifying Chang's variance reduction method, a new variance

reduction technique is developed which further reduces the variance.

New Variance Reduction Method

Define F(yk) to be a function of the stochastic variable Yk· Let k = 0. Thus

Euler's method for Y1 is

Y1 =YO+ hfo + -Yhgozo.

Expanding F(y1) in a Taylor series about yo, we obtain

F(Y1) = F(yo) + F'(yo)[hfo + -Yhgozo] + i F"(yo)[h2t~ + 2-Yhhtogozo + hg~z~] + ....

Define u(yk) and v(yk) fork= 0 as

1 2 2 2 2 u(yo) = F'(yo)[hfo] + 2 F"(YO)[h t0 + hg0z0]

and v(yo) = F'(yo)[hto] + i F"(YO)[h2t~ + hg~.

Note that v is just a nonstochastic function of YO·

Thus,

F(y1) = F(yo) + F'(yo)[-Yhgozo] + F"(yo)[..Jhhtogozo] + u(yo) + ....

* Now, define F (Y1) as

F* (Y1) = F(Y1) - \l'hF'(YO)QOZO- \l'hhF"(YO)fOQOZO- U(YO) + v(yo).

Note that E[F. (Y1 )] = E[F(Y1 )]. We define v*(Y1) using Chang's method and thus

obtaining

v*(Y1) = v(yo) + [v(y1)- v(yo)- -Yhv'(yo)gozo].

17

Page 23: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

Thus, in general we can define the new method as

v*(ii)) = v*(ii) ) + [v(y(i))- v(y(i) ) - -.Thv '(y(i) )g(i) z(i) ] k k-1 k k-1 k-1 k-1 k-1

F* (y~~ 1) = F* (y~)) + [F(y~~ 1)- F(y~)- -ftiF'(y~))QkZk- -ftihF"(y~))fkQkZk

- u(y~))] + v*(y~))

fork= 0,1 ,2, ... and i = 1 ,2, ... ,N where N is the number of time intervals.

Note that

v(yk) = hF'(yk)fk + ~ F"(Yk)[h2~ + h~

d 1 22 22 an u(yk) = hF'(yk)fk +2 F"(Yk)[h fk + hgkzk].

Thus, v'(yk) = hF'(yk)f'k + F''(Yk)[h2fkf'k + hQkQ'k + hfk] + F"'(Yk)~ hg~ +~ h2f~].

The new variance reduction method is easily extended to solutions of systems

of stochastic differential equations. Consider the system

dy = 1 &(t))dt + g(y(t))dW(t)

where y(t) = [Y1(t), Y2(t), ... , Yn(t)]T, f(y) = [f1(Y).f2(Y) •... ,fn(Y)]T and g is ann x n matrix of

functions of y(t). The new variance reduction method for the solution of systems

of stochastic differential equations is defined as

v*(y .... (i)) = v*(y .... (Q ) + rv ,~(i) ) - -ftlv·(y .... (Q )g(i) z-(i) 1 k k-1 \1 k-1 k-1 k-1 k-1

!Y~+1 J(il = IY~J(Q +{!Yk+1J(il- IYkJ(il- ,,hly~(Qg~Z~ - ii~1l} +V·&~1l

fork= O, 1 ,2, ... and i = 1 ,2, ... ,N where N is the number of time intervals.

Note that

v(yk) = hf k and u(yk) = hf k·

.... .... {51 k 5 f k 5 f k} Thus, v'(yk)=h ~ + ~ + ... + ~. uY1 uY2 oYn

In a similar manner, the variance reduction method can be extended to

functions of solutions of systems of stochastic differential equations.

18

Page 24: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

CHAPTER IV

NUMERICAL RESULTS

We present several numerical examples which indicate O(h2) accuracy

acquired by applying the extrapolated Euler method. Both variance reduction

methods are used to show the decrease of stochastic error when the new

variance reduction is applied in comparison to Chang's variance reduction

method. We first provide numerical examples for the scalar stochastic

differential equations. Linear and nonlinear examples are given for both

approximations of expectations of solutions of stochastic differential equations

and functions of stochastic differential equations. In every example 1 0,000

simulations were performed.

Example 1 : Numerical results for y(t)

{ dy = ~dt + ydW(t)

y(O) = 1

Fort = 1.0, the exact value of Ey(t) is 1.6487213 (see tables 4.1 and 4.2).

Table 4.1: Euler's Method for a Scalar Linear Equation

Chang's Method New Method

#of time Expected Standard Expected Standard

intervals Value Deviation Value Deviation

2 1.55914 0.169478 1.56249 0.006836

4 1.60353 0.285398 1.60204 0.026704

8 1.62238 0.359972 1.62394 0.047025

16 1.63535 0.404022 1.63625 0.060508

19

Page 25: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

Table 4.2: Richardson Extrapolation for a Scalar Linear Equation

Chang's Method New Method #of time Expected Richardson Expected Richardson intervals Value Extrapolation Value Extrapolation

2 1.55914 1.56249 4 1.60353 1.64792 1.60204 1.64159 8 1.62238 1.64123 1.62394 1.64584 16 1.63535 1.64832 1.63625 1.64856

Example 2: Numerical results for y(t)

{ dy = (3y213 + 3y113)dt + 3y213dW(t)

y(O) = 1

Fort= 1.0, the exact value of Ey(t) is 14 (see tables 4.3 and 4.4).

Table 4.3: Euler's Method for a Scalar Nonlinear Equation

Chang's Method New Method

#of time Expected Standard Expected Standard

intervals Value Deviation Value Deviation

2 9.9892 1.8163 9.9915 1.5143

4 12.278 3.7204 12.233 1.4030

8 13.297 5.3767 13.326 1.2796

16 13.672 6.4234 13.676 1.3080

Table 4.4: Richardson Extrapolation for a Scalar Nonlinear Equation

Chang's Method New Method

#of time Expected Richardson Expected Richardson

intervals Value Extrapolation Value Extrapolation

2 9.9892 9.9915

4 12.278 14.664 12.233 14.475

8 13.297 14.316 13.326 14.419

16 13.672 14.047 13.676 14.026

20

Page 26: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

Example 3: Numerical results for F(y(t)) = y3(t)

{ dy = ~dt + ydW(t)

y(O) = 1

Fort= 1.0, the exact value of EF(y(t)) = Ey3(t) is

90.017131 (see tables 4.5 and 4.6).

Table 4.5: Euler's Method for a Function of a Scalar Unear Equation

Chang's Method New Method

#of time Expected Standard Expected Standard

intervals Value Deviation Value Deviation

2 14.3271 28.2689 14.4303 15.2022

4 25.1840 121.2213 25.5036 56.6832

8 40.2521 177.2859 40.8489 110.7710

16 51.0455 302.9312 54.6746 192.0087

32 70.3139 666.4928 72.4737 422.3107

Table 4.6: Richardson Extrapolation for a Function of a Scalar Unear Equation

Chang's Method New Method

#of time Expected Richardson Expected Richardson

intervals Value Extrapolation Value Extrapolation

2 14.3271 14.4303

4 25.1840 36.0409 25.5036 36.5769

8 40.2521 55.3202 40.8489 56.1942

16 51.0455 61.8389 54.6746 68.5003

32 70.3139 89.5823 72.4737 90.2728

21

Page 27: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

Example 4: Numerical results for F(y(t)) = y2(t) where

{ dy = (3y213 + 3y113)dt + 3y213dW(t)

y(O) = 1

Fort= 1.0, the exact value of EF(y(t)) = Ey2(t) is

499 (see tables 4. 7 and 4.8).

Table 4.7: Euler's Method for a Function of a Scalar Nonlinear Equation

Chanq's Method New Method

#of time Expected Standard Expected Standard

intervals Value Deviation Value Deviation

2 146.1652 178.1462 146.0280 74.0173

4 264.2405 466.5475 266.4706 293.3166

8 356.7456 759.6386 360.3555 544.7186

16 419.3520 1021.878 424.4428 777.2277

32 462.9625 1452.823 461.8799 1160.892

Table 4.8: Richardson Extrapolation for a Function of a Scalar Nonlinear

Equation

Chang's Method New Method

#of time Expected Richardson Expected Richardson

intervals Value Extrapolation Value Extrapolation

2 146.1652 146.0280

4 264.2405 382.3158 266.4706 386.9132

8 356.7456 449.2507 360.3555 454.2404

16 419.3520 481.9584 424.4428 488.5301

32 462.9625 506.5730 461.8799 499.3170

We now present numerical examples for systems of stochastic differential

equations. A linear and a nonlinear example are given for the approximation of

expectations of the solutions of systems.

22

Page 28: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

Example 5: Numerical results for Y1 and Y2

#of time

intervals

2

4

8

16

{dY1 = f1dl + Y2dW1 (I)

dy2 = 2)'2dt + Y1 dW2(t)

Y1(0) = 1

Y2(0) = 1

Fort = 1.0, the exact value of Ey1 (t) = Ey2(t) is

1.6487213 (see tables 4.9 and 4.1 0).

Table 4.9: Euler's Method for a System of Linear Equations

Chang's Method for Y1 and Y2 New Method for Y1 and Y2

Expected Standard Expected Standard Value Deviation Value Deviation

1.5626 0.17768 1.5625 0.012010

1.5986 0.28276 1.6014 0.026085

1.6245 0.35612 1.6240 0.046799

1.6374 0.40269 1.6362 0.060350

Table 4.10: Richardson Extrapolation for a System of Linear Equations

Chang's Method for Y1 and Y2 New Method for Y1 and Y2

#of time Expected Richardson Expected Richardson

intervals Value Extrapolation Value Extrapolation

2 1.5626 1.5625

4 1.5986 1.6346 1.6014 1.6403

8 1.6245 1.6504 1.6240 1.6466

16 1.6374 1.6503 1.6362 1.6484

23

Page 29: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

Example 6: Numerical results for Y1 and Y2

{

2 1~~ dy1 = y2dt + (y2e-Y1 - 1 )dW1 (t) + 4V y2 + 1 dW2(t)

dy2 = -~ Y2dt + ..J 2y1 + 6 dW1(t)

Y1 (0) = 1

Y2(0) = 1

Fort= 1.0, the exact value of Ey1 (t) is 1.8504061 (see tables 4.11 and 4.12)

and of Ey2(t) = 0.6065307 (see tables 4.13 and 4.14).

Table 4.11: Euler's Method for a System of Nonlinear Equations­Solution of Y1

Chang's Method for Y1 New Method for Y1

#of time Expected Standard Expected Standard

intervals Value Deviation Value Deviation

2 1.9244 0.20994 1.9236 0.12129

4 1.8906 0.24314 1.8911 0.22051

8 1.8780 0.26035 1.8763 0.25978

16 1.8684 0.28427 1.8663 0.26479

Table 4.12: Richardson Extrapolation for a System of Nonlinear Equations -Solution of Y1

Chang's Method for Y1 New Method for Y1

#of time Expected Richardson Expected Richardson

intervals Value Extrapolation Value Extrapolation

2 1.9244 1.9236

4 1.8906 1.8568 1.8911 1.8586

8 1.8780 1.8654 1.8763 1.8615

16 1.8684 1.8588 1.8663 1.8563

24

Page 30: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

Table 4.13: Euler's Method for a System of Nonlinear Equations­Solution of Y2

Chang's Method for Y2 New Method for Y2

#of time Expected Standard Expected Standard intervals Value Deviation Value Deviation

2 0.56189 0.17764 0.56247 0.00622 4 0.58691 0.21029 0.58610 0.02367 8 0.59738 0.22604 0.59676 0.03583 16 0.60197 0.22993 0.60147 0.04169

Table 4.14: Richardson Extrapolation for a System of Nonlinear Equations­Solution of Y2

Chang's Method for Y2 New Method for Y2

#of time Expected Richardson Expected Richardson

intervals Value Extrapolation Value Extrapolation

2 0.56189 0.56247

4 0.58691 0.61193 0.58610 0.60973

8 0.59738 0.60785 0.59676 0.60742

16 0.60197 0.60656 0.60147 0.60618

The numerical examples given show that Euler's method is a first-order

method. When Richardson extrapolation is applied, the accuracy of the method

increases to second-order. This increase in accuracy is consistent regardless

to the variance reduction technique used.

25

Page 31: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

CHAPTERV

CONCLUSION

The extrapolated Euler method has been proven to approximate the

expectations of functions of scalar stochastic differential equations and of

solutions of systems of stochastic differential equations with second-order

accuracy. Second-order accurate results were found regardless to which

variance reduction technique, Chang's method or the new method, was

implemented. It should be noted, however, that because the new variance

reduction technique reduces more stochastic error, better results were found

when the new variance reduction technique was used.

Theoretic work concerning the expectations of functions of systems of

stochastic differential equations remains for future work. Numerical examples

indicate that accuracy may be O(h2) for expectations of functions for systems

and no counter-examples were found to indicate order less than two.

26

Page 32: AN EXTRAPOLATED EULER METHOD OF SECOND-ORDER …

BIBLIOGRAPHY

[1] Chien-Cheng Chang, Numerical Solution of Stochastic Differential Equations with Constant Diffusion Coefficients, Mathematics of Computation, 49, 523-542, 1987.

[2] Thomas C. Gard, Introduction to Stochastic Differential Equations, Marcel Dekker, Inc., New York, 1988.

[3] John R. Klauder and Wesley P. Petersen, Numerical integration of multiplicative-noise stochastic differential equations, SIAM Journal on Numerical Analysis, 22, 1153-1166, 1985.

[4] G. N. Milshtein, Approximate integration of stochastic differential equations, Theory of Probability and its Applications, 19, 557-562, 1974.

[5] G. N. Milshtein, A method of second-order accuracy integration of stochastic differential equations, Theory of Probability and its Applications, 23, 396-401' 1978.

[6] D. Talay, Efficient numerical schemes for the approximation of expectations of functionals of the solution of a stochastic differential equation, and applications, Lecture Notes in Control and Information Science, 61, 294-313, 1984.

[7] Wolfgang Wagner, Monte Carlo evaluation of functionals of solutions of stochastic differential equations. Variance reduction and numerical examples, Stochastic Analysis and Applications, 6, 447-468, 1988.

27