7
A stable differential equation approach for inequality constrained optimization problems q Li Jin School of Mathematics, Physics and Information Science, Zhejiang Ocean University, Zhoushan, Zhejiang 316004, PR China article info Keywords: Inequality constrained optimization Differential equation Asymptotical stability Quadratic convergence Modified barrier function abstract This paper presents a first-order derivatives based and a second-order derivatives based differential equation systems for inequality constrained optimization problems by using the modified barrier function. Under the suitable conditions, we prove the asymptotic sta- bility of the two differential systems and local convergence properties of their Euler dis- crete schemes, including the locally quadratic convergence rate of the discrete algorithm for second-order derivatives based differential equation system. Numerical tests are pre- sented that confirm the robustness and efficiency of the approach. Ó 2008 Elsevier Inc. All rights reserved. 1. Introduction The first differential equation method for solving equality constrained optimization problems was proposed by Arrow and Hurwicz [1]. Fiacco and McCormick [2] used differential equation systems for studying optimality conditions for nonlinear programming problems. A systematic study on differential equation methods for solving nonlinear optimization started from the work of Evtushenko and Zhadan in early 1970s, see for instances Yamadhita [3], Pan [4], Evtushenko [5,6] and Evtushenko and Zhadan [7–9]. The main idea of this type of methods is to construct a system of differential equations such that the equilibrium points of this system coincide with the solutions (or KKT points) to the constrained optimization prob- lem. Along this line, Zhang [10] and Zhang et al. [11] studied a modified Evtushenko–Zhadan systems for solving equality constrained optimization problems, Jin et al. [12] proposed differential equation systems for inequality constrained optimization. This paper constructs a simple differential system for solving inequality constrained optimization problems. In Section 2, the modified barrier function, F(x, y, t), is proposed, based on which the system of differential equations is constructed. It is proved that KKT points of an inequality constrained optimization problem is its asymptotically stable equilibrium points. Euler discrete schemes with constant stepsizes for the differential system is presented and the convergence theorem is dem- onstrated. In Section 3, the second differential system is generated by using the second-order derivatives of problem func- tions. We prove its asymptotic stability and local convergence property of its Euler discrete schemes, including the locally quadratic convergence rate of the discrete algorithm with Armijo line searches. Finally, the algorithm with Armijo line searches and the Runge–Kutta algorithm for the differential system are employed to solve several numerical examples, the numerical results show that the Runge–Kutta method has better stability and higher precision than the Armijo stepsize method. 0096-3003/$ - see front matter Ó 2008 Elsevier Inc. All rights reserved. doi:10.1016/j.amc.2008.09.007 q This material is based upon work funded by Zhejiang Provincial Natural Science Foundation of China under Grant No. Y6080388. E-mail address: [email protected] Applied Mathematics and Computation 206 (2008) 186–192 Contents lists available at ScienceDirect Applied Mathematics and Computation journal homepage: www.elsevier.com/locate/amc

A stable differential equation approach for inequality constrained optimization problems

  • Upload
    li-jin

  • View
    220

  • Download
    0

Embed Size (px)

Citation preview

Page 1: A stable differential equation approach for inequality constrained optimization problems

Applied Mathematics and Computation 206 (2008) 186–192

Contents lists available at ScienceDirect

Applied Mathematics and Computation

journal homepage: www.elsevier .com/ locate /amc

A stable differential equation approach for inequality constrainedoptimization problems q

Li JinSchool of Mathematics, Physics and Information Science, Zhejiang Ocean University, Zhoushan, Zhejiang 316004, PR China

a r t i c l e i n f o

Keywords:Inequality constrained optimizationDifferential equationAsymptotical stabilityQuadratic convergenceModified barrier function

0096-3003/$ - see front matter � 2008 Elsevier Incdoi:10.1016/j.amc.2008.09.007

q This material is based upon work funded by ZheE-mail address: [email protected]

a b s t r a c t

This paper presents a first-order derivatives based and a second-order derivatives baseddifferential equation systems for inequality constrained optimization problems by usingthe modified barrier function. Under the suitable conditions, we prove the asymptotic sta-bility of the two differential systems and local convergence properties of their Euler dis-crete schemes, including the locally quadratic convergence rate of the discrete algorithmfor second-order derivatives based differential equation system. Numerical tests are pre-sented that confirm the robustness and efficiency of the approach.

� 2008 Elsevier Inc. All rights reserved.

1. Introduction

The first differential equation method for solving equality constrained optimization problems was proposed by Arrow andHurwicz [1]. Fiacco and McCormick [2] used differential equation systems for studying optimality conditions for nonlinearprogramming problems. A systematic study on differential equation methods for solving nonlinear optimization startedfrom the work of Evtushenko and Zhadan in early 1970s, see for instances Yamadhita [3], Pan [4], Evtushenko [5,6] andEvtushenko and Zhadan [7–9]. The main idea of this type of methods is to construct a system of differential equations suchthat the equilibrium points of this system coincide with the solutions (or KKT points) to the constrained optimization prob-lem. Along this line, Zhang [10] and Zhang et al. [11] studied a modified Evtushenko–Zhadan systems for solving equalityconstrained optimization problems, Jin et al. [12] proposed differential equation systems for inequality constrainedoptimization.

This paper constructs a simple differential system for solving inequality constrained optimization problems. In Section 2,the modified barrier function, F(x,y, t), is proposed, based on which the system of differential equations is constructed. It isproved that KKT points of an inequality constrained optimization problem is its asymptotically stable equilibrium points.Euler discrete schemes with constant stepsizes for the differential system is presented and the convergence theorem is dem-onstrated. In Section 3, the second differential system is generated by using the second-order derivatives of problem func-tions. We prove its asymptotic stability and local convergence property of its Euler discrete schemes, including the locallyquadratic convergence rate of the discrete algorithm with Armijo line searches. Finally, the algorithm with Armijo linesearches and the Runge–Kutta algorithm for the differential system are employed to solve several numerical examples,the numerical results show that the Runge–Kutta method has better stability and higher precision than the Armijo stepsizemethod.

. All rights reserved.

jiang Provincial Natural Science Foundation of China under Grant No. Y6080388.

Page 2: A stable differential equation approach for inequality constrained optimization problems

L. Jin / Applied Mathematics and Computation 206 (2008) 186–192 187

2. A differential system

Consider the inequality constrained optimization problems

min f ðxÞ;s:t: giðxÞP 0; i ¼ 1; . . . ;m;

ð2:1Þ

where f ; gi : Rn ! R; i ¼ 1; . . . ;m. Here each function is twice continuously differentiable. Let Lðx;lÞ ¼ f ðxÞ þPm

i¼1ligiðxÞ bethe (classical) Lagrangian for problem (2.1) and IðxÞ ¼ fijgiðxÞ ¼ 0; i ¼ 1; . . . ;mg denote the active set of indices at x. Let x� bea local minimizer to problem (2.1) and ðx�;l�Þ be the corresponding KKT point, which satisfies the following conditions:

rxLðx�;l�Þ ¼ rf ðx�Þ �Xm

i¼1

l�irgiðx�Þ ¼ 0;

l� P 0; l�i giðx�Þ ¼ 0; giðx�ÞP 0; i ¼ 1; . . . ;m:

ð2:2Þ

Furthermore, let the Jacobian uniqueness conditions, proposed in [2], hold at ðx�;l�Þ

(1) The multipliers l� > 0; i 2 Iðx�Þ.(2) The gradients rgiðx�Þ; i 2 Iðx�Þ are linearly independent.(3) yTr2

xxLðx�;l�Þy > 0 8 0–y 2 fyjrgiðx�ÞTy ¼ 0; i 2 Iðx�Þg.

To find a solution of problem (2.1), we propose a nonlinear Lagrangian for problem (2.1), in the following form:

Fðx; y; tÞ ¼ f ðxÞ � tXm

i¼1

y2i lgðt�1giðxÞ þ 1Þ ðt > 0Þ;

Let zT ¼ ðxT; yTÞ, by differentiating F(x,y,t) with respect to z, we construct the following system of differential equations:

dzdt¼

�rf ðxÞ þPmi¼1

y2irgiðxÞ

t�1giðxÞþ1

�2ty1 lgðt�1g1ðxÞ þ 1Þ...

�2tym lgðt�1gmðxÞ þ 1Þ

2666664

3777775ðt > 0Þ: ð2:3Þ

The following lemma will be used in the proof of the forthcoming theorem.

Lemma 2.1 [13]. Let A be a n� n symmetrical matrix, B be a r � n matrix, U ¼ ½diagli�mi¼1, where l ¼ ðl1; . . . ;lrÞ > 0. If k > 0

is a scalar and

By ¼ 0 implies hAy; yiP khy; yi:

Then there are scalars k0 > 0 and c 2 ð0; kÞ such that, for any k P k0,

hðAþ kBTUBÞx; xiP chx; xi 8x 2 Rn:

Lemma 2.2. Let ðx�;l�Þ be a KKT point of (2.1), the Jacobian uniqueness conditions hold at ðx�;l�Þ. Then there is y� 2 Rm, suchthat ðy�i Þ

2 ¼ l�i ði ¼ 1; . . . ;mÞ, the following conclusions are satisfied:

(i) Fðx�; y�; tÞ ¼ Lðx�;l�Þ ¼ f ðx�Þ.(ii) rxFðx�; y�; tÞ ¼ rxLðx�;l�Þ ¼ 0.

(iii) r2xxFðx�; y�; tÞ ¼ r2

xxLðx�;l�Þ þ t�1rgðx�ÞTDrgðx�Þ,where D ¼ ½diagðy�i Þ

2=ðt�1giðx�Þ þ 1Þ2�mi¼1;rgðx�Þ ¼ ½rg1ðx�Þ; . . . ;rgrðx�Þ�

T.(iv) There is t0 > 0 and c > 0, for any t 2 ð0; t0�, such that

hr2xxFðx�; y�; tÞw;wiP chw;wi 8w 2 Rn:

Proof. Without loss of generality, we assume that Iðx�Þ ¼ f1; . . . ; rg, where r 6 m. It follows from the KKT conditions that

Fðx�; y�; tÞ ¼ f ðx�Þ � tXm

i¼1

ðy�i Þ2 lgðt�1giðx�Þ þ 1Þ ¼ f ðx�Þ � t

Xr

i¼1

ðy�i Þ2 lgðt�1giðx�Þ þ 1Þ ¼ f ðx�Þ;

where ðy�i Þ2 ¼ l�i ði ¼ 1; . . . ;mÞ. Furthermore, Fðx; y; tÞ is differential at x, then we have

rxFðx�; y�; tÞ ¼ rf ðx�Þ �Xm

i¼1

ðy�i Þ2rgiðx�Þ

t�1giðx�Þ þ 1¼ rf ðx�Þ �

Xm

i¼1

l�irgiðx�Þ ¼ rxLðx�;l�Þ ¼ 0:

Page 3: A stable differential equation approach for inequality constrained optimization problems

188 L. Jin / Applied Mathematics and Computation 206 (2008) 186–192

By calculating, we get

r2xxFðx�; y�; tÞ ¼ r2f ðx�Þ �

Xm

i¼1

ðy�i Þ2r2giðx�Þ

t�1giðx�Þ þ 1þXm

i¼1

ðy�i Þ2rgiðx�Þrgiðx�Þ

T

ðt�1giðx�Þ þ 1Þ2¼ r2

xxLðx�;l�Þ þ t�1rgðx�ÞTDrgðx�Þ;

where D ¼ ½diagl�i �mi¼1. Let A ¼ r2

xxLðx�;l�Þ; B ¼ rgðx�Þ and U ¼ D in Lemma 2.1, using the Jacobian uniqueness conditions,we obtain (iv). The proof is completed. h

Lemma 2.3. Let the Jacobian uniqueness conditions hold at ðx�;l�Þ. Then ðx�; y�Þ is an equilibrium point of system (2.3) if and onlyif ðx�;l�Þ be a KKT point of (2.1), where ðy�i Þ

2 ¼ l�i ; ði ¼ 1; . . . ;mÞ.

Proof. Let ðx�;l�Þ be a Kuhn–Tucker point of (2.1), by introducing ðy�i Þ2 ¼ l�i ði ¼ 1; . . . ;mÞ, we have from Lemma 2.2 that

rxFðx�; y�; tÞ ¼ rxLðx�;l�Þ ¼ 0;

which implies that ðx�; y�Þ is an equilibrium point of (2.2).On the contrary, let ðx�; y�Þ be an equilibrium point of (2.2). Existing l�i ¼ ðy�i Þ

2ði ¼ 1; . . . ;mÞ, we have from Lemma 2.2 andthe Jacobian uniqueness conditions that

rxLðx�;l�Þ ¼ 0; l� P 0; l�i giðx�Þ ¼ 0; giðx�Þ 6 0; i ¼ 1; . . . ;m

implying that ðx�;l�Þ be a KKT point of (2.1). h

Theorem 2.1. Let ðx�;l�Þ be a KKT point of (2.1), the Jacobian uniqueness conditions hold at ðx�;l�Þ. Then the system (2.3) witht > 0 is asymptotically stable at ðx�; y�Þ, where ðy�i Þ

2 ¼ l�i ði ¼ 1; . . . ;mÞ.

Proof. Without loss of generality, we assume that Iðx�Þ ¼ f1; . . . ; rgwith r 6 m. Linearizing system (2.3) in the neighborhoodof z�T ¼ ðx�T; y�TÞ, we obtain

dzdt¼ �Qðz� z�Þ; ð2:4Þ

where

Q ¼Q 1 00 Q 2

� �ð2:5Þ

and Q 1 2 RðnþrÞ�ðnþrÞ and Q2 2 Rðm�rÞ�ðm�rÞ are given by

Q 1 ¼

r2xxFðx�; y�; tÞ �2y�1rg1ðx�Þ; . . . ;�2y�rrgrðx�Þ

2y�1rg1ðx�ÞT

..

.0

2y�rrgrðx�ÞT

2666664

3777775

ð2:6Þ

and

Q2 ¼ 2t½diag lgðt�1giðxÞ þ 1Þ�mi¼rþ1:

The stability of system (2.3) is determined by the properties of the roots of the characteristic equation

detðQ � kInþmÞ ¼ 0; ð2:7Þ

which is equivalent to the following two equations:

jQ 1 � kInþrj ¼ 0; jQ2 � kIm�rj ¼ 0:

The solutions of the second equation are explicitly expressed as

ki ¼ 2t lgðt�1giðxÞ þ 1Þ; nþm� r þ 1 6 i 6 nþm

and from the assumptions we have

�k ¼ minnþm�rþ16i6nþm

ki > 0: ð2:8Þ

Let y denote the conjugate vector of a complex vector y, ReðbÞ denote the real part of the complex number b. Let a be aneigenvalue of Q1 and ðz1; z2ÞT 2 Rn � Rr the corresponding nonzero eigenvector of the matrix Q1. Then

Re ½z1; z2�Q 1z1

z2

� �� �¼ Re a½z1; z2�

z1

z2

� �� �¼ ReðaÞðjz1j2 þ jz2j2Þ: ð2:9Þ

It follows from the definition of Q1 that

ReðaÞðjz1j2 þ jz2j2Þ ¼ Refz1Tr2

xxFðx�; y�; tÞz1g: ð2:10Þ

Page 4: A stable differential equation approach for inequality constrained optimization problems

L. Jin / Applied Mathematics and Computation 206 (2008) 186–192 189

In view of Lemma 2.2, the matrix r2xxFðx�; y�; tÞ is positive definite, we obtain for z1–0 that

Refz1Tr2

xxFðx�; y�; tÞz1g > 0: ð2:11Þ

If z1 ¼ 0, then from the Jacobian uniqueness conditions, yields z2 ¼ 0. This contradicts with ðz1; z2Þ–0. Therefore, by notingthat (2.8) and (2.11), we have that all eigenvalues of �Q have negative real parts. It follows from Lyapunov’s stability the-orem of the first-order approximation that ðx�; y�Þ is a local asymptotically stable equilibrium point of (2.3).

Integrating the system (2.3) by the Euler method, one obtains the iterate process

xkþ1 ¼ xk � hkðrf ðxkÞ �Pmi¼1

ðykiÞ2rgiðxkÞ

t�1giðxkÞþ1Þ ;

ykþ1i ¼ yk

i � 2hkyki ðlg t�1giðxkÞ þ 1Þ; i ¼ 1; . . . ;m;

8><>: ð2:12Þ

where yk ¼ ðyk1; y

k2; . . . ; yk

mÞT and hk is a stepsize. h

The following theorem tells us that the Euler scheme with a constant stepsize is locally convergent.

Theorem 2.2. Let conditions of Theorem 2.1 be satisfied at the point ðx�;l�Þ. Then there exists a �h > 0 such that for any0 < hk <

�h, the iterations defined by (2.12) with t > 0 converge locally to ðx�; y�Þ, where l�i ¼ ðy�i Þ2; i ¼ 1; . . . ;m.

Proof. Let zT ¼ ðxT; yTÞ, since z� is a fixed point of the mapping

FðzÞ ¼ z� t

rf ðxÞ �Pmi¼1

y2irgiðxÞ

t�1giðxÞþ1

2ty1 lgðt�1g1ðxÞ þ 1Þ...

2tym lgðt�1gmðxÞ þ 1Þ

266666664

377777775: ð2:13Þ

The convergence of the iterations (2.12) will be proved if we demonstrate that a �h can be chosen such that the iterationsdefined by

zkþ1 ¼ FðzkÞ

converge to z� whenever z0 is in a neighborhood of z� and 0 < hk <�h. LetrFðzÞ be Jacobian transpose of F(z) and m1; . . . ; mnþm

be the eigenvalues of the matrix rFðz�ÞT with the expression

mj ¼ ð1� tajÞ � iðtbjÞ:

From the proof of Theorem 2.1 one has that aj > 0. The condition jmjj < 1 can be written as

h < 2aj=ða2j þ b2

j Þ; 1 6 j 6 nþm:

Let

�h ¼minf2aj=ða2j þ b2

j Þj1 6 j 6 nþmg;

then the spectral radius ofrFðz�ÞT is strictly less than 1 for h < �h, and the iterations generated by the scheme (2.12) is locallyconvergent to ðx�; y�Þ (see Evtushenko [6]). The proof is completed. h

3. The second differential system

The method in Section 2 can be viewed as a gradient approach, through the system (2.3), for solving problem (2.1). Nowwe discuss the Newton version. The continuous version of Newton method leads to the initial value problem for the follow-ing system of ordinary differential equations:

KðzÞdzdt¼ diag16i6nþmci

�rf ðxÞ þPmi¼1

y2irgiðxÞ

ðt�1giðxÞþ1Þ

�2ty1 lgðt�1g1ðxÞ þ 1Þ

..

.

�2tym lgðt�1gmðxÞ þ 1Þ

2666666664

3777777775; ð3:1Þ

Page 5: A stable differential equation approach for inequality constrained optimization problems

190 L. Jin / Applied Mathematics and Computation 206 (2008) 186–192

where zT ¼ ðxT; yTÞ; ðc1; . . . ; cnþmÞT > 0 is a scaling vector and KðzÞ is the Jacobian matrix of the mapping

/ðzÞ ¼

rf ðxÞ �Pmi¼1

y2irgiðxÞ

ðt�1giðxÞþ1Þ

2ty1 lgðt�1g1ðxÞ þ 1Þ...

2tym lgðt�1gmðxÞ þ 1Þ

266666664

377777775: ð3:2Þ

We can easily get the formula

KðzÞ ¼

r2xxFðx; y; tÞ � 2y1rg1ðxÞ

t�1g1ðxÞþ1� � � � 2ymrgmðxÞ

t�1gmðxÞþ1

2y1rg1ðxÞT

t�1g1ðxÞþ12t lgðt�1g1ðxÞ þ 1Þ

..

. . ..

2ymrgmðxÞT

t�1gmðxÞþ12t lgðt�1gmðxÞ þ 1Þ

26666664

37777775:

Lemma 3.1. Let conditions of Theorem 2.1 be satisfied at the point ðx�;l�Þ. Then, for t > 0, Kðz�Þ is a nonsingular matrix, wherel�i ¼ ðy�i Þ

2ði ¼ 1; . . . ;mÞ and z�T ¼ ðx�T; y�TÞ.

Proof. Let ðx�;l�Þ be a KKT point of (2.1), the Jacobian uniqueness conditions hold at ðx�;l�Þ, and denote Iðx�Þ ¼f1; . . . ; rg; r 6 m. Then we obtain

Kðz�Þ ¼

r2xxFðx�; y�; tÞ �2y�1rg1ðx�Þ � � � �2y�rrgrðx�Þ 0 � � � 0

2y�1rg1ðx�ÞT 0 0

..

. . .. . .

.0

2y�rrgrðx�ÞT 0

0rþ1 0 2t lgðt�1grþ1ðx�Þ þ 1Þ... . .

. . ..

0m 0 2t lgðt�1gmðx�Þ þ 1Þ

266666666666664

377777777777775

:

Since ðx�;l�Þ be a KKT point of (2.1), we have, equivalently, by Lemma 2.2 that the matrix r2xxFðx�; y�; tÞ is positive definite.

Since the Jacobian uniqueness conditions, we have that y�1rg1ðx�Þ; . . . ; y�rrgrðx�Þ are linearly independent. Moreover,giðx�Þ > 0 for all i 2 fr þ 1; . . . ;mg. Hence, we obtain that Kðz�Þ is a nonsingular matrix. h

Theorem 3.1. Let conditions of Theorem 2.1 be satisfied at the point ðx�;l�Þ. Then the system (3.1) with t > 0 is asymptoticallystable at z�, where l�i ¼ ðy�i Þ

2ði ¼ 1; . . . ;mÞ and z�T ¼ ðx�T; y�TÞ.

Proof. From the second-order smoothness of / around z�, we have

/ðzÞ ¼ /ðz�Þ þKðz�ÞdðzÞ þHðdðzÞÞ;

where dðzÞ ¼ z� z�; HðdðzÞÞ ¼ OðkdðzÞk2Þ.Linearizing system (3.1) at the point z�, we obtain

ddðzÞdt¼ �Qðz�ÞdðzÞ; Qðz�Þ ¼ Kðz�Þ�1diag16i6nþmciKðz�Þ:

Matrix Qðz�Þ is similar to matrix diag16i6nþmci; therefore, they have the same eigenvalues ki ¼ ci > 0;1 6 i 6 nþm.According to Lyapunov linearization principle, we have that the equilibrium point z� is asymptotically stable.

Integrating the system (3.1) by the Euler method, one obtains the iterate process

zkþ1 ¼ zk � akdiag16i6nþmciKðzkÞ�1/ðzkÞ: � ð3:3Þ

Theorem 3.2. Let conditions of Theorem 2.1 be satisfied at the point ðx�;l�Þ. Then there exists a �a ¼ 2=max16i6nþmci such that forany 0 < ak < �a, the iterations fzkg defined by (3.3) converge locally to z�.

We adopt the Armijo line search rule to determine step lengths when the Euler method is used. Let EðzÞ ¼ k/ðzÞk2 be themerit function.

Algorithm 3.1

Step 1: Given a 2 ð0;1Þ; ci ¼ 1ð1 6 i 6 nþmÞ;q 2 ð0;1=2Þ; e P 0 and initial point z0 ¼ ðx0; y0Þ; k :¼ 0.

Page 6: A stable differential equation approach for inequality constrained optimization problems

L. Jin / Applied Mathematics and Computation 206 (2008) 186–192 191

Step 2: If EðzkÞ 6 e, then stop; otherwise, computer the search direction

dk ¼ �KðzkÞ�1/ðzkÞ:

Step 3: Computer the approximate iterate point

zkþ1 ¼ zk þ akdk;

where ak ¼ bik , ik is the smallest nonnegative integer i satisfying

Eðzk þ bidkÞ 6 ð1� 2qbiÞEðzkÞ:

Step 4: k :¼ kþ 1, go to Step 2.

Theorem 3.3. Let problem functions of problem (2.1) be twice continuously differentiable. Let fzkg be generated by Algorithm 3.1satisfying that for any cluster point z of this sequence, KðzÞ is nonsingular. Then the iterations fzkg converge to a solution toEðzÞ ¼ 0.

Proof. Obviously, we have that the sequence fEðzkÞg is monotone decreasing and convergent. Let �z be a cluster point of fzkg,then there is fzkj

g, such that

zkj! �z;

where fkjg# f1;2;3; . . .g. From Armijo line search rule, we have that

EðzkjÞ � Eðzkj

þ akj dkjÞP 2qakj Eðzkj

Þ;

EðzkjÞ � Eðzkj

þ akj�1dkjÞ < 2qakj�1Eðzkj

Þ:

Assume that Eð�zÞ–0, then from the first inequality above, we have for j!1, bkj ! 0. So from the second inequality, weobtain

Eðzkjþ akj�1dkj

Þ � EðzkjÞ

akj�1 > �2qEðzkjÞ;

hrEð�zÞ; �diP �2qEð�zÞ; ð3:4Þ

where �d ¼ �Kð�zÞ�1/ð�zÞ. Since

hrEð�zÞ; �di ¼ �h2Kð�zÞ/ð�zÞ;Kð�zÞ�1/ð�zÞi ¼ �2k/ð�zÞk2 ¼ �2Eð�zÞ; ð3:5Þ

by combining (3.4) and (3.5), we obtain that

0 6 ð2� 2qÞEð�zÞ 6 0;

therefore Eð�zÞ ¼ 0, which contradicts with the assumption Eð�zÞ–0. Therefore we must have Eð�zÞ ¼ 0. We conclude that anycluster point of fzkg is the solution of EðzÞ ¼ 0. h

Theorem 3.4. Let conditions of Theorem 2.1 be satisfied at the point ðx�;l�Þ, KðzÞ satisfies a Lipschitz condition in a neighborhoodof z�. Assume that the sequence fzkg generated by Algorithm 3.1 converges to z�, then there is an integer K large enough such thathk ¼ 1 for k > K, the sequence fzkg generated by Algorithm 3.1 with t > 0 converges quadratically to z�, wherel�i ¼ ðy�i Þ

2ði ¼ 1; . . . ;mÞ, z�T ¼ ðx�T; y�TÞ.

Proof. Let HðzÞ ¼ /ðzÞ, then Hðz�Þ ¼ 0nþm. From Lemma 3.1 we have that Kðz�Þ is a nonsingular matrix, then there are con-stants c0 > 0, c1 > 0, c2 > 0, c3 > 0, and a neighborhood Nðz�Þ, such that

kHðzÞk ¼ kHðzÞ � Hðz�ÞkP c0kz� z�k;kHðzÞk ¼ kHðzÞ � Hðz�Þk 6 c1kz� z�k;kKðzÞ �Kðz�Þk 6 c2kx� x�k;kKðzÞ�1k 6 c3; z 2Nðz�Þ

ð3:6Þ

for sufficiently large K, when k P K , we have

kzk � z�k 6 r0; zk 2Nðz�Þ; ð3:7Þ

where r0 ¼ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi1� 2rp

c0=ðc1c3c2ð2þ c3c1ÞÞ, and therefore

kdkk ¼ kKðzkÞ�1HðzkÞk 6 c3c1kzk � z�k: ð3:8Þ

Page 7: A stable differential equation approach for inequality constrained optimization problems

Table 1Numerical results

No. Algorithm IT k/ðzÞk2 Accuracy

1 [15, Tp100] R(3.1) 80 1:156948� 10�6 3:7042� 10�5

n ¼ 7, m ¼ 4 3.1 9 3:246312� 10�6 7:2006� 10�5

2 [15, Tp113] R(3.1) 296 7:028284� 10�7 2:2439� 10�5

n ¼ 10, m ¼ 8 3.1 77 3:861312� 10�6 7:1947� 10�4

3 [15, Tp45] R(3.1) 65 5:610742� 10�9 1:1439� 10�4

n ¼ 5, m ¼ 10 3.1 21 9:996834� 10�9 1:6342� 10�3

n stands for the dimension of variables, m for the number of constraints, IT for the number of iterations and R(3.1) for Runge–Kutta algorithm.

192 L. Jin / Applied Mathematics and Computation 206 (2008) 186–192

We can deduce from (3.6)–(3.8) that

kHðzk þ dkÞk ¼ kHðzk þ dkÞ � HðzkÞ þKðzkÞdkk 6 sup06t61

kðKðzk þ tdkÞ �KðzkÞÞdkk

6 sup06t61

kðKðzk þ tdkÞ �Kðz�ÞÞdkk þ kðKðzkÞ �Kðz�ÞÞdkk 6 2c2kzk � z�kkdkk þ c2kdkk2

6 c1c3c2ð2þ c3c1Þkzk � z�k2; ð3:9Þ

therefore, when k P K ,

EðzkÞ � Eðzk þ dkÞ � 2qEðzkÞ ¼ ð1� 2rÞkHðzkÞk2 � kHðzk þ dkÞk2

P ð1� 2qÞc20kzk � z�k2 � ½c1c3c2ð2þ c3c1Þ�2kzk � z�k4

P ½ð1� 2qÞc20 � ½c1c3c2ð2þ c3c1Þ�2r2

0�kzk � z�k2 P 0; ð3:10Þ

which implies that the step length hk ¼ 1 will satisfy Armijo line search rule when k P K in Algorithm 3.1 and hence Algo-rithm 3.1 becomes the classical Newton method and the rate of convergence is quadratic, see for instance [14]. h

4. Computation results

In this section, we gives three examples to illustrate the theoretical results achieved and the efficiency of Algorithm 3.1 inSection 3. The results are obtained by a preliminary MATLAB implementation of Algorithm 3.1 and the Runge–Kutta algo-rithm for the differential equation system (3.1). To compare their numerical results, we adopted the same initial pointand the parameter t in computing each of the problems. The numerical results given in Table 1 show that Runge–Kutta meth-od has better stability and higher precision than the Armijo stepsize method.

References

[1] K.J. Arrow, L. Hurwicz, Reduction of constrained maxima to saddle point problems, in: J. Neyman (Ed.), Proceedings of the Third Berkeley Symposiumon Mathematical Statistics and Probability, University of California Press, Berkeley, 1956, pp. 1–26.

[2] A.V. Fiacco, G.P. McCormick, Sequential unconstrained minimization techniques, in: Nonlinear Programming, John Wiley, New York, 1968.[3] H. Yamadhita, A differential equation approach to nonlinear programming, Math. Program. 18 (1980) 115–168.[4] P.Q. Pan, New ODE methods for equality constrained optimization (1)-equations, J. Comput. Math. 10 (1992) 77–92.[5] Yu.G. Evtushenko, Two numerical methods of solving nonlinear programming problems, Sov. Math. Dokl. 15 (2) (1974) 420–423.[6] Yu.G. Evtushenko, Numerical Optimization Techniques, Optimization Software, Inc. Publication Division, New York, 1985.[7] Yu.G. Evtushenko, V.G. Zhadan, Barrier-projective methods for nonlinear programming, Comput. Math. Math. Phys. 34 (5) (1994) 579–590.[8] Yu.G. Evtushenko, V.G. Zhandan, Stable barrier-projection and barrier-Newton methods in nonlinear programming, Opt. Software 3 (1994) 237–256.[9] Yu.G. Evtushenko, V.G. Zhandan, Stable barrier-projection and barrier-Newton methods for linear and nonlinear programming, in: E. Spedicato (Ed.),

Algorithms for Continuous Optimization, Kluwer Academic Publishers, 1994, pp. 255–285.[10] L.W. Zhang, A modified version to the differential system of Evtushenko and Zhanda for solving nonlinear programming, in: Ya-Xiang Yuan (Ed.),

Numerical Linear Algebra and Optimization, Science Press, 1999, pp. 161–168.[11] Liwei Zhang, Qin Li, Xin Zhang, Two differential systems for solving nonlinear programming problems, OR Trans. 4 (4) (2000) 33–46.[12] Li Jin, Li-Wei Zhang, Xian-Tao Xiao, Two differential equation systems for inequality constrained optimization, Appl. Math. Comput. 188 (2007) 1334–

1343.[13] R. Cominetti, J.P. Dussault, A stable exponential-algorithm with super-linear convergence, J. Optimiz. Theor. Appl. 2 (83) (1994) 285–309.[14] J.M. Ortega, W.C. Rheinboldt, Iterative Solution of Nonlinear Equations in Several Variables, Academic Press, New York, 1970.[15] W. Hock, K. Schittkowski, Test examples for nonlinear programming codes, Lecture Notes Economics and Mathematical Systems, vol. 187, Springer-

Verlag, Berlin, 1981.