Upload
li-jin
View
214
Download
1
Embed Size (px)
Citation preview
Applied Mathematics and Computation 188 (2007) 1334–1343
www.elsevier.com/locate/amc
Two differential equation systems for inequalityconstrained optimization q
Li Jin a,b,*, Li-Wei Zhang a, Xian-Tao Xiao a
a Department of Applied Mathematics, Dalian University of Technology, Dalian 116024, Liaoning Province, Chinab Department of Mathematics, Anshan Normal University, Anshan 114005, Liaoning Province, China
Abstract
This paper presents two differential systems, involving first and second order derivatives of problem functions, respec-tively, for solving inequality constrained optimization problems. Local minimizers to the optimization problems areproved to be asymptotically stable equilibrium points of the two differential systems. First, the Euler discrete schemes withconstant stepsizes for the two differential systems are presented and their convergence theorems are demonstrated. Second,an Euler discrete scheme of the second differential system with an Armijo line search rule, is proposed and proved to havethe locally quadratic convergence rate. The numerical results based on solving the second system of differential equationsshow that the Runge–Kutta method for the second system has good stability and the Euler discrete scheme the Armijo linesearch rule has fast convergence.� 2006 Elsevier Inc. All rights reserved.
Keywords: Inequality-constrained optimization; Constraint qualification; Differential equation; Asymptotical stability; Equilibrium point
1. Introduction
Consider the inequality-constrained optimization problems
0096-3
doi:10
q ThScienti
* CoProvin
E-m
min f ðxÞs:t: giðxÞ 6 0; i ¼ 1; . . . ;m;
ð1:1Þ
where f : Rn ! R; gi : Rn ! R, i = 1, . . . ,m twice continuously differentiable. Differential equation methods,for solving equality constrained optimization problems, have been proposed by Arrow and Hurwicz [1],
003/$ - see front matter � 2006 Elsevier Inc. All rights reserved.
.1016/j.amc.2006.10.085
e research is supported by the National Natural Science Foundation of China under project Grant No. 10471015 and by thefic Research Foundation for the Returned Overseas Chinese Scholars, State Education Ministry.rresponding author. Address: Department of Applied Mathematics, Dalian University of Technology, Dalian 116024, Liaoningce, China.ail addresses: [email protected] (L. Jin), [email protected] (L.-W. Zhang).
L. Jin et al. / Applied Mathematics and Computation 188 (2007) 1334–1343 1335
Fiacco and Mccormick [2] and Evtushenko [3], and developed by Yamadhita [4], Pan [5] and Evtushenko andZhadan [6–10]. The main idea of this type of methods is to construct a system of differential equations so thatits equilibrium points of this system coincide with the solutions to the equality constrained problem. Evtush-enko and Zhadan have studied, by using the stability theory of differential equations, differential equation ap-proaches for optimization problems over closed sets with equality constraints. This class of problems includeinequality constrained problems as special cases. They convert the concerned problem to an equality con-strained one by using the so-called space transformations, and then construct a system of differential equationsto solve the resulted equality constrained problem. For equality constrained problems, the systems defined by[4] and Evtushenko and Zhadan [9] are quite similar, and both are of the form
BðxÞdx=dt ¼ �ðrf ðxÞ þ rhðxÞuðxÞÞ;dhðxÞ=dt ¼ �shðxÞ; ð1:2Þ
where BðxÞ 2 Rn;n is symmetrically positive definite, h : Rn ! Rl, $h(x) is the Jacobian transpose of h at x ands > 0 is a constant. Actually, if we choose s � 1, then the system in [4] is obtained and if B(x) � In·n, then theabove system becomes the one of Evtushenko and Zhadan [9]. Along this line, Zhang [11] studied a modifiedversion to (1.2) for solving equality constrained optimization problems.
In this paper, we consider a direct way for constructing a system of differential equations to solve inequalityconstrained problems without using space transformations of Evtushenko and Zhadan [9]. In Section 2, wepropose an nonlinear Lagrangian, Fr(x,y), and construct a system of differential equations by using this func-tion, which does not need a space transformation. We prove that the system is asymptotically stable at localminimum points of the problems. In Section 3, a differential system which involve the second order derivativesof problem functions is constructed and the Euler discrete scheme for the differential system with constantstepsizes is proved to be convergent. Furthermore, an Euler discrete scheme with Armijo line search for thesecond system is proposed and its global convergence and the locally quadratic convergence rate are demon-strated. In Section 4, the Runge–Kutta method and the Euler discrete scheme with Armijo line search areimplemented to solve the second differential equation system. The numerical results obtained show thatRunge–Kutta method has good stability and high precision and the Euler discrete scheme with Armijo linesearch has the fast convergence.
2. The first differential system
Let Lðx; lÞ ¼ f ðxÞ þPm
i¼1ligiðxÞ be the (classical) Lagrangian for problem (1.1) and I(x) = {ijgi(x) = 0,i = 1, . . . ,m} denote the active set of indices at x. Let x* be a local minimizer to problem (1.1) and (x*,l*)be the corresponding Kuhn–Tucker point, which satisfies the following conditions:
rxLðx�; l�Þ ¼ rf ðx�Þ þXm
i¼1
l�irgiðx�Þ ¼ 0;
l� P 0; l�i giðx�Þ ¼ 0; giðx�Þ 6 0; i ¼ 1; . . . ;m: ð2:1Þ
Furthermore, let the Jacobian uniqueness conditions, proposed in [2], hold at (x*,l*):
(1) The multipliers l* > 0, i 2 I (x*).(2) The gradients $gi(x*), i 2 I (x*) are linearly independent.(3) yTr2
xxLðx�; l�Þy > 0 80 6¼ y 2 fyjrgiðx�ÞTy ¼ 0; i 2 Iðx�Þg.
To obtain the numerical solution, we propose a nonlinear Lagrangian for problem (1.1), in the following form:
F rðx; yÞ ¼ f ðxÞ þ rXm
i¼1
y2i ðer�1giðxÞ � 1Þ:
Let zT = (xT,yT). By differentiating Fr(x,y) with respect to x and y, respectively, we construct the followingsystem of differential equations:
1336 L. Jin et al. / Applied Mathematics and Computation 188 (2007) 1334–1343
dz=dt ¼
�rf ðxÞ �Pmi¼1
y2i er�1giðxÞrgiðxÞ
2ry1ðer�1g1ðxÞ � 1Þ2ry2ðer�1g2ðxÞ � 1Þ
..
.
2rymðer�1gmðxÞ � 1Þ
26666666664
37777777775: ð2:2Þ
The following lemma will be used in the proof of the forthcoming theorem.
Lemma 2.1 [12]. Let A be a n · n symmetrical matrix, B be a r · n matrix, U ¼ ½diag li�ri¼1 where
l = (l1, . . . ,lr) > 0. If k > 0 is a scalar and
By ¼ 0 implies hAy; yiP khy; yi
then there are scalars k0 > 0 and c 2 (0,k) such that, for any k P k0,
hðAþ kBTUBÞx; xiP chx; xi 8x 2 Rn:
Lemma 2.2. Let (x*,l*) be a Kuhn–Tucker point of (1.1), the Jacobian uniqueness conditions hold at (x*, l*),
then there is y� 2 Rm, such that ðy�i Þ2 ¼ l�i ði ¼ 1; . . . ;mÞ, the following conclusions are satisfied:
(i) Fr(x*, y*) = L(x*,l*) = f(x*);
(ii) $xFr(x*, y*) = $xL(x*,l*) = 0;
(iii) r2xxF rðx�; y�Þ ¼ r2
xxLðx�; l�Þ þ r�1rgðx�ÞT Drgðx�Þ; where D ¼ ½diagl�i �mi¼1;
(iv) there is r0 > 0 and c > 0, for any r 2 (0,r0], such that
hr2xxF rðx�; y�Þw;wiP chw;wi 8w 2 Rn:
Proof. Without loss of generality, we assume that I(x*) = {1, . . . , r} where r 6 m. It follows from the definitionof a Kuhn–Tucker point that
F rðx�; y�Þ ¼ f ðx�Þ þ rXm
i¼1
ðy�i Þ2ðer�1giðx�Þ � 1Þ
¼ f ðx�Þ þ rXr
i¼1
ðy�i Þ2ðer�1giðx�Þ � 1Þ ¼ f ðx�Þ:
Let ðy�i Þ2 ¼ l�i ði ¼ 1; . . . ;mÞ, we have from the definition of a Kuhn–Tucker point that
rxF rðx�; y�Þ ¼ rf ðx�Þ þXm
i¼1
ðy�i Þ2er�1giðx�Þrgiðx�Þ ¼ rf ðx�Þ þ
Xr
i¼1
l�irgiðx�Þ ¼ rxLðx�; l�Þ ¼ 0:
By calculating, we have that
r2xxF rðx�; y�Þ ¼ r2
xxf ðx�Þ þXm
i¼1
ðy�i Þ2er�1giðx�Þr2
xxgiðx�Þ þXm
i¼1
ðy�i Þ2er�1giðx�Þrgiðx�Þrgiðx�Þ
T
¼ r2xxLðx�; l�Þ þ r�1rgðx�ÞTDrgðx�Þ;
where D ¼ ½diagl�i �mi¼1. Let A ¼ r2
xxLðx�; l�Þ; B ¼ ½rg1ðx�Þ; . . . ; rgrðx�Þ�T and U = D in Lemma 2.1, using the
Jacobian uniqueness conditions, we obtain (iv). The proof is completed. h
Lemma 2.3. Let the Jacobian uniqueness conditions hold at (x*,l*). Then (x*, y*) is an equilibrium point of system
(2.2) if and only if (x*,l*) be a Kuhn–Tucker pair of (1.1) with ðy�i Þ2 ¼ l�i ði ¼ 1; . . . ;mÞ.
L. Jin et al. / Applied Mathematics and Computation 188 (2007) 1334–1343 1337
Proof. Let (x*,l*) be a Kuhn–Tucker pair of (1.1). By introducing y�i ¼ffiffiffiffiffil�ip ði ¼ 1; . . . ;mÞ, we have from
Lemma 2.2 that
rxF rðx�; y�Þ ¼ rxLðx�; l�Þ ¼ 0;
which implies that (x*,y*) is an equilibrium point of (2.2).On the contrary, let (x*,y*) be an equilibrium point of (2.2). Defining l�i ¼ ðy�i Þ
2 ði ¼ 1; . . . ;mÞ, we havefrom Lemma 2.2 and the Jacobian uniqueness conditions that
rxLðx�; l�Þ ¼ 0; l� P 0; l�i giðx�Þ ¼ 0; giðx�Þ 6 0; i ¼ 1; . . . ;m
implying that (x*,l*) be a Kuhn–Tucker point of (1.1).Now we prove the system (2.2) is asymptotically stable and the Euler iterative scheme with constant
stepsizes is locally convergent. h
Theorem 2.1. Let(x*,l*) be a Kuhn–Tucker point of (1.1), the Jacobian uniqueness conditions hold at (x*,l*).
Then there exists k0 > 0 such that for k P k0,the system with r > 0 is asymptotically stable at (x*, y*), where
l�i ¼ ðy�i Þ2 ði ¼ 1; . . . ;mÞ.
Proof. Without loss of generality, we assume that I(x*) = {1, . . . , r} with r 6 m. Linearizing system (2.2) in theneighborhood of z*T = (x*T,y*T), we obtain the equation of the first approximation of (2.2) about the equilib-rium point z*
dz=dt ¼ �Qðz� z�Þ; ð2:3Þ
where Q ¼Q1 0
0 Q2
� �ð2:4Þ
and Q1 2 RðnþrÞ�ðnþrÞ and Q2 2 Rðm�rÞ�ðm�rÞ are given by
Q1 ¼
r2xxF rðx�; y�Þ 2y�1rg1ðx�Þ; . . . ; 2y�rrgrðx�Þ
�2y�1rg1ðx�ÞT
..
.0
�2y�rrgrðx�ÞT
2666664
3777775 ð2:5Þ
and
Q2 ¼ �½diag2rðer�1giðxÞ � 1Þ�mi¼rþ1:
The stability of system (2.2) is determined by the properties of the roots of the characteristic equation
detðQ� kInþmÞ ¼ 0; ð2:6Þ
which is equivalent to the following two equations:jQ1 � kInþrj ¼ 0; jQ2 � kIm�rj ¼ 0:
The solutions of the second equation are explicitly expressed as
ki ¼ �2rðer�1giðxÞ � 1Þ; nþ m� r þ 1 6 i 6 nþ m
and from the assumptions we have
�k ¼ minnþm�rþ16i6nþm
ki > 0: ð2:7Þ
Let y denote the conjugate vector of a complex vector y, Re(b) denote the real part of the complex numberb. Let a be an eigenvalue of Q1 and ðz1; z2ÞT 2 Rn � Rr the corresponding nonzero eigenvector of the matrixQ1. Then
Re ½z1; z2�Q1
z1
z2
� �� �¼ Re a½z1; z2�
z1
z2
� �� �¼ ReðaÞðjz1j2 þ jz2j2Þ: ð2:8Þ
1338 L. Jin et al. / Applied Mathematics and Computation 188 (2007) 1334–1343
It follows from the definition of Q1 that
Re ½z1; z2�Q1
z1
z2
� �� �¼ Re z1
Tr2xxF rðx�; y�Þz1 þ z2
TQ21z1 þ z1TQ12z2
� �; ð2:9Þ
where Q12 ¼ ð2y�1rg1ðx�Þ; . . . ; 2y�rrgrðx�ÞÞ, Q21 ¼ �QT12. Since for any real matrix M,
RefzTMTwg ¼ RefwTMzg; ð2:10Þ
we have from (2.9) and (2.8) that
Re ½z1; z2�Q1
z1
z2
� �� �¼ Refz1
Tr2xxF rðx�; y�Þz1g: ð2:11Þ
It follows from (2.7) and (2.10) that
ReðaÞðjz1j2 þ jz2j2Þ ¼ Refz1Tr2
xxF rðx�; y�Þz1g: ð2:12Þ
In view of Lemma 2.2, the matrix r2xxF rðx�; y�Þ is positive definite, we obtain for z1 5 0 that
Refz1Tr2
xxF rðx�; y�Þz1g > 0: ð2:13Þ
If z1 = 0, then
Q1
z1
z2
� �¼ a
z1
z2
� �
and
ð2y�1rg1ðx�Þ; . . . ; 2y�rrgrðx�ÞÞz2 ¼ 0;
which, from the Jacobian uniqueness conditions, yields z2 = 0. This contradicts with (z1,z2) 5 0. Therefore, bynoting that (2.7) and (2.13), we have that all eigenvalues of Q have negative real parts. It follows from Lyapu-nov’s stability theorem of the first-order approximation that (x*,y*) is a local asymptotically stable equilib-rium point of (2.2).
Integrating the system defined by (2.2) using Euler method, one obtains the iterate process
xkþ1 ¼ xk � tkðrf ðxkÞ þPmi¼1
y2ki
er�1giðxkÞrgiðxkÞÞ;
yðkþ1Þi ¼ ykiþ 2rtkyki
ðer�1giðxkÞ � 1Þ; i ¼ 1; . . . ;m;
8><>: ð2:14Þ
where tk is a stepsize. h
The following theorem tells us that the Euler scheme with constant constants is locally convergent.
Theorem 2.2. Let (x*,l*) be a Kuhn–Tucker pair of (1.1), the Jacobian uniqueness conditions hold at (x*,l*).
Then there exists a �t > 0 such that for any 0 < tk < �t the iterations defined by (2.14) with r > 0 converge locally
to (x*, y*), where l�i ¼ ðy�i Þ2 ði ¼ 1; . . . ;mÞ.
Proof. Let zT = (xT,yT), since z* is a fixed point of the mapping
F ðzÞ ¼ z� t
rf ðxÞ þPmi¼1
y2i er�1giðxÞrgiðxÞ
�2ry1ðer�1g1ðxÞ � 1Þ�2ry2ðer�1g2ðxÞ � 1Þ
..
.
�2rymðer�1gmðxÞ � 1Þ
26666666664
37777777775: ð2:15Þ
L. Jin et al. / Applied Mathematics and Computation 188 (2007) 1334–1343 1339
The convergence of the iterations (2.14) will be proved if we demonstrate that a �t can be chosen such that theiterations defined by
zkþ1 ¼ F ðzkÞ
converge to z* whenever z0 is in a neighborhood of z* and 0 < tk < �t. Let $F(z) be Jacobian transpose of F(z)and m1, . . . ,mn+m be the eigenvalues of the matrix $F(z*)T with the expressionmj ¼ ð1� tajÞ � iðtbjÞ:
From the proof of Theorem 2.1 one has that aj > 0. The condition jmjj < 1 can be written ast < minf2aj=ða2j þ b2
j Þ j 1 6 j 6 nþ mg:
Let
�t ¼ minf2aj=ða2j þ b2
j Þj1 6 j 6 nþ mg
then the spectral radius of $F(z*)T is strictly less than 1 for t < �t, and the iterations generated by the scheme(2.14) is locally convergent to (x*,y*) (see [6]). The proof is completed. h
3. The second differential system
The method in Section 2 can be viewed as a gradient approach, through the system (2.2), for solving prob-lem (1.1). Now we discuss the Newton version. The continuous version of Newton method leads to the initialvalue problem for the following system of ordinary differential equations:
KðzÞdz=dt ¼ diag16i6nþmci
�rf ðxÞ �Pmi¼1
y2i er�1giðxÞrgiðxÞ
2ry1ðer�1g1ðxÞ � 1Þ2ry2ðer�1g2ðxÞ � 1Þ
..
.
2rymðer�1gmðxÞ � 1Þ
26666666664
37777777775; ð3:1Þ
where (c1, . . . ,cn+m) is a scaling vector, K(z) is the Jacobian matrix of the mapping
/ðzÞ ¼
rf ðxÞ þPmi¼1
y2i er�1giðxÞrgiðxÞ
�2ry1ðer�1g1ðxÞ � 1Þ�2ry2ðer�1g2ðxÞ � 1Þ
..
.
�2rymðer�1gmðxÞ � 1Þ
26666666664
37777777775; zT ¼ ðxT; yTÞ: ð3:2Þ
We can easily get the formula
KðzÞ ¼
r2xxF rðx; yÞ 2y1er�1g1ðxÞrg1ðxÞ � � � 2ymer�1gmðxÞrgmðxÞ
�2y1er�1g1ðxÞrg1ðxÞT �2rðer�1g1ðxÞ � 1Þ
..
. . ..
�2ymer�1gmðxÞrgmðxÞT �2rðer�1gmðxÞ � 1Þ
2666664
3777775:
Lemma 3.1. Let conditions of Theorem 2.1 be satisfied at the the point (x*,l*). Then, for r > 0, K(z*) is a
nonsingular matrix, where l�i ¼ ðy�i Þ2 ði ¼ 1; . . . ;mÞ, z* = (x*, y*).
1340 L. Jin et al. / Applied Mathematics and Computation 188 (2007) 1334–1343
Proof. Let (x*,l*) be a Kuhn–Tucker point of (1.1), the Jacobian uniqueness conditions hold at (x*,l*). LetI(x*) = {1, . . . , r}, r 6 m, we obtain
Kðz�Þ ¼
r2xxF rðx�; y�Þ 2y1er�1g1ðx�Þrg1ðx�Þ � � � 2y�r er�1grðx�Þrgrðx�Þ 0 � � � 0
�2y�1er�1g1ðx�Þrg1ðx�ÞT 0 0
..
. . .. . .
.
�2y�r er�1grðx�Þrgrðx�ÞT 0 0
0rþ1 0 2yrþ1er�1g1ðx�Þrg1ðx�Þ
..
. . .. . .
.
0m 0 2y�mer�1gmðx�Þrgmðx�Þ
2666666666666664
3777777777777775
:
Hence the matrix K(z*) coincides with the matrix Q introduced in (2.4). It was shown in the proof of Theorem2.1 that all eigenvalues of this are strictly positive. Therefore, K(z*) is nonsingular. h
Theorem 3.1. Let conditions of Theorem 2.1 be satisfied at the point (x*,l*). Then the system (3.1) with r > 0 is
asymptotically stable at z*, where l�i ¼ ðy�i Þ2 ði ¼ 1; . . . ;mÞ and z*T = (x*T, y*T).
Proof. From the second order smoothness of / around z*, we have
/ðzÞ ¼ /ðz�Þ þ Kðz�ÞdðzÞ þHðdðzÞÞ;
where d(z) = z � z*, H(d(z)) = O(kd(z)k2).Linearizing system (3.1) at the point z*, we obtain
ddðzÞ=dt ¼ �bQðz�ÞdðxÞ; bQðz�Þ ¼ Kðz�Þ�1diag16i6nþmciKðz�Þ:
The stability of this system is determined by the properties of the roots of the characteristic equation
detðbQðz�Þ � kInþmÞ ¼ 0:
Matrix bQðz�Þ is similar to matrix diag16i6n+mci; therefore, they have the same eigenvalues ki = ci > 0,1 6i 6 n + m. According to Lyapunov linearization principle, we have that the equilibrium point z* is asymptot-ically stable.
Integrating the system (3.1) by the Euler method, one obtains the iterate process
zkþ1 ¼ zk � hkdiag16i6nþmciKðzkÞ�1/ðzkÞ: � ð3:3Þ
Theorem 3.2. Let conditions of Theorem 2.1 be satisfied at the the point (x*,l*). Then there exists a �t > 0 such
that for any 0 < tk < �t, the iterations defined by (3.3) with r > 0 converge locally to (x*, y*), where
l�i ¼ ðy�i Þ2 ði ¼ 1; . . . ;mÞ.
We adopt the Armijo line search rule to determine step lengths when the Euler method is used. Let E(z) =
k/(z)k2 be the merit function.
Algorithm 3.1.
Step 1: Given a 2 (0, 1), ci = 1(1 6 i 6 n + m), q 2 (0,1/2), e P 0 and initial point zT0 ¼ ðxT
0 ; yT0 Þ, k :¼ 0.
Step 2: If E(zk) 6 e, then stop; otherwise, computer the search direction
dk ¼ �KðzkÞ�1/ðzkÞ:
Step 3: Computer the approximate iterate pointzkþ1 ¼ zk þ hkdk;
where hk ¼ aik , ik is the smallest nonnegative integer i satisfying
Eðzk þ aidkÞ 6 ð1� 2qaiÞEðzkÞ:
Step 4: k :¼ k + 1, go to Step 2.L. Jin et al. / Applied Mathematics and Computation 188 (2007) 1334–1343 1341
Theorem 3.3. Let problem functions of problem (1.1) be twice continuously differentiable. Let {zk} be generated
by Algorithm 3.1 satisfying that for any cluster point z of this sequence, K(z) is nonsingular. Then the iterations
{zk} converge to a solution to E(z) = 0.
Proof. Obviously, we have that the sequence {E(zk)} is monotone decreasing and convergent. Let �z be a clus-ter point of {zk}, then there is fzkjg, such that
zkj ! �z;
where {kj} � {1,2,3 ,. . .}. From Armijo line search rule, we have that
EðzkjÞ � Eðzkj þ bkj dkjÞP 2rbkj EðzkjÞ;EðzkjÞ � Eðzkj þ bkj�1dkjÞ < 2rbkj�1EðzkjÞ:
Assume that Eð�zÞ 6¼ 0, then from the first inequality above, we have for j!1, bkj ! 0. So from the secondinequality, we obtain
Eðzkj þ bkj�1dkjÞ � EðzkjÞbkj�1
> �2rEðzkjÞ;
hrEð�zÞ; �diP �2rEð�zÞ; ð3:4Þ
where �d ¼ �Kð�zÞ�1/ð�zÞ. Since
hrEð�zÞ; �di ¼ �h2Kð�zÞ/ð�zÞ;Kð�zÞ�1/ð�zÞi ¼ �2k/ð�zÞk2 ¼ �2Eð�zÞ ð3:5Þ
by combining (2.13) and (2.14), we obtain that
0 6 ð2� 2rÞEð�zÞ 6 0
therefore Eð�zÞ ¼ 0, which contradicts with the assumption Eð�zÞ 6¼ 0. Therefore we must have Eð�zÞ ¼ 0. Weconclude that any cluster point of {zk} is the solution of E(z) = 0. h
Theorem 3.4. Let conditions of Theorem 2.1 be satisfied at the point (x*,l*), K(z) satisfies a Lipschitz condition
in a neighborhood of z*. Assume that the sequence {zk} generated by Algorithm 3.1 converges to z*, then there is
an integer K large enough such that hk = 1 for k > K, the sequence {zk} generated by Algorithm 3.1 with r > 0
converges quadratically to z*, where l�i ¼ ðy�i Þ2 ði ¼ 1; . . . ;mÞ, z* = (x*, y*).
Proof. Let H(z) = /(z), then H (z*) = 0n+m. From Lemma 3.1 we have that K(z*) is a nonsingular matrix, thenthere are constants c0 > 0, c1 > 0, c2 > 0, c3 > 0, and a neighborhood Nðz�Þ, such that
kHðzÞk ¼ kHðzÞ � Hðz�ÞkP c0kz� z�k;
kHðzÞk ¼ kHðzÞ � Hðz�Þk 6 c1kz� z�k;
kKðzÞ � Kðz�Þk 6 c2kx� x�k;
kKðzÞ�1k 6 c3; z 2Nðz�Þ ð3:6Þ
for sufficiently large K, when k P K, we have
kzk � z�k 6 r0; zk 2Nðz�Þ; ð3:7Þ
where r0 ¼ffiffiffiffiffiffiffiffiffiffiffiffiffiffi1� 2rp
c0=ðc1c3c2ð2þ c3c1ÞÞ, and therefore
kdkk ¼ kKðzkÞ�1HðzkÞk 6 c3c1kzk � z�k: ð3:8Þ
Table 1Numerical results
No. Algorithm IT k/(z)k2 Accuracy Time (s)
1 [14,Tp113] R (3.1) 80 1.595012 · 10�8 8.4655 · 10�6 4504.7n = 10, m = 8 3.1 21 8.886450 · 10�8 2.6308 · 10�5 2752.3
2 [14,Tp108] R (3.1) 83 6.449789 · 10�11 0.0328 6968.8n = 9, m = 14 3.1 20 9.917368 · 10�11 0.0328 383.1880
3 [14,Tp100] R (3.1) 128 1.156948 · 10�6 3.7042 · 10�5 2882.3n = 7, m = 4 3.1 15 3.114798 · 10�6 6.7895 · 10�5 56.3750
4 [14,Tp45] R (3.1) 62 4.833404 · 10�9 7.0017 · 10�5 40.4530n = 5, m = 10 3.1 17 9.219.51 · 10�9 1.0039 · 10�4 10.7180
n stands for the dimension of variables, m for the number of constraints, IT for the number of iterations and R (3.1) for Runge–Kuttaalgorithm.
1342 L. Jin et al. / Applied Mathematics and Computation 188 (2007) 1334–1343
We can deduce from (3.6)–(3.8) that
kHðzk þ dkÞk ¼ kHðzk þ dkÞ � HðzkÞ þ KðzkÞdkk6 sup
06t61kðKðzk þ tdkÞ � KðzkÞÞdkk
6 sup06t61
kðKðzk þ tdkÞ � Kðz�ÞÞdkk þ kðKðzkÞ � Kðz�ÞÞdkk
6 2c2kzk � z�kkdkk þ c2kdkk2
6 c1c3c2ð2þ c3c1Þkzk � z�k2; ð3:9Þ
therefore, when k P K,
EðzkÞ � Eðzk þ dkÞ � 2qEðzkÞ ¼ ð1� 2rÞkHðzkÞk2 � kHðzk þ dkÞk2
P ð1� 2qÞc20kzk � z�k2 � ½c1c3c2ð2þ c3c1Þ�2kzk � z�k4
P ½ð1� 2qÞc20 � ½c1c3c2ð2þ c3c1Þ�2r2
0�kzk � z�k2 P 0; ð3:10Þ
which implies that the step length hk = 1 will satisfy Armijo line search rule when k P K in Algorithm 3.1 andhence Algorithm 3.1 becomes the classical Newton method and the rate of convergence is quadratic, see forinstance [13]. h
4. Computation results
To verify the computational efficiency, we implement a code complied based on Algorithm 3.1 and theRunge–Kutta method for the differential equation system (3.1). The test problems chosen from [14] are usedto compare the numerical performances of these two approaches.
Table 1 gives numerical results of Algorithm 3.1 and the Runge–Kutta method on four test problems inwhich, for the sake of brevity, only six significant digits of the solutions are given. Problem Tp108 is interestingin the sense that we obtained the approximate solution x�com ¼ ð0:8764; 0:4817; 0:0210; 0:9998; 0:8764;0:4817; 0:0210; 0:9998; 0:0000Þ and f �com ¼ �0:8660257, which is better than the one given in [14]. Actuallywe find that most of our numerical results are considerably more accurate than those reported in [14].
To compare the numerical results obtained by Algorithm 3.1 and Runge–Kutta algorithm, we adopted thesame the initial point and r in computing each of the problems. The numerical results given show that Runge–Kutta method has good stability and high precision and Algorithm 3.1 is fast convergent.
References
[1] K.J. Arrow, L. Hurwicz, Reduction of constrained maxima to saddle point problems, in: J. Neyman (Ed.), Proceedings of the 3rdBerkeley Symposium on Mathematical Statistics and Probability, University of California Press, Berkeley, 1956, pp. 1–26.
[2] A.V. Fiacco, G.P. Mccormick, Nonlinear Programming: Sequential Unconstrained Minimization Techniques, John Wiely,New York, 1968.
L. Jin et al. / Applied Mathematics and Computation 188 (2007) 1334–1343 1343
[3] Yu.G. Evtushenko, Two numerical methods of solving nonlinear programming problems, Sov. Math. Dokl 15 (2) (1974) 420–423.[4] H. Yamadhita, A differential equation approach to nonlinear programming, Math. Prog. 18 (1980) 115–168.[5] P.Q. Pan, New ODE methods for equality constrained optimization (1)-equations, J. Comput. Math. 10 (1) (1992) 77–92.[6] Yu.G. Evtushenko, Numerical Optimization Techniques, Optimization Software, Inc. Publication Division, New York, 1985.[7] Yu.G. Evtushenko, V.G. Zhadan, Barrier-projective methods for nonlinear programming, Comp. Maths Math. Phys. 34 (5) (1994)
579–590.[8] Yu.G. Evtushenko, V.G. Zhadan, Stable barrier-projection and barrier-Newton methods in nonlinear programming, Opt. Software 3
(1994) 237–256.[9] Yu.G. Evtushenko, V.G. Zhadan, Stable barrier-projection and barrier-Newton methods for linear and nonlinear programming, in:
E. Spedicato (Ed.), Algorithms for Continuous Optimization, Kulwer Academic Publishers, 1994, pp. 255–285.[10] Yu.G. Evtushenko, V.G. Zhadan, Stable barrier-projection and barrier-Newton methods in linear programming, Comput. Opt. Appl.
3 (1994) 289–303.[11] L.W. Zhang, A modified version to the differential system of Evtushenko and Zhanda for solving nonlinear programming, in:
Ya-xiang Yuan (Ed.), Numerical Linear Algebra and Optimization, Science Press, 1999, pp. 161–168.[12] R. Cominetti, J.P. Dussault, A stable exponential-algorithm with super-linear convergence, J. Optim. Theor. Appl. 2 (83) (1994) 285–
309.[13] J.M. Ortega, W.C. Rheinboldt, Iterative Solution of Nonlinear Equations in Several Variables, Academic Press, New York, 1970.[14] W. Hock, K. Schittkowski, Test examples for nonlinear programming codes’, Lecture Notes Econom. Math. Syst. 187 (1981).