Upload
manu-bhat
View
216
Download
0
Embed Size (px)
Citation preview
7/29/2019 SSR_CH5optimisation
1/107
1
One-Dimensional Minimization Methods
The basic philosophy of most of the numerical methods of optimization isto produce a sequence of improved approximations to the optimumaccording to the following scheme:
1. Start with an initial trial pointX1.
2. Find a suitable direction Si (i = 1 to start with) that points in the generaldirection of the optimum.
3. Find an appropriate step length*ifor movement along the direction Si.
4. Obtain the new approximationXi+1as
Xi+1 = Xi + *i Si (5.1)
7/29/2019 SSR_CH5optimisation
2/107
2
5. Test whether Xi+1 is optimum. If Xi+1 is optimum, stop the procedure.Otherwise, set a new i = i + 1 and repeat step (2) onward.
7/29/2019 SSR_CH5optimisation
3/107
3
7/29/2019 SSR_CH5optimisation
4/107
4
If f(X) is the objective function to be minimized, the problem ofdetermining*ireduces to finding the valuei =
*ithat minimizes
f (Xi+1) = f (Xi+ i Si) = f (i)
for fixed values ofXi and Si . Since f becomes a function of one variable ionly, the methods of finding *i in Eq. (5.1) are called one-dimensional
minimization methods. Several methods are available for solving a one-dimensional minimization problem. These can be classified as shown inTable 5.1.
7/29/2019 SSR_CH5optimisation
5/107
5
7/29/2019 SSR_CH5optimisation
6/107
6
7/29/2019 SSR_CH5optimisation
7/107
7
7/29/2019 SSR_CH5optimisation
8/107
8
7/29/2019 SSR_CH5optimisation
9/107
9
Elimination Methods
Unrestricted search
In most practical problems, the optimum solution is known to lie within
restricted ranges of the design variables. In some cases this range is notknown, and hence the search has to be made with no restrictions on thevalues of the variables.
Search with Fixed Step Size
The most elementary approach for such a problem is to use a fixed stepsize and move from an initial guess point in a favorable direction (positiveor negative). The step size used must be small in relation to the finalaccuracy desired. Although this method is very simple to implement, it is
7/29/2019 SSR_CH5optimisation
10/107
10
not efficient in many cases. This method is described in the followingsteps:
1. Start with an initial guess point, say,x1.
2. Findf1 = f (x1).
3. Assuming a step size s, findx2 = x1 + s.
4. Findf2 = f (x2).
5. If f2 < f1, and if the problem is one of minimization, the assumption ofunimodality indicates that the desired minimum cannot lie atx < x1. Hence
the search can be continued further along points x3, x4 . . . using theunimodality assumption while testing each pair of experiments. Thisprocedure is continued until a point,xi = x1+ (i 1)s, shows an increase inthe function value.
7/29/2019 SSR_CH5optimisation
11/107
11
6. The search is terminated at xi, and either xi1 or xi can be taken as the
optimum point.
7. Originally, iff2 >f1, the search should be carried in the reverse directionat pointsx2, x3, . . . , wherexj = x1(j 1)s.
8. If f2 = f1, the desired minimum lies in between x1 and x2, and theminimum point can be taken as eitherx1 orx2.
9. If it happens that both f2 and f2 are greater than f1, it implies that thedesired minimum will lie in the double intervalx2 < x < x2.
Example:
Find the minimum of f = x(x 1.5) by starting from 0.0 with an initial stepsize of 0.05.
7/29/2019 SSR_CH5optimisation
12/107
12
Solution:
The function value atx1 isf1 = 0.0. If we try to start moving in the negativex direction, we find that x2= 0.05 and f2 = 0.0775. Since f2 >f1, theassumption of unimodality indicates that the minimum cannot lie towardthe left of x2. Thus we start moving in the positive x direction and obtain
the following results:
7/29/2019 SSR_CH5optimisation
13/107
13
From these results, the optimum point can be seen to be xoptx6 = 0.8. Inthis case, the pointsx6 andx7 do not really bracket the minimum point butprovide information about it. If a better approximation to the minimum isdesired, the procedure can be restarted fromx5 with a smaller step size.
7/29/2019 SSR_CH5optimisation
14/107
14
DICHOTOMOUS SEARCH
The exhaustive search method is a simultaneous search method in whichall the experiments are conducted before any judgment is made regardingthe location of the optimum point. The dichotomous search method, aswell as the Fibonacci and the golden section methods discussed in
subsequent sections, are sequential search methods in which the resultof any experiment influences the location of the subsequent experiment.In the dichotomous search, two experiments are placed as close aspossible at the center of the interval of uncertainty. Based on the relativevalues of the objective function at the two points, almost half of theinterval of uncertainty is eliminated. Let the positions of the two
experiments be given by (Fig. 5.7)
7/29/2019 SSR_CH5optimisation
15/107
15
7/29/2019 SSR_CH5optimisation
16/107
16
(5.2)
Where is a small positive number chosen so that the two experimentsgive significantly different results. Then the new interval of uncertainty isgiven by (L0/2 + /2). The building block of dichotomous search consists ofconducting a pair of experiments at the center of the current interval ofuncertainty. The next pair of experiments is, therefore, conducted at the
center of the remaining interval of uncertainty. This results in thereduction of the interval of uncertainty by nearly a factor of 2. Theintervals of uncertainty at the end of different pairs of experiments aregiven in the following table:
7/29/2019 SSR_CH5optimisation
17/107
17
In general, the final interval of uncertainty after conducting n experiments(n even) is given by
Example:
Find the minimum of f = x(x 1.5) in the interval (0.0, 1.00) to within 10%of the exact value.
7/29/2019 SSR_CH5optimisation
18/107
18
Solution:
The ratio of final to initial intervals of uncertainty is given by [from Eq.(5.3)]
Where is a small quantity, say 0.001, and n is the number ofexperiments. If the middle point of the final interval is taken as theoptimum point, the requirement can be stated as
7/29/2019 SSR_CH5optimisation
19/107
19
Since = 0.001 andL0 = 1.0, we have
7/29/2019 SSR_CH5optimisation
20/107
20
Since n has to be even, this inequality gives the minimum admissiblevalue of n as 6. The search is made as follows. The first two experiments
are made at
with the function values given by
7/29/2019 SSR_CH5optimisation
21/107
21
Since f2 < f1, the new interval of uncertainty will be (0.4995, 1.0). Thesecond pair of experiments is conducted a
which give the function values as
7/29/2019 SSR_CH5optimisation
22/107
22
Since f3 >f4, we delete (0.4995, x3) and obtain the new interval ofuncertainty as
(x3, 1.0) = (0.74925, 1.0)
The final set of experiments will be conducted at
The corresponding function values are
7/29/2019 SSR_CH5optimisation
23/107
23
Sincef5 < f6, the new interval of uncertainty is given by (x3, x6) = (0.74925,
0.875125). The middle point of this interval can be taken as optimum, andhence
xopt 0.8121875 andfopt0.5586327148
INTERVAL HALVING METHOD
In the interval halving method, exactly one-half of the current interval ofuncertainty is deleted in every stage. It requires three experiments in thefirst stage and two experiments in each subsequent stage. The procedurecan be described by the following steps:
1. Divide the initial interval of uncertainty L0 = [a, b] into four equal partsand label the middle pointx0 and the quarter-interval pointsx1andx2.
7/29/2019 SSR_CH5optimisation
24/107
24
2. Evaluate the function f (x) at the three interior points to obtain f1 = f(x1),f0 = f (x0), andf2 = f (x2).
3. (a) Iff2 >f0 >f1as shown in Fig. 5.8a, delete the interval (x0, b), label
x1andx0as the newx0 and b, respectively, and go to step 4.
(b) Iff2 < f0 < f1 as shown in Fig. 5.8b, delete the interval (a, x0), label
x2 andx0 as the newx0 and a, respectively, and go to step 4.
(c) Iff1 >f0 andf2 >f0 as shown in Fig. 5.8c, delete both the intervals(a, x1) and (x2, b), labelx1 andx2 as the new a and b, respectively,and go to step 4.
4. Test whether the new interval of uncertainty, L = b a, satisfies theconvergence criterion L , where is a small quantity. If the
7/29/2019 SSR_CH5optimisation
25/107
25
convergence criterion is satisfied, stop the procedure. Otherwise, set thenewL0 = L and go to step 1.
Remarks:
1. In this method, the function value at the middle point of the interval ofuncertainty,f0, will be available in all the stages except the first stage.
2. The interval of uncertainty remaining at the end of n experiments (n 3and odd) is given by
(5.3)
7/29/2019 SSR_CH5optimisation
26/107
26
7/29/2019 SSR_CH5optimisation
27/107
27
7/29/2019 SSR_CH5optimisation
28/107
28
7/29/2019 SSR_CH5optimisation
29/107
29
7/29/2019 SSR_CH5optimisation
30/107
30
Example:
Find the minimum of f = x(x 1.5) in the interval (0.0, 1.0) to within 10%of the exact value.
Solution:
If the middle point of the final interval of uncertainty is taken as theoptimum point, the specified accuracy can be achieved if
SinceL0 = 1,
7/29/2019 SSR_CH5optimisation
31/107
31
Since n has to be odd, inequality gives the minimum permissible value ofn as 7. With this value of n = 7, the search is conducted as follows. Thefirst three experiments are placed at one-fourth points of the interval L0 =
[a = 0, b = 1] as
x1 = 0.25,f1= 0.25(1.25) = 0.3125x0 = 0.50,f0= 0.50(1.00) = 0.5000x2 = 0.75,f2= 0.75(0.75) = 0.5625
Sincef1 >f0 >f2, we delete the interval (a, x0) = (0.0, 0.5), labelx2 andx0 asthe new x0 and a so that a = 0.5, x0 = 0.75, and b = 1.0. By dividing thenew interval of uncertainty,L3 = (0.5, 1.0) into four equal parts, we obtain
7/29/2019 SSR_CH5optimisation
32/107
32
x1 = 0.625,f1= 0.625(0.875) = 0.546875x0 = 0.750,f0= 0.750(0.750) = 0.562500x2 = 0.875,f2= 0.875(0.625) = 0.546875
Since f1 >f0and f2 >f0, we delete both the intervals (a, x1) and (x2, b), andlabel x1, x0, and x2 as the new a, x0, and b, respectively. Thus the newinterval of uncertainty will be L5 = (0.625, 0.875). Next, this interval isdivided into four equal parts to obtain
x1= 0.6875,f1= 0.6875(0.8125) = 0.558594x0 = 0.75,f0= 0.75(0.75) = 0.5625
x2 = 0.8125,f2= 0.8125(0.6875) = 0.558594
Again we note thatf1 >f0andf2 >f0and hence we delete both the intervals(a, x1) and (x2, b) to obtain the new interval of uncertainty as L7 = (0.6875,
7/29/2019 SSR_CH5optimisation
33/107
33
0.8125). By taking the middle point of this interval (L7) as optimum, weobtain
xopt 0.75 andfopt 0.5625(This solution happens to be the exact solution in this case.)
FIBONACCI METHOD
As stated earlier, the Fibonacci method can be used to find the minimumof a function of one variable even if the function is not continuous. Thismethod, like many other elimination methods, has the followinglimitations:
1. The initial interval of uncertainty, in which the optimum lies, has to beknown.
7/29/2019 SSR_CH5optimisation
34/107
34
2. The function being optimized has to be unimodal in the initial interval ofuncertainty.
3. The exact optimum cannot be located in this method. Only an intervalknown as the final interval of uncertainty will be known. The final intervalof uncertainty can be made as small as desired by using morecomputations.
4. The number of function evaluations to be used in the search or theresolution required has to be specified beforehand.
This method makes use of the sequence of Fibonacci numbers, {Fn}, forplacing the experiments. These numbers are defined as
F0 = F1 = 1
Fn = Fn1 + Fn2, n = 2, 3, 4, . . .
7/29/2019 SSR_CH5optimisation
35/107
35
which yield the sequence 1, 1, 2, 3, 5, 8, 13, 21, 34, 55, 89,. . . .
Procedure:
LetL0be the initial interval of uncertainty defined by a x b and n be thetotal number of experiments to be conducted. Define
(5.4)and place the first two experiments at points x1 and x2, which are located
at a distance ofL2from each end ofL0. This gives
7/29/2019 SSR_CH5optimisation
36/107
36
(5.5)
Discard part of the interval by using the unimodality assumption. Thenthere remains a smaller interval of uncertaintyL2 given by
(5.6)and with one experiment left in it. This experiment will be at a distance of
7/29/2019 SSR_CH5optimisation
37/107
37
(5.7)from one end and
(5.8)from the other end. Now place the third experiment in the interval L2 sothat the current two experiments are located at a distance of
(5.9)
7/29/2019 SSR_CH5optimisation
38/107
38
from each end of the interval L2. Again the unimodality property will allowus to reduce the interval of uncertainty toL3given by
(5.10)This process of discarding a certain interval and placing a newexperiment in the remaining interval can be continued, so that the locationof the j th experiment and the interval of uncertainty at the end of jexperiments are, respectively, given by
(5.11)
7/29/2019 SSR_CH5optimisation
39/107
39
(5.12)The ratio of the interval of uncertainty remaining after conductingj of the npredetermined experiments to the initial interval of uncertainty becomes
(5.13)and forj = n, we obtain
(5.14)The ratio Ln/L0 will permit us to determine n, the required number ofexperiments, to achieve any desired accuracy in locating the optimum
0
7/29/2019 SSR_CH5optimisation
40/107
40
point. Table 5.2 gives the reduction ratio in the interval of uncertaintyobtainable for different number of experiments.
Position of the Final Experiment:
In this method the last experiment has to be placed with some care.
(5.15)
Thus after conducting n 1 experiments and discarding the appropriateinterval in each step, the remaining interval will contain one experimentprecisely at its middle point.
41
7/29/2019 SSR_CH5optimisation
41/107
41
However, the final experiment, namely, the nth experiment, is also to beplaced at the center of the present interval of uncertainty. That is, the
position of the nth experiment will be same as that of (n 1)th one, andthis is true for whatever value we choose for n. Since no new informationcan be gained by placing the nth experiment exactly at the same locationas that of the (n 1)th experiment, we place the nth experiment veryclose to the remaining valid experiment, as in the case of thedichotomous search method. This enables us to obtain the final interval of
uncertainty to within 0.5Ln-1A flowchart for implementing the Fibonacci method of minimization isgiven in Fig. 5.9.
42
7/29/2019 SSR_CH5optimisation
42/107
42
43
7/29/2019 SSR_CH5optimisation
43/107
43
44
7/29/2019 SSR_CH5optimisation
44/107
44
45
7/29/2019 SSR_CH5optimisation
45/107
45
Example:
Minimize f(x) = 0.65 [0.75/(1 +x2)] 0.65x tan1(1/x) in the interval[0,3] by the Fibonacci method using n = 6.
Solution:
Here n = 6andL0 = 3.0, which yield
46
7/29/2019 SSR_CH5optimisation
46/107
46
Thus the positions of the first two experiments are given by x1 = 1.153846andx2= 3.0 1.153846 = 1.846154 with f1 =f (x1)= 0.207270 and f2 =f(x2) = 0.115843.
Since f1 is less than f2, we can delete the interval [x2, 3.0] by using theunimodality assumption (Fig. 5.10a).
The third experiment is placed atx3 = 0 + (x2x1) = 1.846154 1.153846= 0.692308, with the corresponding function value off3= 0.291364.
Sincef1 >f3, we delete the interval [x1,x2] (Fig. 5.10b).
47
7/29/2019 SSR_CH5optimisation
47/107
47
The next experiment is located atx4 = 0 + (x1x3) = 1.153846 0.692308= 0.461538 withf4 = 0.309811. Nothing thatf4 < f3, we delete the interval
[x3,x1] (Fig. 5.10c).
The location of the next experiment can be obtained as x5 = 0 + (x3 x4) =0.692308 0.461538 = 0.230770 with the corresponding objectivefunction value off5 = 0.263678.
Sincef5 >f4, we delete the interval [0, x5] (Fig. 5.10d).
The final experiment is positioned at x6 = x5 + (x3 x4) = 0.230770 +(0.692308 0.461538) = 0.461540 with f6 = 0.309810. (Note that,theoretically, the value of x6should be same as that of x4; however, it isslightly different fromx
4
, due to round-off error).Since f6 >f4, we delete the interval [x6, x3] and obtain the final interval ofuncertainty as L6 = [x5, x6] = [0.230770, 0.461540] (Fig. 5.10e). The ratioof the final to the initial interval of uncertainty is
48
7/29/2019 SSR_CH5optimisation
48/107
48
This value can be compared with Eq. (5.14), which states that if nexperiments (n = 6) are planned, a resolution no finer than 1/Fn = 1/F6 =1/13 = 0.076923 can be expected from the method.
49
7/29/2019 SSR_CH5optimisation
49/107
50
7/29/2019 SSR_CH5optimisation
50/107
51
7/29/2019 SSR_CH5optimisation
51/107
52
7/29/2019 SSR_CH5optimisation
52/107
53
7/29/2019 SSR_CH5optimisation
53/107
54
7/29/2019 SSR_CH5optimisation
54/107
55
7/29/2019 SSR_CH5optimisation
55/107
56
7/29/2019 SSR_CH5optimisation
56/107
GOLDEN SECTION METHOD
The golden section method is same as the Fibonacci method except thatin the Fibonacci method the total number of experiments to be conductedhas to be specified before beginning the calculation, whereas this is notrequired in the golden section method. In the Fibonacci method, the
location of the first two experiments is determined by the total number ofexperiments, N. In the golden section method we start with theassumption that we are going to conduct a large number of experiments.Of course, the total number of experiments can be decided during thecomputation. The intervals of uncertainty remaining at the end of differentnumber of experiments can be computed as follows:
57
7/29/2019 SSR_CH5optimisation
57/107
(5.16)
(5.17)This result can be generalized to obtain
(5.18)
58
7/29/2019 SSR_CH5optimisation
58/107
Using the relation
(5.19)we obtain, after dividing both sides by FN1,
(5.20)By defining a ratio as
(5.21)
So,
59
7/29/2019 SSR_CH5optimisation
59/107
that is,
(5.22)This gives the root = 1.618, and hence
(5.23)the ratios FN2/FN1 and FN1/FN have been taken to be same forlarge values of N. The validity of this assumption can be seen from the
following table:
60
7/29/2019 SSR_CH5optimisation
60/107
The ratio has a historical background. Ancient Greek architects believedthat a building having the sides d and b satisfying the relation
(5.24)would have the most pleasing properties (Fig. 5.11). The origin of thename, golden section method, can also be traced to the Euclidsgeometry. In Euclids geometry, when a line segment is divided into twounequal parts so that the ratio of the whole to the larger part is equal to
61
7/29/2019 SSR_CH5optimisation
61/107
the ratio of the larger to the smaller, the division is called the goldensection and the ratio is called the golden mean.
Procedure:
62
7/29/2019 SSR_CH5optimisation
62/107
The procedure is same as the Fibonacci method except that the locationof the first two experiments is defined by
The desired accuracy can be specified to stop the procedure.
Example:
Minimize the function
using the golden section method with n = 6.
Solution:
63
7/29/2019 SSR_CH5optimisation
63/107
The locations of the first two experiments are defined by
L
2= 0.382L0 = (0.382)(3.0) = 1.1460.
Thusx1 = 1.1460 andx2= 3.0 1.1460 = 1.8540
withf1 =f (x1) = 0.208654 andf2 =f (x2) = 0.115124.
Since f1 < f2, we delete the interval [x2, 3.0] based on the assumption ofunimodality and obtain the new interval of uncertainty as L2 = [0, x2] =[0.0, 1.8540].
The third experiment is placed at x3 = 0 + (x2x1) = 1.8540 1.1460 =
0.7080.
64
7/29/2019 SSR_CH5optimisation
64/107
Since f3 = 0.288943 is smaller than f1 = 0.208654, we delete theinterval [x1,x2] and obtain the new interval of uncertainty as [0.0, x1] = [0.0,
1.1460].The position of the next experiment is given by x4 = 0 + (x1x3) = 1.1460 0.7080 = 0.4380 withf4= 0.308951.
Since f4 < f3, we delete [x3, x1] and obtain the new interval of uncertaintyas [0,x3] = [0.0, 0.7080].
The next experiment is placed at x5 = 0 + (x3x4) = 0.7080 0.4380 =0.2700. Sincef5= 0.278434 is larger thanf4= 0.308951, we delete theinterval [0, x5] and obtain the new interval of uncertainty as [x5, x3] =[0.2700, 0.7080].
The final experiment is placed at x6 = x5 + (x3x4) = 0.2700 + (0.7080 0.4380) = 0.5400 withf6= 0.308234.
65
7/29/2019 SSR_CH5optimisation
65/107
Since f6 >f4, we delete the interval [x6, x3] and obtain the final interval ofuncertainty as [x5, x6] = [0.2700, 0.5400]. Note that this final interval of
uncertainty is slightly larger than the one found in the Fibonacci method,[0.461540, 0.230770]. The ratio of the final to the initial interval ofuncertainty in the present case is
66
7/29/2019 SSR_CH5optimisation
66/107
67
7/29/2019 SSR_CH5optimisation
67/107
68
7/29/2019 SSR_CH5optimisation
68/107
CUBIC INTERPOLATION METHOD
69
7/29/2019 SSR_CH5optimisation
69/107
70
7/29/2019 SSR_CH5optimisation
70/107
71
7/29/2019 SSR_CH5optimisation
71/107
72
7/29/2019 SSR_CH5optimisation
72/107
Example:
Find the minimum of f = 5 53 20 + 5 by the cubic interpolationmethod.
Solution:
Since this problem has not arisen during a multivariable optimizationprocess, we can skip stage 1. We take A = 0 and find that
To findB at which df/d is nonnegative, we start with t0 = 0.4 and evaluatethe derivative at t0, 2t0, 4t0, . . . . This gives
73
7/29/2019 SSR_CH5optimisation
73/107
Thus we find that
Iteration 1:
74
7/29/2019 SSR_CH5optimisation
74/107
To find the value of and to test the convergence criteria, we firstcompute Z and Q as
Hence,
By discarding the negative value, we have
75
7/29/2019 SSR_CH5optimisation
75/107
Convergence criterion:
If is close to the true minimum, , then shouldbe approximately zero. Sincef = 54152 20,
Since this is not small, we go to the next iteration or refitting. As
we take
A =
76
7/29/2019 SSR_CH5optimisation
76/107
Iteration 2
77
7/29/2019 SSR_CH5optimisation
77/107
Convergence criterion:
Since this value is large, we go the next iteration with
78
7/29/2019 SSR_CH5optimisation
78/107
Iteration 3
79
Convergence criterion:
7/29/2019 SSR_CH5optimisation
79/107
g
Assuming that this value is close to zero, we can stop the iterativeprocess and take
80
7/29/2019 SSR_CH5optimisation
80/107
Newton Method
Consider the quadratic approximation of the function f () at = i usingthe Taylors series expansion:
(5.25)
By setting the derivative of Eq. (5.25) equal to zero for the minimum of f(), we obtain
(5.26)
81
If i denotes an approximation to the minimum of f(), Eq. (5.26) can be
7/29/2019 SSR_CH5optimisation
81/107
pp f( ) q ( )rearranged to obtain an improved approximation as
(5.27)Thus the Newton method, Eq. (5.27), is equivalent to using a quadraticapproximation for the function f() and applying the necessary conditions.The iterative process given by Eq. (5.65) can be assumed to have
converged when the derivative,f (i+1), is close to zero:
(5.28)where is a small quantity. The convergence process of the method is
shown graphically in Fig. 5.18a.Remarks:
82
1. The Newton method was originally developed by Newton for solving
7/29/2019 SSR_CH5optimisation
82/107
nonlinear equations and later refined by Raphson, and hence the methodis also known as NewtonRaphson method in the literature of numerical
analysis.2. The method requires both the first- and second-order derivatives off().
3.f (i ) = 0 [in Eq. (5.27)], the Newton iterative method has a powerful(fastest) convergence property, known as quadratic convergence.
4. If the starting point for the iterative process is not close to the true
solution , the Newton iterative process might diverge as illustrated inFig. 5.18b.
83
7/29/2019 SSR_CH5optimisation
83/107
84
7/29/2019 SSR_CH5optimisation
84/107
85
7/29/2019 SSR_CH5optimisation
85/107
Example:
Find the minimum of the function
using the NewtonRaphson method with the starting point 1 = 0.1. Use = 0.01 in Eq. (5.66) for checking the convergence.
Solution:
The first and second derivatives of the functionf() are given by
86
7/29/2019 SSR_CH5optimisation
86/107
Iteration 1
1 = 0.1,f (1)= 0.188197,f (1)= 0.744832,f (1) = 2.68659
Convergence check: |f (2)| = |0.138230|>.
Iteration 2
87
f ( ) 0 303279 f ( ) 0 138230 f ( ) 1 57296
7/29/2019 SSR_CH5optimisation
87/107
f (2)= 0.303279,f (2)= 0.138230,f (2) = 1.57296
Convergence check: |f (3)| = |0.0179078|>.
Iteration 3
f (3) = 0.309881,f (3) = 0.0179078,f (3) = 1.17126
Convergence check: |f (4)| = |0.0005033| < .
88
Since the process has converged, the optimum solution is taken as4
7/29/2019 SSR_CH5optimisation
88/107
= 0.480409.
Quasi-Newton Method
If the function being minimized f () is not available in closed form or isdifficult to differentiate, the derivatives f () andf () in Eq. (5.27) can be
approximated by the finite difference formulas as
(5.29)
(5.30)
89
where is a small step size. Substitution of Eqs. (5.29) and (5.30) into Eq.(5 27) leads to
7/29/2019 SSR_CH5optimisation
89/107
(5.27) leads to
(5.31)The iterative process indicated by Eq. (5.69) is known as the quasi-Newton method. To test the convergence of the iterative process, the
following criterion can be used:
(5.32)where a central difference formula has been used for evaluating thederivative of f and is a small quantity.
90
Remarks:
7/29/2019 SSR_CH5optimisation
90/107
Remarks:
1. The central difference formulas have been used in Eqs. (5.31) and(5.32). However, the forward or backward difference formulas can also beused for this purpose.
2. Equation (5.31) requires the evaluation of the function at the pointsi+and
i in addition to
iin each iteration.
Example:
Find the minimum of the function
91
using quasi-Newton method with the starting point 1 = 0.1 and the stepsize = 0 01 in central difference formulas Use = 0 01 in Eq (5 32) for
7/29/2019 SSR_CH5optimisation
91/107
size = 0.01 in central difference formulas. Use = 0.01 in Eq. (5.32) forchecking the convergence.
Solution:
Iteration 1
1 = 0.1, = 0.01, = 0.01, f1 = f (1) = 0.188197,f + 1 = f (1 + ) =0.195512,f 1 = f(1)= 0.180615
92
7/29/2019 SSR_CH5optimisation
92/107
Convergence check:
Iteration 2f2= f(2)= 0.303368,f+2 =f(2+)=0.304662,f2=f(2)=0.301916
93
Convergence check:
7/29/2019 SSR_CH5optimisation
93/107
Convergence check:
Iteration 3
f3= f(3) = 0.309885, f+3 = f(3+) = 0.310004, f3 = f(3) =0.309650
94
Convergence check:
7/29/2019 SSR_CH5optimisation
94/107
Since the process has converged, we take the optimum solution as 4
= 0.480600
Secant Method
The secant method uses an equation similar to Eq. (5.26) as
95
7/29/2019 SSR_CH5optimisation
95/107
(5.33)where s is the slope of the line connecting the two points (A,f(A)) and (B,f(B)), where A and B denote two different approximations to the correctsolution,. The slope s can be expressed as (Fig. 5.19)
(5.34)Equation (5.33) approximates the function f () between A and B as alinear equation (secant), and hence the solution of Eq. (5.33) gives thenew approximation to the root off () as
96
7/29/2019 SSR_CH5optimisation
96/107
(5.35)The iterative process given by Eq. (5.35) is known as the secant method(Fig. 5.19). Since the secant approaches the second derivative of f () atA as B approaches A, the secant method can also be considered as aquasi-Newton method. It can also be considered as a form of eliminationtechnique since part of the interval, (A,
i+1) in Fig. 5.19, is eliminated in
every iteration. The iterative process can be implemented by using thefollowing step-by-step procedure.
1. Set1 = A = 0 and evaluatef (A). The value off (A) will be negative.Assume an initial trial step length t0. Set i = 1.
2. Evaluatef (t0).
97
3. Iff (t0) < 0, set A =i = t0,f (A) =f (t0), new t0 = 2t0, and go to step 2.4. Iff (t0) 0, set B = t0,f (B) =f (t0), and go to step 5.
7/29/2019 SSR_CH5optimisation
97/107
f ( 0) , 0, f ( ) f ( 0), g p
5. Find the new approximate solution of the problem as
(5.36)6. Test for convergence:
|f (i + 1)| (5.37)
where is a small quantity. If Eq. (5.37) is satisfied, take i+1 and stop
the procedure. Otherwise, go to step 7.
98
7. Iff (i+1) 0, set new B =i+1,f (B) =f (i+1), i = i + 1, and go to step 5.
7/29/2019 SSR_CH5optimisation
98/107
8. Iff (i+1) < 0, set new A =i+1,f (A) =f (i+1), i = i + 1, and go to step 5.
Remarks:
1. The secant method is identical to assuming a linear equation for f ().This implies that the original function,f (), is approximated by a quadraticequation.
2. In some cases we may encounter a situation where the function f ()varies very slowly with , as shown in Fig. 5.20. This situation can beidentified by noticing that the point B remains unaltered for severalconsecutive refits. Once such a situation is suspected, the convergence
process can be improved by taking the next value of i+1 as (A + B)/2instead of finding its value from Eq. (5.36).
99
7/29/2019 SSR_CH5optimisation
99/107
100
7/29/2019 SSR_CH5optimisation
100/107
101
7/29/2019 SSR_CH5optimisation
101/107
102
Example:
7/29/2019 SSR_CH5optimisation
102/107
Find the minimum of the function
using the secant method with an initial step size of t0 = 0.1,1= 0.0, and
= 0.01.
Solution:
1 = A = 0.0, t0 = 0.1,f (A)= 1.02102, B = A + t0= 0.1,f (B)= 0.744832.
Sincef (B) < 0, we set new A = 0.1,f (A) = 0.744832, t0= 2(0.1) = 0.2, B=1 + t0= 0.2, and computef (B) = 0.490343.
103
Sincef (B) < 0, we set new A = 0.2,f (A) = 0.490343, t0 = 2(0.2) = 0.4, B=1 + t0= 0.4, and computef (B) = 0.103652.
7/29/2019 SSR_CH5optimisation
103/107
Sincef (B) < 0, we set new A = 0.4,f (A) =0.103652, t0 = 2(0.4) = 0.8, B=1 + t0= 0.8, and computef (B) = +0.180800.
Sincef (B)>0, we proceed to find 2.
Iteration 1
Since A = 1 = 0.4, f (A) = 0.103652,B = 0.8, f (B) = +0.180800, wecompute
Convergence check:|f (2)| = |+0.0105789|>.
104
Iteration 2
7/29/2019 SSR_CH5optimisation
104/107
Since f (2) = +0.0105789>0, we set new A = 0.4, f (A) = 0.103652,B =2 = 0.545757,f (B) =f (2) = +0.0105789, and compute
Convergence check: |f (3)| = |+0.00151235| < .
Since the process has converged, the optimum solution is given by3
= 0.490632.
105
7/29/2019 SSR_CH5optimisation
105/107
PRACTICAL CONSIDERATIONS
How to Make the Methods Efficient and More Reliable
In some cases, some of the interpolation methods discussed may be veryslow to converge, may diverge, or may predict the minimum of thefunction,f(), outside the initial interval of uncertainty, especially when theinterpolating polynomial is not representative of the variation of thefunction being minimized. In such cases we can use the Fibonacci or
golden section method to find the minimum. In some problems it mightprove to be more efficient to combine several techniques. For example,the unrestricted search with an accelerated step size can be used to
106
bracket the minimum and then the Fibonacci or the golden sectionmethod can be used to find the optimum point. In some cases theFibonacci or golden section method can be used in conjunction with an
7/29/2019 SSR_CH5optimisation
106/107
Fibonacci or golden section method can be used in conjunction with an
interpolation method.
Comparison of Methods
It has been shown in Section 5.9 that the Fibonacci method is the most
efficient elimination technique in finding the minimum of a function if theinitial interval of uncertainty is known. In the absence of the initial intervalof uncertainty, the quadratic interpolation method or the quasi-Newtonmethod is expected to be more efficient when the derivatives of thefunction are not available. When the first derivatives of the function being
minimized are available, the cubic interpolation method or the secantmethod are expected to be very efficient. On the other hand, if both thefirst and second derivatives of the function are available, the Newton
107
method will be the most efficient one in finding the optimal step length,.
In general, the efficiency and reliability of the various methods are
7/29/2019 SSR_CH5optimisation
107/107
g , y yproblem dependent and any efficient computer program must includemany heuristic additions not indicated explicitly by the method. Theheuristic considerations are needed to handle multimodal functions(functions with multiple extreme points), sharp variations in the slopes(first derivatives) and curvatures (second derivatives) of the function, andthe effects of round-off errors resulting from the precision used in the
arithmetic operations.