SSR_CH5optimisation

Embed Size (px)

Citation preview

  • 7/29/2019 SSR_CH5optimisation

    1/107

    1

    One-Dimensional Minimization Methods

    The basic philosophy of most of the numerical methods of optimization isto produce a sequence of improved approximations to the optimumaccording to the following scheme:

    1. Start with an initial trial pointX1.

    2. Find a suitable direction Si (i = 1 to start with) that points in the generaldirection of the optimum.

    3. Find an appropriate step length*ifor movement along the direction Si.

    4. Obtain the new approximationXi+1as

    Xi+1 = Xi + *i Si (5.1)

  • 7/29/2019 SSR_CH5optimisation

    2/107

    2

    5. Test whether Xi+1 is optimum. If Xi+1 is optimum, stop the procedure.Otherwise, set a new i = i + 1 and repeat step (2) onward.

  • 7/29/2019 SSR_CH5optimisation

    3/107

    3

  • 7/29/2019 SSR_CH5optimisation

    4/107

    4

    If f(X) is the objective function to be minimized, the problem ofdetermining*ireduces to finding the valuei =

    *ithat minimizes

    f (Xi+1) = f (Xi+ i Si) = f (i)

    for fixed values ofXi and Si . Since f becomes a function of one variable ionly, the methods of finding *i in Eq. (5.1) are called one-dimensional

    minimization methods. Several methods are available for solving a one-dimensional minimization problem. These can be classified as shown inTable 5.1.

  • 7/29/2019 SSR_CH5optimisation

    5/107

    5

  • 7/29/2019 SSR_CH5optimisation

    6/107

    6

  • 7/29/2019 SSR_CH5optimisation

    7/107

    7

  • 7/29/2019 SSR_CH5optimisation

    8/107

    8

  • 7/29/2019 SSR_CH5optimisation

    9/107

    9

    Elimination Methods

    Unrestricted search

    In most practical problems, the optimum solution is known to lie within

    restricted ranges of the design variables. In some cases this range is notknown, and hence the search has to be made with no restrictions on thevalues of the variables.

    Search with Fixed Step Size

    The most elementary approach for such a problem is to use a fixed stepsize and move from an initial guess point in a favorable direction (positiveor negative). The step size used must be small in relation to the finalaccuracy desired. Although this method is very simple to implement, it is

  • 7/29/2019 SSR_CH5optimisation

    10/107

    10

    not efficient in many cases. This method is described in the followingsteps:

    1. Start with an initial guess point, say,x1.

    2. Findf1 = f (x1).

    3. Assuming a step size s, findx2 = x1 + s.

    4. Findf2 = f (x2).

    5. If f2 < f1, and if the problem is one of minimization, the assumption ofunimodality indicates that the desired minimum cannot lie atx < x1. Hence

    the search can be continued further along points x3, x4 . . . using theunimodality assumption while testing each pair of experiments. Thisprocedure is continued until a point,xi = x1+ (i 1)s, shows an increase inthe function value.

  • 7/29/2019 SSR_CH5optimisation

    11/107

    11

    6. The search is terminated at xi, and either xi1 or xi can be taken as the

    optimum point.

    7. Originally, iff2 >f1, the search should be carried in the reverse directionat pointsx2, x3, . . . , wherexj = x1(j 1)s.

    8. If f2 = f1, the desired minimum lies in between x1 and x2, and theminimum point can be taken as eitherx1 orx2.

    9. If it happens that both f2 and f2 are greater than f1, it implies that thedesired minimum will lie in the double intervalx2 < x < x2.

    Example:

    Find the minimum of f = x(x 1.5) by starting from 0.0 with an initial stepsize of 0.05.

  • 7/29/2019 SSR_CH5optimisation

    12/107

    12

    Solution:

    The function value atx1 isf1 = 0.0. If we try to start moving in the negativex direction, we find that x2= 0.05 and f2 = 0.0775. Since f2 >f1, theassumption of unimodality indicates that the minimum cannot lie towardthe left of x2. Thus we start moving in the positive x direction and obtain

    the following results:

  • 7/29/2019 SSR_CH5optimisation

    13/107

    13

    From these results, the optimum point can be seen to be xoptx6 = 0.8. Inthis case, the pointsx6 andx7 do not really bracket the minimum point butprovide information about it. If a better approximation to the minimum isdesired, the procedure can be restarted fromx5 with a smaller step size.

  • 7/29/2019 SSR_CH5optimisation

    14/107

    14

    DICHOTOMOUS SEARCH

    The exhaustive search method is a simultaneous search method in whichall the experiments are conducted before any judgment is made regardingthe location of the optimum point. The dichotomous search method, aswell as the Fibonacci and the golden section methods discussed in

    subsequent sections, are sequential search methods in which the resultof any experiment influences the location of the subsequent experiment.In the dichotomous search, two experiments are placed as close aspossible at the center of the interval of uncertainty. Based on the relativevalues of the objective function at the two points, almost half of theinterval of uncertainty is eliminated. Let the positions of the two

    experiments be given by (Fig. 5.7)

  • 7/29/2019 SSR_CH5optimisation

    15/107

    15

  • 7/29/2019 SSR_CH5optimisation

    16/107

    16

    (5.2)

    Where is a small positive number chosen so that the two experimentsgive significantly different results. Then the new interval of uncertainty isgiven by (L0/2 + /2). The building block of dichotomous search consists ofconducting a pair of experiments at the center of the current interval ofuncertainty. The next pair of experiments is, therefore, conducted at the

    center of the remaining interval of uncertainty. This results in thereduction of the interval of uncertainty by nearly a factor of 2. Theintervals of uncertainty at the end of different pairs of experiments aregiven in the following table:

  • 7/29/2019 SSR_CH5optimisation

    17/107

    17

    In general, the final interval of uncertainty after conducting n experiments(n even) is given by

    Example:

    Find the minimum of f = x(x 1.5) in the interval (0.0, 1.00) to within 10%of the exact value.

  • 7/29/2019 SSR_CH5optimisation

    18/107

    18

    Solution:

    The ratio of final to initial intervals of uncertainty is given by [from Eq.(5.3)]

    Where is a small quantity, say 0.001, and n is the number ofexperiments. If the middle point of the final interval is taken as theoptimum point, the requirement can be stated as

  • 7/29/2019 SSR_CH5optimisation

    19/107

    19

    Since = 0.001 andL0 = 1.0, we have

  • 7/29/2019 SSR_CH5optimisation

    20/107

    20

    Since n has to be even, this inequality gives the minimum admissiblevalue of n as 6. The search is made as follows. The first two experiments

    are made at

    with the function values given by

  • 7/29/2019 SSR_CH5optimisation

    21/107

    21

    Since f2 < f1, the new interval of uncertainty will be (0.4995, 1.0). Thesecond pair of experiments is conducted a

    which give the function values as

  • 7/29/2019 SSR_CH5optimisation

    22/107

    22

    Since f3 >f4, we delete (0.4995, x3) and obtain the new interval ofuncertainty as

    (x3, 1.0) = (0.74925, 1.0)

    The final set of experiments will be conducted at

    The corresponding function values are

  • 7/29/2019 SSR_CH5optimisation

    23/107

    23

    Sincef5 < f6, the new interval of uncertainty is given by (x3, x6) = (0.74925,

    0.875125). The middle point of this interval can be taken as optimum, andhence

    xopt 0.8121875 andfopt0.5586327148

    INTERVAL HALVING METHOD

    In the interval halving method, exactly one-half of the current interval ofuncertainty is deleted in every stage. It requires three experiments in thefirst stage and two experiments in each subsequent stage. The procedurecan be described by the following steps:

    1. Divide the initial interval of uncertainty L0 = [a, b] into four equal partsand label the middle pointx0 and the quarter-interval pointsx1andx2.

  • 7/29/2019 SSR_CH5optimisation

    24/107

    24

    2. Evaluate the function f (x) at the three interior points to obtain f1 = f(x1),f0 = f (x0), andf2 = f (x2).

    3. (a) Iff2 >f0 >f1as shown in Fig. 5.8a, delete the interval (x0, b), label

    x1andx0as the newx0 and b, respectively, and go to step 4.

    (b) Iff2 < f0 < f1 as shown in Fig. 5.8b, delete the interval (a, x0), label

    x2 andx0 as the newx0 and a, respectively, and go to step 4.

    (c) Iff1 >f0 andf2 >f0 as shown in Fig. 5.8c, delete both the intervals(a, x1) and (x2, b), labelx1 andx2 as the new a and b, respectively,and go to step 4.

    4. Test whether the new interval of uncertainty, L = b a, satisfies theconvergence criterion L , where is a small quantity. If the

  • 7/29/2019 SSR_CH5optimisation

    25/107

    25

    convergence criterion is satisfied, stop the procedure. Otherwise, set thenewL0 = L and go to step 1.

    Remarks:

    1. In this method, the function value at the middle point of the interval ofuncertainty,f0, will be available in all the stages except the first stage.

    2. The interval of uncertainty remaining at the end of n experiments (n 3and odd) is given by

    (5.3)

  • 7/29/2019 SSR_CH5optimisation

    26/107

    26

  • 7/29/2019 SSR_CH5optimisation

    27/107

    27

  • 7/29/2019 SSR_CH5optimisation

    28/107

    28

  • 7/29/2019 SSR_CH5optimisation

    29/107

    29

  • 7/29/2019 SSR_CH5optimisation

    30/107

    30

    Example:

    Find the minimum of f = x(x 1.5) in the interval (0.0, 1.0) to within 10%of the exact value.

    Solution:

    If the middle point of the final interval of uncertainty is taken as theoptimum point, the specified accuracy can be achieved if

    SinceL0 = 1,

  • 7/29/2019 SSR_CH5optimisation

    31/107

    31

    Since n has to be odd, inequality gives the minimum permissible value ofn as 7. With this value of n = 7, the search is conducted as follows. Thefirst three experiments are placed at one-fourth points of the interval L0 =

    [a = 0, b = 1] as

    x1 = 0.25,f1= 0.25(1.25) = 0.3125x0 = 0.50,f0= 0.50(1.00) = 0.5000x2 = 0.75,f2= 0.75(0.75) = 0.5625

    Sincef1 >f0 >f2, we delete the interval (a, x0) = (0.0, 0.5), labelx2 andx0 asthe new x0 and a so that a = 0.5, x0 = 0.75, and b = 1.0. By dividing thenew interval of uncertainty,L3 = (0.5, 1.0) into four equal parts, we obtain

  • 7/29/2019 SSR_CH5optimisation

    32/107

    32

    x1 = 0.625,f1= 0.625(0.875) = 0.546875x0 = 0.750,f0= 0.750(0.750) = 0.562500x2 = 0.875,f2= 0.875(0.625) = 0.546875

    Since f1 >f0and f2 >f0, we delete both the intervals (a, x1) and (x2, b), andlabel x1, x0, and x2 as the new a, x0, and b, respectively. Thus the newinterval of uncertainty will be L5 = (0.625, 0.875). Next, this interval isdivided into four equal parts to obtain

    x1= 0.6875,f1= 0.6875(0.8125) = 0.558594x0 = 0.75,f0= 0.75(0.75) = 0.5625

    x2 = 0.8125,f2= 0.8125(0.6875) = 0.558594

    Again we note thatf1 >f0andf2 >f0and hence we delete both the intervals(a, x1) and (x2, b) to obtain the new interval of uncertainty as L7 = (0.6875,

  • 7/29/2019 SSR_CH5optimisation

    33/107

    33

    0.8125). By taking the middle point of this interval (L7) as optimum, weobtain

    xopt 0.75 andfopt 0.5625(This solution happens to be the exact solution in this case.)

    FIBONACCI METHOD

    As stated earlier, the Fibonacci method can be used to find the minimumof a function of one variable even if the function is not continuous. Thismethod, like many other elimination methods, has the followinglimitations:

    1. The initial interval of uncertainty, in which the optimum lies, has to beknown.

  • 7/29/2019 SSR_CH5optimisation

    34/107

    34

    2. The function being optimized has to be unimodal in the initial interval ofuncertainty.

    3. The exact optimum cannot be located in this method. Only an intervalknown as the final interval of uncertainty will be known. The final intervalof uncertainty can be made as small as desired by using morecomputations.

    4. The number of function evaluations to be used in the search or theresolution required has to be specified beforehand.

    This method makes use of the sequence of Fibonacci numbers, {Fn}, forplacing the experiments. These numbers are defined as

    F0 = F1 = 1

    Fn = Fn1 + Fn2, n = 2, 3, 4, . . .

  • 7/29/2019 SSR_CH5optimisation

    35/107

    35

    which yield the sequence 1, 1, 2, 3, 5, 8, 13, 21, 34, 55, 89,. . . .

    Procedure:

    LetL0be the initial interval of uncertainty defined by a x b and n be thetotal number of experiments to be conducted. Define

    (5.4)and place the first two experiments at points x1 and x2, which are located

    at a distance ofL2from each end ofL0. This gives

  • 7/29/2019 SSR_CH5optimisation

    36/107

    36

    (5.5)

    Discard part of the interval by using the unimodality assumption. Thenthere remains a smaller interval of uncertaintyL2 given by

    (5.6)and with one experiment left in it. This experiment will be at a distance of

  • 7/29/2019 SSR_CH5optimisation

    37/107

    37

    (5.7)from one end and

    (5.8)from the other end. Now place the third experiment in the interval L2 sothat the current two experiments are located at a distance of

    (5.9)

  • 7/29/2019 SSR_CH5optimisation

    38/107

    38

    from each end of the interval L2. Again the unimodality property will allowus to reduce the interval of uncertainty toL3given by

    (5.10)This process of discarding a certain interval and placing a newexperiment in the remaining interval can be continued, so that the locationof the j th experiment and the interval of uncertainty at the end of jexperiments are, respectively, given by

    (5.11)

  • 7/29/2019 SSR_CH5optimisation

    39/107

    39

    (5.12)The ratio of the interval of uncertainty remaining after conductingj of the npredetermined experiments to the initial interval of uncertainty becomes

    (5.13)and forj = n, we obtain

    (5.14)The ratio Ln/L0 will permit us to determine n, the required number ofexperiments, to achieve any desired accuracy in locating the optimum

    0

  • 7/29/2019 SSR_CH5optimisation

    40/107

    40

    point. Table 5.2 gives the reduction ratio in the interval of uncertaintyobtainable for different number of experiments.

    Position of the Final Experiment:

    In this method the last experiment has to be placed with some care.

    (5.15)

    Thus after conducting n 1 experiments and discarding the appropriateinterval in each step, the remaining interval will contain one experimentprecisely at its middle point.

    41

  • 7/29/2019 SSR_CH5optimisation

    41/107

    41

    However, the final experiment, namely, the nth experiment, is also to beplaced at the center of the present interval of uncertainty. That is, the

    position of the nth experiment will be same as that of (n 1)th one, andthis is true for whatever value we choose for n. Since no new informationcan be gained by placing the nth experiment exactly at the same locationas that of the (n 1)th experiment, we place the nth experiment veryclose to the remaining valid experiment, as in the case of thedichotomous search method. This enables us to obtain the final interval of

    uncertainty to within 0.5Ln-1A flowchart for implementing the Fibonacci method of minimization isgiven in Fig. 5.9.

    42

  • 7/29/2019 SSR_CH5optimisation

    42/107

    42

    43

  • 7/29/2019 SSR_CH5optimisation

    43/107

    43

    44

  • 7/29/2019 SSR_CH5optimisation

    44/107

    44

    45

  • 7/29/2019 SSR_CH5optimisation

    45/107

    45

    Example:

    Minimize f(x) = 0.65 [0.75/(1 +x2)] 0.65x tan1(1/x) in the interval[0,3] by the Fibonacci method using n = 6.

    Solution:

    Here n = 6andL0 = 3.0, which yield

    46

  • 7/29/2019 SSR_CH5optimisation

    46/107

    46

    Thus the positions of the first two experiments are given by x1 = 1.153846andx2= 3.0 1.153846 = 1.846154 with f1 =f (x1)= 0.207270 and f2 =f(x2) = 0.115843.

    Since f1 is less than f2, we can delete the interval [x2, 3.0] by using theunimodality assumption (Fig. 5.10a).

    The third experiment is placed atx3 = 0 + (x2x1) = 1.846154 1.153846= 0.692308, with the corresponding function value off3= 0.291364.

    Sincef1 >f3, we delete the interval [x1,x2] (Fig. 5.10b).

    47

  • 7/29/2019 SSR_CH5optimisation

    47/107

    47

    The next experiment is located atx4 = 0 + (x1x3) = 1.153846 0.692308= 0.461538 withf4 = 0.309811. Nothing thatf4 < f3, we delete the interval

    [x3,x1] (Fig. 5.10c).

    The location of the next experiment can be obtained as x5 = 0 + (x3 x4) =0.692308 0.461538 = 0.230770 with the corresponding objectivefunction value off5 = 0.263678.

    Sincef5 >f4, we delete the interval [0, x5] (Fig. 5.10d).

    The final experiment is positioned at x6 = x5 + (x3 x4) = 0.230770 +(0.692308 0.461538) = 0.461540 with f6 = 0.309810. (Note that,theoretically, the value of x6should be same as that of x4; however, it isslightly different fromx

    4

    , due to round-off error).Since f6 >f4, we delete the interval [x6, x3] and obtain the final interval ofuncertainty as L6 = [x5, x6] = [0.230770, 0.461540] (Fig. 5.10e). The ratioof the final to the initial interval of uncertainty is

    48

  • 7/29/2019 SSR_CH5optimisation

    48/107

    48

    This value can be compared with Eq. (5.14), which states that if nexperiments (n = 6) are planned, a resolution no finer than 1/Fn = 1/F6 =1/13 = 0.076923 can be expected from the method.

    49

  • 7/29/2019 SSR_CH5optimisation

    49/107

    50

  • 7/29/2019 SSR_CH5optimisation

    50/107

    51

  • 7/29/2019 SSR_CH5optimisation

    51/107

    52

  • 7/29/2019 SSR_CH5optimisation

    52/107

    53

  • 7/29/2019 SSR_CH5optimisation

    53/107

    54

  • 7/29/2019 SSR_CH5optimisation

    54/107

    55

  • 7/29/2019 SSR_CH5optimisation

    55/107

    56

  • 7/29/2019 SSR_CH5optimisation

    56/107

    GOLDEN SECTION METHOD

    The golden section method is same as the Fibonacci method except thatin the Fibonacci method the total number of experiments to be conductedhas to be specified before beginning the calculation, whereas this is notrequired in the golden section method. In the Fibonacci method, the

    location of the first two experiments is determined by the total number ofexperiments, N. In the golden section method we start with theassumption that we are going to conduct a large number of experiments.Of course, the total number of experiments can be decided during thecomputation. The intervals of uncertainty remaining at the end of differentnumber of experiments can be computed as follows:

    57

  • 7/29/2019 SSR_CH5optimisation

    57/107

    (5.16)

    (5.17)This result can be generalized to obtain

    (5.18)

    58

  • 7/29/2019 SSR_CH5optimisation

    58/107

    Using the relation

    (5.19)we obtain, after dividing both sides by FN1,

    (5.20)By defining a ratio as

    (5.21)

    So,

    59

  • 7/29/2019 SSR_CH5optimisation

    59/107

    that is,

    (5.22)This gives the root = 1.618, and hence

    (5.23)the ratios FN2/FN1 and FN1/FN have been taken to be same forlarge values of N. The validity of this assumption can be seen from the

    following table:

    60

  • 7/29/2019 SSR_CH5optimisation

    60/107

    The ratio has a historical background. Ancient Greek architects believedthat a building having the sides d and b satisfying the relation

    (5.24)would have the most pleasing properties (Fig. 5.11). The origin of thename, golden section method, can also be traced to the Euclidsgeometry. In Euclids geometry, when a line segment is divided into twounequal parts so that the ratio of the whole to the larger part is equal to

    61

  • 7/29/2019 SSR_CH5optimisation

    61/107

    the ratio of the larger to the smaller, the division is called the goldensection and the ratio is called the golden mean.

    Procedure:

    62

  • 7/29/2019 SSR_CH5optimisation

    62/107

    The procedure is same as the Fibonacci method except that the locationof the first two experiments is defined by

    The desired accuracy can be specified to stop the procedure.

    Example:

    Minimize the function

    using the golden section method with n = 6.

    Solution:

    63

  • 7/29/2019 SSR_CH5optimisation

    63/107

    The locations of the first two experiments are defined by

    L

    2= 0.382L0 = (0.382)(3.0) = 1.1460.

    Thusx1 = 1.1460 andx2= 3.0 1.1460 = 1.8540

    withf1 =f (x1) = 0.208654 andf2 =f (x2) = 0.115124.

    Since f1 < f2, we delete the interval [x2, 3.0] based on the assumption ofunimodality and obtain the new interval of uncertainty as L2 = [0, x2] =[0.0, 1.8540].

    The third experiment is placed at x3 = 0 + (x2x1) = 1.8540 1.1460 =

    0.7080.

    64

  • 7/29/2019 SSR_CH5optimisation

    64/107

    Since f3 = 0.288943 is smaller than f1 = 0.208654, we delete theinterval [x1,x2] and obtain the new interval of uncertainty as [0.0, x1] = [0.0,

    1.1460].The position of the next experiment is given by x4 = 0 + (x1x3) = 1.1460 0.7080 = 0.4380 withf4= 0.308951.

    Since f4 < f3, we delete [x3, x1] and obtain the new interval of uncertaintyas [0,x3] = [0.0, 0.7080].

    The next experiment is placed at x5 = 0 + (x3x4) = 0.7080 0.4380 =0.2700. Sincef5= 0.278434 is larger thanf4= 0.308951, we delete theinterval [0, x5] and obtain the new interval of uncertainty as [x5, x3] =[0.2700, 0.7080].

    The final experiment is placed at x6 = x5 + (x3x4) = 0.2700 + (0.7080 0.4380) = 0.5400 withf6= 0.308234.

    65

  • 7/29/2019 SSR_CH5optimisation

    65/107

    Since f6 >f4, we delete the interval [x6, x3] and obtain the final interval ofuncertainty as [x5, x6] = [0.2700, 0.5400]. Note that this final interval of

    uncertainty is slightly larger than the one found in the Fibonacci method,[0.461540, 0.230770]. The ratio of the final to the initial interval ofuncertainty in the present case is

    66

  • 7/29/2019 SSR_CH5optimisation

    66/107

    67

  • 7/29/2019 SSR_CH5optimisation

    67/107

    68

  • 7/29/2019 SSR_CH5optimisation

    68/107

    CUBIC INTERPOLATION METHOD

    69

  • 7/29/2019 SSR_CH5optimisation

    69/107

    70

  • 7/29/2019 SSR_CH5optimisation

    70/107

    71

  • 7/29/2019 SSR_CH5optimisation

    71/107

    72

  • 7/29/2019 SSR_CH5optimisation

    72/107

    Example:

    Find the minimum of f = 5 53 20 + 5 by the cubic interpolationmethod.

    Solution:

    Since this problem has not arisen during a multivariable optimizationprocess, we can skip stage 1. We take A = 0 and find that

    To findB at which df/d is nonnegative, we start with t0 = 0.4 and evaluatethe derivative at t0, 2t0, 4t0, . . . . This gives

    73

  • 7/29/2019 SSR_CH5optimisation

    73/107

    Thus we find that

    Iteration 1:

    74

  • 7/29/2019 SSR_CH5optimisation

    74/107

    To find the value of and to test the convergence criteria, we firstcompute Z and Q as

    Hence,

    By discarding the negative value, we have

    75

  • 7/29/2019 SSR_CH5optimisation

    75/107

    Convergence criterion:

    If is close to the true minimum, , then shouldbe approximately zero. Sincef = 54152 20,

    Since this is not small, we go to the next iteration or refitting. As

    we take

    A =

    76

  • 7/29/2019 SSR_CH5optimisation

    76/107

    Iteration 2

    77

  • 7/29/2019 SSR_CH5optimisation

    77/107

    Convergence criterion:

    Since this value is large, we go the next iteration with

    78

  • 7/29/2019 SSR_CH5optimisation

    78/107

    Iteration 3

    79

    Convergence criterion:

  • 7/29/2019 SSR_CH5optimisation

    79/107

    g

    Assuming that this value is close to zero, we can stop the iterativeprocess and take

    80

  • 7/29/2019 SSR_CH5optimisation

    80/107

    Newton Method

    Consider the quadratic approximation of the function f () at = i usingthe Taylors series expansion:

    (5.25)

    By setting the derivative of Eq. (5.25) equal to zero for the minimum of f(), we obtain

    (5.26)

    81

    If i denotes an approximation to the minimum of f(), Eq. (5.26) can be

  • 7/29/2019 SSR_CH5optimisation

    81/107

    pp f( ) q ( )rearranged to obtain an improved approximation as

    (5.27)Thus the Newton method, Eq. (5.27), is equivalent to using a quadraticapproximation for the function f() and applying the necessary conditions.The iterative process given by Eq. (5.65) can be assumed to have

    converged when the derivative,f (i+1), is close to zero:

    (5.28)where is a small quantity. The convergence process of the method is

    shown graphically in Fig. 5.18a.Remarks:

    82

    1. The Newton method was originally developed by Newton for solving

  • 7/29/2019 SSR_CH5optimisation

    82/107

    nonlinear equations and later refined by Raphson, and hence the methodis also known as NewtonRaphson method in the literature of numerical

    analysis.2. The method requires both the first- and second-order derivatives off().

    3.f (i ) = 0 [in Eq. (5.27)], the Newton iterative method has a powerful(fastest) convergence property, known as quadratic convergence.

    4. If the starting point for the iterative process is not close to the true

    solution , the Newton iterative process might diverge as illustrated inFig. 5.18b.

    83

  • 7/29/2019 SSR_CH5optimisation

    83/107

    84

  • 7/29/2019 SSR_CH5optimisation

    84/107

    85

  • 7/29/2019 SSR_CH5optimisation

    85/107

    Example:

    Find the minimum of the function

    using the NewtonRaphson method with the starting point 1 = 0.1. Use = 0.01 in Eq. (5.66) for checking the convergence.

    Solution:

    The first and second derivatives of the functionf() are given by

    86

  • 7/29/2019 SSR_CH5optimisation

    86/107

    Iteration 1

    1 = 0.1,f (1)= 0.188197,f (1)= 0.744832,f (1) = 2.68659

    Convergence check: |f (2)| = |0.138230|>.

    Iteration 2

    87

    f ( ) 0 303279 f ( ) 0 138230 f ( ) 1 57296

  • 7/29/2019 SSR_CH5optimisation

    87/107

    f (2)= 0.303279,f (2)= 0.138230,f (2) = 1.57296

    Convergence check: |f (3)| = |0.0179078|>.

    Iteration 3

    f (3) = 0.309881,f (3) = 0.0179078,f (3) = 1.17126

    Convergence check: |f (4)| = |0.0005033| < .

    88

    Since the process has converged, the optimum solution is taken as4

  • 7/29/2019 SSR_CH5optimisation

    88/107

    = 0.480409.

    Quasi-Newton Method

    If the function being minimized f () is not available in closed form or isdifficult to differentiate, the derivatives f () andf () in Eq. (5.27) can be

    approximated by the finite difference formulas as

    (5.29)

    (5.30)

    89

    where is a small step size. Substitution of Eqs. (5.29) and (5.30) into Eq.(5 27) leads to

  • 7/29/2019 SSR_CH5optimisation

    89/107

    (5.27) leads to

    (5.31)The iterative process indicated by Eq. (5.69) is known as the quasi-Newton method. To test the convergence of the iterative process, the

    following criterion can be used:

    (5.32)where a central difference formula has been used for evaluating thederivative of f and is a small quantity.

    90

    Remarks:

  • 7/29/2019 SSR_CH5optimisation

    90/107

    Remarks:

    1. The central difference formulas have been used in Eqs. (5.31) and(5.32). However, the forward or backward difference formulas can also beused for this purpose.

    2. Equation (5.31) requires the evaluation of the function at the pointsi+and

    i in addition to

    iin each iteration.

    Example:

    Find the minimum of the function

    91

    using quasi-Newton method with the starting point 1 = 0.1 and the stepsize = 0 01 in central difference formulas Use = 0 01 in Eq (5 32) for

  • 7/29/2019 SSR_CH5optimisation

    91/107

    size = 0.01 in central difference formulas. Use = 0.01 in Eq. (5.32) forchecking the convergence.

    Solution:

    Iteration 1

    1 = 0.1, = 0.01, = 0.01, f1 = f (1) = 0.188197,f + 1 = f (1 + ) =0.195512,f 1 = f(1)= 0.180615

    92

  • 7/29/2019 SSR_CH5optimisation

    92/107

    Convergence check:

    Iteration 2f2= f(2)= 0.303368,f+2 =f(2+)=0.304662,f2=f(2)=0.301916

    93

    Convergence check:

  • 7/29/2019 SSR_CH5optimisation

    93/107

    Convergence check:

    Iteration 3

    f3= f(3) = 0.309885, f+3 = f(3+) = 0.310004, f3 = f(3) =0.309650

    94

    Convergence check:

  • 7/29/2019 SSR_CH5optimisation

    94/107

    Since the process has converged, we take the optimum solution as 4

    = 0.480600

    Secant Method

    The secant method uses an equation similar to Eq. (5.26) as

    95

  • 7/29/2019 SSR_CH5optimisation

    95/107

    (5.33)where s is the slope of the line connecting the two points (A,f(A)) and (B,f(B)), where A and B denote two different approximations to the correctsolution,. The slope s can be expressed as (Fig. 5.19)

    (5.34)Equation (5.33) approximates the function f () between A and B as alinear equation (secant), and hence the solution of Eq. (5.33) gives thenew approximation to the root off () as

    96

  • 7/29/2019 SSR_CH5optimisation

    96/107

    (5.35)The iterative process given by Eq. (5.35) is known as the secant method(Fig. 5.19). Since the secant approaches the second derivative of f () atA as B approaches A, the secant method can also be considered as aquasi-Newton method. It can also be considered as a form of eliminationtechnique since part of the interval, (A,

    i+1) in Fig. 5.19, is eliminated in

    every iteration. The iterative process can be implemented by using thefollowing step-by-step procedure.

    1. Set1 = A = 0 and evaluatef (A). The value off (A) will be negative.Assume an initial trial step length t0. Set i = 1.

    2. Evaluatef (t0).

    97

    3. Iff (t0) < 0, set A =i = t0,f (A) =f (t0), new t0 = 2t0, and go to step 2.4. Iff (t0) 0, set B = t0,f (B) =f (t0), and go to step 5.

  • 7/29/2019 SSR_CH5optimisation

    97/107

    f ( 0) , 0, f ( ) f ( 0), g p

    5. Find the new approximate solution of the problem as

    (5.36)6. Test for convergence:

    |f (i + 1)| (5.37)

    where is a small quantity. If Eq. (5.37) is satisfied, take i+1 and stop

    the procedure. Otherwise, go to step 7.

    98

    7. Iff (i+1) 0, set new B =i+1,f (B) =f (i+1), i = i + 1, and go to step 5.

  • 7/29/2019 SSR_CH5optimisation

    98/107

    8. Iff (i+1) < 0, set new A =i+1,f (A) =f (i+1), i = i + 1, and go to step 5.

    Remarks:

    1. The secant method is identical to assuming a linear equation for f ().This implies that the original function,f (), is approximated by a quadraticequation.

    2. In some cases we may encounter a situation where the function f ()varies very slowly with , as shown in Fig. 5.20. This situation can beidentified by noticing that the point B remains unaltered for severalconsecutive refits. Once such a situation is suspected, the convergence

    process can be improved by taking the next value of i+1 as (A + B)/2instead of finding its value from Eq. (5.36).

    99

  • 7/29/2019 SSR_CH5optimisation

    99/107

    100

  • 7/29/2019 SSR_CH5optimisation

    100/107

    101

  • 7/29/2019 SSR_CH5optimisation

    101/107

    102

    Example:

  • 7/29/2019 SSR_CH5optimisation

    102/107

    Find the minimum of the function

    using the secant method with an initial step size of t0 = 0.1,1= 0.0, and

    = 0.01.

    Solution:

    1 = A = 0.0, t0 = 0.1,f (A)= 1.02102, B = A + t0= 0.1,f (B)= 0.744832.

    Sincef (B) < 0, we set new A = 0.1,f (A) = 0.744832, t0= 2(0.1) = 0.2, B=1 + t0= 0.2, and computef (B) = 0.490343.

    103

    Sincef (B) < 0, we set new A = 0.2,f (A) = 0.490343, t0 = 2(0.2) = 0.4, B=1 + t0= 0.4, and computef (B) = 0.103652.

  • 7/29/2019 SSR_CH5optimisation

    103/107

    Sincef (B) < 0, we set new A = 0.4,f (A) =0.103652, t0 = 2(0.4) = 0.8, B=1 + t0= 0.8, and computef (B) = +0.180800.

    Sincef (B)>0, we proceed to find 2.

    Iteration 1

    Since A = 1 = 0.4, f (A) = 0.103652,B = 0.8, f (B) = +0.180800, wecompute

    Convergence check:|f (2)| = |+0.0105789|>.

    104

    Iteration 2

  • 7/29/2019 SSR_CH5optimisation

    104/107

    Since f (2) = +0.0105789>0, we set new A = 0.4, f (A) = 0.103652,B =2 = 0.545757,f (B) =f (2) = +0.0105789, and compute

    Convergence check: |f (3)| = |+0.00151235| < .

    Since the process has converged, the optimum solution is given by3

    = 0.490632.

    105

  • 7/29/2019 SSR_CH5optimisation

    105/107

    PRACTICAL CONSIDERATIONS

    How to Make the Methods Efficient and More Reliable

    In some cases, some of the interpolation methods discussed may be veryslow to converge, may diverge, or may predict the minimum of thefunction,f(), outside the initial interval of uncertainty, especially when theinterpolating polynomial is not representative of the variation of thefunction being minimized. In such cases we can use the Fibonacci or

    golden section method to find the minimum. In some problems it mightprove to be more efficient to combine several techniques. For example,the unrestricted search with an accelerated step size can be used to

    106

    bracket the minimum and then the Fibonacci or the golden sectionmethod can be used to find the optimum point. In some cases theFibonacci or golden section method can be used in conjunction with an

  • 7/29/2019 SSR_CH5optimisation

    106/107

    Fibonacci or golden section method can be used in conjunction with an

    interpolation method.

    Comparison of Methods

    It has been shown in Section 5.9 that the Fibonacci method is the most

    efficient elimination technique in finding the minimum of a function if theinitial interval of uncertainty is known. In the absence of the initial intervalof uncertainty, the quadratic interpolation method or the quasi-Newtonmethod is expected to be more efficient when the derivatives of thefunction are not available. When the first derivatives of the function being

    minimized are available, the cubic interpolation method or the secantmethod are expected to be very efficient. On the other hand, if both thefirst and second derivatives of the function are available, the Newton

    107

    method will be the most efficient one in finding the optimal step length,.

    In general, the efficiency and reliability of the various methods are

  • 7/29/2019 SSR_CH5optimisation

    107/107

    g , y yproblem dependent and any efficient computer program must includemany heuristic additions not indicated explicitly by the method. Theheuristic considerations are needed to handle multimodal functions(functions with multiple extreme points), sharp variations in the slopes(first derivatives) and curvatures (second derivatives) of the function, andthe effects of round-off errors resulting from the precision used in the

    arithmetic operations.