SlideShare a Scribd company logo
1 of 46
Download to read offline
Optimization
Optimization
Dr Reshma Khemchandani
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
OUTLINE
1 UNCONSTRAINED OPTIMIZATION PROBLEMS
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
An optimization problem where decision vector x is allowed to
take any value in Rn
is called an unconstrained optimization
problem (UMP).
Aim is to study standard algorithms for unconstrained
minimization of functions of one variable as well as functions of
several variables.
Consider the unconstrained minimization problem (UMP)
Min
x∈Rn
f(x)
to get a point x ∈ Rn
such that f(x) ≤ f(x) for all x ∈ Rn
.
Unfortunately, finding local minimum for general function is very
difficult task.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Solution Set: Let f : Rn
−→ R be continuously differentiable.
Then the set Ω = { x ∈ Rn
: ∇f(x) = 0 } is called the solution
set of (1).
Let f : Rn
−→ R be differentiable convex function and x be a
point of the solution set. Then x is a global min point of UMP.
We now describe a common basic scheme of the form
x(k+1)
= x(k)
+ αk d(k)
for solving the UMP where x(k)
is the current solution, d(k)
is the
direction of movement from x(k)
and αk > 0 (called the step
size) is the distance upto which we move in the direction d(k)
from the current point x(k)
.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Descent Property: An algorithm for solving the UMP is said to
have the descent property if the objective function value
decreases as we go through the sequence {x(k)
}, i.e.
f(x(k+1)
) < f(x(k)
) for all k.
In other words, for the algorithm to possess the descent
property, the objective function should decrease as we proceed.
Quadratic Termination Property: An algorithm for UMP is said to
have quadratic termination property (q.t.p) if the minimum of a
positive definite quadratic form in n variables is reached in at
most n iterations.
The motivation for defining q.t.p stems from the fact that near a
local min point, the function behaves like a strictly convex
function(like a parabola in R or a positive definite quadratic form
in Rn
) and therefore if the algorithm behaves well on such a
function, it will ‘hopefully’ do well on the other functions as well.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Globally Convergent: An algorithm for the UMP is said to be
globally convergent if starting from any point x(0)
∈ Rn
, the
sequence {x(k)
} always converges to a point of the solution set
Ω.
The property of ‘global convergence’ guarantees that no matter
from any arbitrary point x(0)
∈ Rn
we start, we are guaranteed to
generate a sequence {x(k)
} that converges to a point of the
solution set. Here it must be noted that two different starting
points will, in general, generate two different sequences of
iterates but both converging to a point of solution set.
Order of Convergence: Let the sequence {x(k)
} converge to a
point x and let x(k)
6= x for sufficiently large k. The quantity
kx(k)
− xk is called the error of the kth
iterate x(k)
.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Suppose that there exists p and 0 < a < ∞ such that
lim
k→∞
||x(k+1)
− x||
||x(k) − x||p
= a (0 < a < ∞)
then p is called the order of convergence of the sequence {x(k)
}.
Thus kx(k+1)
− xk = a||x(k)
− x||p
asymptotically.
If p = 1, the sequence {x(k)
} is said to have linear convergence
rate and for p = 2, it is said to have quadratic convergence rate.
In case p = 1 but a = 0, then the sequence {x(k)
} is said to
have super linear convergence rate.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
LINE SEARCH METHODS FOR UNIMODAL
FUNCTIONS
Unimodal Min Function: The function f : [a, b] −→ R is said to be
a unimodal (to be specific unimodal min) function if it has only
one mode i.e. it has a single relative min, i.e. ∃ a ≤ α ≤ b such
that
(i) f is strictly decreasing (↓) in [a, α).
(ii) f is strictly increasing (↑) in [α, b].
Similar definition hold for unimodal max function.
Unimodal function may not be differentiable - in fact it may not
even be continuous
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
PLOT OF UNI-MODAL FUNCTIONS
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Basic Strategy
Let us choose two distinct points (say) x1 and x2 in [a, b] i.e.
a ≤ x1 < x2 ≤ b.
Let xmin denote the point giving the minimum value of f(x) over
[a, b].
Since f is uni-modal min function, it is clear that
(i) f(x1) < f(x2) ⇒ xmin ∈ [a, x2]
(ii) f(x1) > f(x2) ⇒ xmin ∈ [x1, b]
(iii) f(x1) = f(x2) ⇒ xmin ∈ [x1, x2] .
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
LINE SEARCH METHODS
(I) The Golden Section Rule.
(II) The Fibonacci Search Method.
In discussing the aforesaid methods we shall make use of the
following notations
xL,k = lower limit of the search interval at the kth
iteration (so xL,1 = a)
xU,k = upper limit of the search interval at the kth
iteration (so xU,1 = b)
xp,k = 1st
trial point at the kth
iteration
xq,k = 2nd
trial point at the kth
iteration
Ep,k = f(xp,k ) value of the function at the 1st
trial point
Eq,k = f(xq,k ) value of the function at the 2nd
trial point
Ik = xU,k − xL,k = length of the search interval at the kth
iteration
(so I1 = (b - a))
IL
k = length of the left part of the search interval Ik
IR
k = length of the right part of the search interval Ik .
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
x x
p,k
L,k
xq,k= xp,k+1 xq,k+1 x
U,k
I
I
Ik+1
k+1
k+2
I
k+2
L
L
R
R
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
THE GOLDEN SECTION RULE
Here the two trial points xp,k and xq,k are chosen as per the following
criteria
(I) IL
k = IR
k for all k , i.e. no undue advantage is given to left or right
part.
(II) xp,k+1 = xq,k and xq,k+1 is computed afresh for the interval IR
k+1 (
or xq,k+1 = xp,k and xp,k+1 is computed afresh for the interval
IL
k+1), i.e. only one new trial point is computed at the kth
iteration,
the other trial point is obtained from the previous iteration. In
view of this we have
Ik = Ik+1 + Ik+2 for all k.
(III)
Ik
Ik+1
=
Ik+1
Ik+2
= c (constant) for all k. This is called the golden
section criterion.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Thus we wish to choose trial points xp,k and xq,k according to
criterion (i), (ii) and (iii). From (ii) we have
Ik = Ik+1 + Ik+2
i.e.
Ik
Ik+2
= Ik+1
Ik+2
+ 1
i.e.
Ik
Ik+1
Ik+1
Ik+2
= Ik+1
Ik+2
+ 1 (by (iii))
i.e.
c2
= c + 1. Therefore c = 1±
√
5
2
Since c can not be negative, we get c =
1 +
√
5
2
= 1.618 (golden
section ratio and
1
1.618
= 0.618 .
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
STEPWISE DESCRIPTION
Step 1 Input data- xL,1, xU,1, , f (Here   0 is the tolerance to be
prescribed by the user).
Step 2 Compute the first two trial points xp,1 and xq,1, where
xp,1 = xU,1 − 0.618 (xU,1 − xL,1)
xq,1 = xL,1 + 0.618 (xU,1 − xL,1).
Set k = 1.
Step 3 Evaluate the function f at the two trial points xp,k and xq,k . Let
Ep,k = f(xp,k )
Eq,k = f(xq,k ) .
Step 4 Test the interval which contains the minimum, i.e. if Ep,k ≤ Eq,k ,
go to Step 5 otherwise go to Step 6.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Step 5 Use following relations to update the data
xL,k+1 = xL,k
xU,k+1 = xq,k
xp,k+1 = xU,k+1 - 0.618 Ik+1
xq,k+1 = xp,k
Ep,k+1 = f(xp,k+1)
Eq,k+1 = f(xq,k+1) = Ep,k .
Step 6 Use the following relations to update the data
xL,k+1 = xp,k
xU,k+1 = xU,k
xp,k+1 = xq,k
xq,k+1 = xL,k + 0.618 Ik+1
Ep,k+1 = f(xp,k+1) = Eq,k
Eq,k+1 = f(xq,k+1) .
Step 7 Test for the end of optimization , i.e. if Ik  , go to Step 8,
otherwise set k = (k+1) and go to Step 4.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Step 8 Output xL,n, xU,n. Then xmin ∈ (xL,n, xU,n) and
Emin ≤ Min(Ep,n−1, Eq,n−1).
It can be noted that for the golden section rule,
In
I1
=

1
1.618
n−1
= (0.618)n−1
. Thus knowing I1 and In we can
find out the number of iterations as well as the number of points
at which the function is to be evaluated.
Note that if we are going upto the 7th
iteration (i.e. getting xL,7
and xU,7) then we shall be having n = 7 functional evaluations,
namely, at xp,1, xq,1, and one each at 2nd
, 3rd
, 4th
, 5th
and 6th
iteration. To stop at the 7th
iteration we shall be computing xp,6
and xq,6 and Emin ≤ Min (xp,6, xq,6).
Thus N functional evaluations will mean stopping at the Nth
iteration and computing points upto xp,N−1 and xq,N−1. Also
Emin ≤ Min (xp,N−1, xq,N−1) and there are only (N − 1) interval
reductions.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Find min x2
over [-5, 15] by the golden section rule. Take  = 1.5.
k xL,k xU,k xp,k xq,k Ep,k Eq,k L/R
1 - 5.0 15.00 2.64 7.36 6.96 54.1 L
 
2 -5.0 7.36 -0.2 2.64 0.077 6.96 L
 
3 -5.0 2.64 -2.08 -0.27 4.33 0.077 R
. .
4 -2.08 2.64 -0.2 0.84 0.077 0.71 L
 
5 -2.08 0.84 -0.96 -0.27 0.92 0.077 R
. .
6 -0.96 0.84 -0.27 0.15 0.077 0.023 R
7 -0.27 0.84
I7 = 1.11  1.5(= ), stop, xmin ∈ [-0.27, 0.84] Emin ≤ Min (0.077,
0.023) = 0.023.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
The Fibonacci Search Method
Fibonacci Sequence: Let F0 = 1, F1 = 1 and Fi =
Fi−1 + Fi−2 (i ≥ 2). Then {Fn} is called the sequence of
Fibonacci numbers or in short, the Fibonacci sequence.
Thus, the Fibonacci sequence is {1, 1, 2, 3, 5, 8, 13, 21, . . .}
where F0 = 1, F1 = 1, F2 = 2, F3 = 3, F4 = 5, F5 = 8, F6 = 8,
F7 = 21... and so on.
(I) IL
k = IR
k for all k
(II) Ik = Ik+1 + Ik+2
(III)
Ik
Ik+1
=
Fn−k+1
Fn−k
, where n denotes the number of iterations
to be performed.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
We now give a justification of the condition (iii) given above. If we
wish to stop at the nth
iteration, then
In+1 ' In = 1 In
In = In = 1 In
In−1 = In + In+1 = 2 In
In−2 = In + In−1 = 3 In
In−3 = 5 In
In−4 = 8 In
.
.
.
I1 = FnIn.
In the golden section rule Ik
Ik+1
= c = 1.618 for all k and all n. In
Fibonacci search method this ratio depends on both the current
iteration k and also the total number of iterations n to be performed.
Therefore instead of 0.618 (i.e.
1
c
) we shall be using the number
Fn−k
Fn−k+1
for determining the points xp,k and xq,k .
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
STEPWISE APPROACH OF THE FIBONACCI
SEARCH METHOD
Step 1 Input data: xL,1, xU,1, n, f.
Step 2 Compute Fibonacci numbers Fn and Fn−1 and find
xp,1 = xU,1 -
Fn−1
Fn
(xU,1 − xL,1)
xq,1 = xL,1 +
Fn−1
Fn
(xU,1 − xL,1)
Set k = 1.
Step 3 Evaluate
Ep,k = f(xp,k )
Eq,k = f(xq,k )
and test the interval which contains the minimum. If Ep,k ≤ Eq,k ,
go to Step 4, otherwise go to Step 5.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Step 4 Use the following relations to update the data
xL,k+1 = xL,k
xU,k+1 = xq,k
xp,k+1 = xU,k+1 -
Fn−k
Fn−k+1
Ik+1
xq,k+1 = xp,k
Ep,k+1 = f(xp,k+1)
Eq,k+1 = f(xq,k+1) = Ep,k .
Step 5 Use the following relations to update the data
xL,k+1 = xp,k
xU,k+1 = xU,k
xp,k+1 = xq,k+1
xq,k+1 = xL,k +
Fn−k
Fn−k+1
Ik+1
Ep,k+1 = f(xp,k+1) = Eq,k
Eq,k+1 = f(xq,k+1) .
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Step 6 If k ≤ (n − 2), then set k = k + 1 and go to step 3. If k = (n − 1)
then go to Step 7.
Step 7 As F0 = F1 = 1, it can be seen that at the (n − 1)th
iteration, the
two trial points xp,n−1 and xq,n−1 will come out to be the same ,
i.e. xp,n−1 = xq,n−1. To make them distinct we have to add /
subtract a fixed positive number  (say 0.01) to one of these. For
this we look at the (n − 2)th
iteration. If there (i.e. at the (n − 2)th
iteration) we are searching in the left then  is subtracted from
xp,n−1 otherwise it is added to xq,n−1. We then determine new
xp,n−1 and xq,n−1 and hence xL,n and xU,n. Then
xmin ∈ (xL,n, xU,n)
and fmin ≤ Min(Ep,n−1, Eq,n−1)
As
In
I1
=
1
Fn
, so given I1 and In, the value of n can be computed in
advance.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
EXAMPLE
With n = 7, find min x2
over [-5,15] via Fibonacci search method.
k Fn−k /Fn−k+1 xL,k xU,k xp,k xq,k Ep,k Eq,k
1 13/21 -5.0 15.00 2.64 7.38 6.88 54.6
 
2 8/13 -5.0 7.38 -0.24 2.62 0.058 6.88
 
3 5/8 -5.0 2.62 -2.14 -0.24 4.67 0.058
. .
4 3/5 -2.14 2.62 -0.24 0.72 0.058 0.52
 
5 2/3 -2.14 0.72 -1.2 -0.24 1.44 0.058
. .
6 1/2 - 1.2 0.72 -0.24 -0.24+0.01 0.058 0.053
7 1 -0.24 0.72
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
From this table we observe that xmin ∈ [−0.24, 0.72] and
fmin ≤ 0.053.
In the given example, I1 = 20 and therefore if we take  = 1.5,
then In = 1.5.
This gives
I1
In
=
20
1.5
= 13.3 which from the sequence of
Fibonacci numbers determines n = 7 as F6 = 13 and F7 = 21.
Also as explained in Step 7, the two trial points at the (n − 1)th
iteration, i.e. xp,6 and xq,6 both becomes equal to 0.24.
Therefore we look at the (n − 2)th
iteration, i.e. the 5th
iteration in
our example. As there we are seeking the right part, we add
0.01 to xq,6 to get new xq,6 and then continue as explained.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
RELATION BETWEEN THE FIBONACCI SEARCH
METHOD AND THE GOLDEN SECTION RULE
For the constant c appearing in the golden section rule we have
c = 1 + 1/c = 1.618.
Fn '
cn+1
√
5
(for large n).
Let RF and RGS denote the reduction factors for the Fibonacci
search method and the golden section rule respectively.
RF =
In
I1
=
1
Fn
=
√
5
cn+1
RGS =
In
I1
=
1
c
n−1
=
1
cn−1
.
Therefore
RGS
RF
=
1
cn−1
×
cn+1
√
5
=
c2
√
5
' 1.17 .
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
THE STEEPEST DESCENT METHOD
The steepest descent method is gradient based methods for
solving UMP.
Simple to implement and therefore used widely in various
applications.
The only drawback of the steepest descent method is its slow
convergence (as its order of convergence is one ).
The UMP: Min
x∈Rn
f(x), f has continuous first order partial
derivatives on Rn
.
Then the basic scheme involved is x(k+1)
= x(k)
+ αk d(k)
where
d(k)
= −∇f(x(k)
)/k∇f(x(k)
)k, is the direction of movement and
the step size αk ≥ 0 is chosen such that h(αk ) = Min
αk ≥0
h(αk ).
Here the function h(αk ) is given by h(αk ) = f(x(k)
+ αk d(k)
),
and we stop when k∇f(x(k)
)k  .
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
A stepwise description of the steepest descent method could be
Step 1 Choose x(0)
∈ Rn
and tolerance   0. Set k = 0.
Step 2 Compute ∇f(x(k)
) and d(k)
= −∇f(x(k)
)/k∇f(x(k)
)k.
Step 3 Evaluate x(k+1)
= x(k)
+ αk d(k)
where αk  0 is chosen such
that h(αk ) = Min
αk ≥0
h(αk ) ,
h(αk ) = f(x(k)
+ αk d(k)
) .
The Steps 2 and 3 above are repeated till k∇f(x(k)
)k  . In that case
x(k)
becomes a point of the solution set and therefore, if f is a convex
function then it becomes an optimal solution.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
The steepest descent algorithm
(I) has descent property
(II) does not have quadratic termination property
(III) is globally convergent and
(IV) has order of convergence p = 1.
Therefore if we are using the steepest descent method then we can
start from any point x(0)
and as we proceed, the objective function
value will decrease, but the algorithm may take lot many iterations
near the optimal solution. Also it may, in general, take more than n
iterations to minimize a positive definite quadratic form of n variables.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Example: Use the steepest descent method to minimize
f(x1, x2) = 3x2
1 − 4x1x2 + 2x2
2 + 4x1 + 6 over (x1, x2) ∈ R2
.
The given function is a convex function and so the steepest descent
method will give a global optimal solution. Starting from x(0)
= (0, 0)T
and following Steps 1-3 described above, we get
k x(k)
∇f(x(k)
) d(k)
αk
1 (0, 0)T
(4, 0)T
(−1, 0)T
2/3
2 (−2/3, 0)T
(0, 8/3)T
(0, −1)T
2/3
3 (−2/3, −2/3)T
(8/3, 0)T
(−1, 0)T
1/6
4 (−10/9, −2/3)T
(0, 16/9)T
(0, −1)T
1/4
5 (−38/27, −10/9)T
(16/9, 0)T
(−1, 0)T
1/6
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
The function f(x1, x2) is a positive definite quadratic form in two
variables but its optimal solution has not been obtained in atmost
two iterations. This illustrates that the method of steepest
descent does not possess quadratic termination property.
We observe that the directions d(k)
are repeated alternately
(−1, 0)T
, (0, −1)T
, (−1, 0)T
, (0, −1)T
etc. Any two consecutive
directions d(k)
and d(k+1)
given by the steepest descent method
are mutually orthogonal. Therefore in R2
, if the first two
directions are d(1)
and d(2)
then d(3)
has to be d(1)
and d(4)
has
to be d(2)
so on. But this repetition may not be in R3
and higher
dimensional spaces because there if d(1)
and d(2)
are
orthogonal then we may have d(3)
, different from d(1)
, which is
orthogonal to d(2)
. So the important thing is the orthogonality of
consecutive directions and NOT their alternate repetition.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
x
x2
1
FIGURE: −∇ f(x(k)
) which is normal to the surface f(x) = constant at
x(k)
, is tangent to the surface at the point x(k+1)
.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
NEWTON’S METHOD
Newton’s method: The basic scheme: yk+1 = yk − g(yk )/g0
(yk )
where yk is the current iterate or the current approximation,
g(y) = 0, y ∈ R.
Looking at the standard basic scheme of Newton’s method
x(k+1)
= x(k)
− (Hf (x(k)
))−1
∇f(x(k)
)
for finding a solution of the system ∇f(x) = 0.
Newton’s method for solving UMP’s has quadratic termination
property (in fact here for minimizing a positive definite quadratic
form of n variables it will take exactly one iteration).
Order of convergence is 2 and it will have descent property, but it
will not have the property of global convergence.
Therefore except for the case when we are minimizing a positive
definite quadratic form, we cannot start the method from an
arbitrary point x(0) ∈ Rn
.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Use Newton’s method to minimize
f(x1, x2) = 8x2
1 − 4x1x2 + 5x2
2 , (x1, x2) ∈ R2
.
As the function f is a positive definite quadratic form we can start
from any arbitrary point x(0)
∈ R2
.
To be specific, let x(0)
= (5, 2)T
. Then
∇f(x(0)
) =

16x1 − 4x2
−4x1 + 10x2

(x1=5,x2=2)
= 72 0

,
Hf (x(0)
) =

16 −4
−4 10

,
(Hf (x(0)
))−1
=
1
144

10 4
4 16

.
x(1)
= x(0)
− (Hf (x(0)
))−1
∇f(x(0)
)
=

5
2

−
1
444

10 4
4 16
 
72
0

=

0
0

,
giving x1 = 0, x2 = 0 as the minimizing point.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
THE CONJUGATE GRADIENT METHOD
Conjugate based methods are better than the steepest descent
method (in terms of order of convergence) and are simpler to
implement than the modified Newton’s method.
Conjugate Directions: Let Q be an (n × n) positive definite
matrix. Any two non-zero vectors (directions) d(1)
, d(2)
∈ Rn
are
said to be conjugate vectors or conjugate directions with respect
to Q, if (d(1)
)T
Q d(2)
= 0.
For Q = I, conjugacy reduces to the usual concept of
orthogonality. Therefore if d(1)
and d(2)
are conjugate with
respect to Q, sometimes we also call them Q - orthogonal.
A set {d(0)
, . . . , d(k)
} of (k + 1) vectors in Rn
is said to be
conjugate if every two of them are so, i.e.
(d(i)
)T
Q d(j)
= 0 (i 6= j).
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Let {d(0)
, d(1)
, . . . , d(k)
} be a set of (k + 1) non-zero vectors
which are conjugate with respect to a given positive definite
matrix Q. Then the vectors d(0)
, d(1)
, . . . , d(k)
are linearly
independent.
Consider the problem Min
x∈Rn
1
2
xT
Q x − bT
x where Q is an
(n × n) symmetric positive definite matrix and hence the problem
is strictly convex and has unique minimizing point x ∈ Rn
.
Further the KKT conditions give ∇

1
2
xT
Qx − bT
x

= 0, i.e.
Qx = b, which implies x = Q−1
b (note that as Q is positive
definite, Q−1
exists), which is equivalent to finding the unique
solution of the system of equations Qx = b.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Let {d(0)
, d(1)
, . . . , d(n−1)
} be a set of n non-zero vectors in Rn
which are conjugate with respect to Q. Then x, the unique
solution to the system Qx = b or equivalently, the unique
minimizing point of problem, is given by
x =
n−1
X
k=0

(d(k)
)T
b
(d(k))T Qd(k)

d(k)
.
Conjugate Direction Theorem: Let {d(0)
, d(1)
, . . . , d(n−1)
} be a
set of n non-zero vectors in Rn
which are conjugate with respect
to Q. For any x(0)
∈ Rn
, the sequence {x(k)
} is generated
x(k+1)
= x(k)
+ αk d(k)
,
αk = −
(g(k)
)T
d(k)
(d(k))T Q d(k)
,
g(k)
= Qx(k)
− b,
converges to the unique solution x of the system Q x = b
exactly after n steps, i.e. x(n)
= x.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
CONJUGATE GRADIENT METHOD FOR THE
QUADRATIC CASE
Step 1 Choose x(0)
∈ Rn
arbitrary. Define d(0)
= −g(0)
= b − Qx(0)
. Set
k = 0.
Step 2 Use the scheme
x(k+1)
= x(k)
+ αk d(k)
αk =
−(g(k)
)T
d(k)
(d(k))T Qd(k)
d(k+1)
= −g(k+1)
+ βk d(k)
βk =
(g(k+1)
)T
Qd(k)
(d(k))T Qd(k)
g(k)
= Qx(k)
− b .
Step 3 Continue till we get x(n)
. Then x(n)
= x and hence stop.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
The main point to note here is that we do not need all directions
d(0)
, d(1)
, . . . , d(k)
at one go.
We start from d(0)
and generate subsequent conjugate
directions as we proceed with the algorithm.
As soon as we determine d(n−1)
, the point x(n)
is known and that
is precisely the point x.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
EXAMPLE
Use the conjugate gradient method to minimize
f(x1, x2) = 3x2
1 − 4x1x2 + 2x2
2 + 4x1 + 6, (x1, x2) ∈ R2
.
Express the given quadratic function in the form
f(x) =
1
2
xT
Qx − bT
x.
It is simple to check here that
Q =

6 −4
−4 4

, b =

−4
0

and x =

x1
x2

.
Q is positive definite.
For the sake of illustration let x(0)
= (0, 0)T
.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Step 1 For x(0)
= (0, 0)T
, g(0)
= Qx(0)
− b =

6 −4
−4 4
 
0
0

−

−4
0

=

4
0

.
Therefore, d(0)
= −g(0)
= (−4, 0)T
.
Step 2 Next x(1)
= x(0)
+ α0d(0)
where
α0 =
−(g(0)
)T
d(0)
(d(0))T Qd(0)
=
−(4, 0)

4
0

(−4, 0)

6 −4
−4 4
 
−4
0
 = 1/6
Therefore,
x(1)
=

0
0

+ 1/6

−4
0

=

−2/3
0

Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Step 3 Now,
g(1)
= Qx(1)
− b =

6 −4
−4 4
 
−2/3
0

−

−4
0

=

0
8/3

,
β0 =
(g(1)
)T Qd(0)
(d(0))T Qg(0)
=
(0, 8/3)

6 −4
−4 4
 
−4
0

96
= 4/9,
and d(1)
= −g(1)
+ β0d(0)
=

0
−8/3

+
4
9

−4
0

=

−16/9
−8/3

.
Step 4 Now we obtain x(2)
as
x(2)
= x(1)
+ α1d(1)
, where α1 =
−(g(1)
)T
d(1)
(d(1))T Qd(1)
= 3/4.
x(2)
=

−2/3
0

+
3
4

−16/9
−8/3

=

−2
−2

= x,
i.e. the minimizing point of f(x1, x2) is x1 = −2, x2 = −2.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
For the conjugate gradient method, following are true
(I) (g(k+1)
)T
d(k)
= 0
(II) αk =
(g(k)
)T
g(k)
(d(k))T Q d(k)
(III) βk =
(g(k+1)
)T
g(k+1)
(g(k))T g(k)
(IV) (d(k)
)T
Q d(i)
= 0, (i = 0, 1, . . . , k − 1).
Here g(k)
is the gradient of the objective function at x(k)
i.e.
g(k)
= Q x(k)
− b.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
FLETCHER AND REEVES’ METHOD
Let us consider UMP Min
x∈Rn
f(x) and attempt to translate the steps of
the conjugate gradient method for UMP.
Step 1 Choose x(0)
∈ Rn
arbitrary and take d(0)
= −g(0)
= −∇f(x(0)
).
Set k = 0.
Step 2 Obtain αk  0 such that
h(αk ) = Min
αk 0
h(αk ) ,
where
h(αk ) = f(x(k)
+ αk d(k)
) .
Step 3 Define x(k+1)
= x(k)
+ αk d(k)
and take d(k+1)
= −g(k+1)
+ βk d(k)
where g(k+1)
= ∇f(x(k+1)
) and βk is given by
βk =
(g(k+1)
)T
g(k+1)
(g(k))T g(k)
.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
Step 4 Continue till we get a point x(p)
of the solution set, i.e.
k∇f(x(p)
)k   for some preassigned tolerance   0.
Above method which has the descent property and also has the
quadratic termination property, but is not globally convergent, i.e.
we can not start the method from an arbitrary point x(0)
∈ Rn
.
To make this algorithm globally convergent we incorporate
Powell’s correction in the above procedure.
Dr Reshma Khemchandani Optimization
Optimization
Unconstrained Optimization Problems
POWELL’S CORRECTION
The procedure is as follows:
Step 1 Choose x(0)
∈ Rn
arbitrary and take d(0)
= −g(0)
= −∇f(x(0)
).
Step 2 For k = 0, 1, . . . , (n − 1),
(A) Set x(k+1)
= x(k)
+ αk d(k)
, where αk  0 is chosen such
that h(αk ) = Min
αk 0
h(αk ), h(αk ) = f(x(k)
+ αk d(k)
) .
(B) Compute g(k+1)
= ∇f(x(k+1)
) .
(C) Unless k = (n − 1), set
d(k+1)
= −g(k+1)
+ βk d(k)
,
βk =
(g(k+1)
)T
g(k+1)
(g(k))T g(k)
.
Step 3 Replace x(0)
by x(n)
and go back to Step 1.
Dr Reshma Khemchandani Optimization

More Related Content

Similar to opt_slides_ump.pdf

Non linearequationsmatlab
Non linearequationsmatlabNon linearequationsmatlab
Non linearequationsmatlabsheetslibrary
 
Solution of non-linear equations
Solution of non-linear equationsSolution of non-linear equations
Solution of non-linear equationsZunAib Ali
 
Non linearequationsmatlab
Non linearequationsmatlabNon linearequationsmatlab
Non linearequationsmatlabZunAib Ali
 
Complexity Analysis of Recursive Function
Complexity Analysis of Recursive FunctionComplexity Analysis of Recursive Function
Complexity Analysis of Recursive FunctionMeghaj Mallick
 
AMS_502_13, 14,15,16 (1).pptx
AMS_502_13, 14,15,16 (1).pptxAMS_502_13, 14,15,16 (1).pptx
AMS_502_13, 14,15,16 (1).pptxbhavypatel2228
 
Delayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsDelayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsChristian Robert
 
Fractional Newton-Raphson Method and Some Variants for the Solution of Nonlin...
Fractional Newton-Raphson Method and Some Variants for the Solution of Nonlin...Fractional Newton-Raphson Method and Some Variants for the Solution of Nonlin...
Fractional Newton-Raphson Method and Some Variants for the Solution of Nonlin...mathsjournal
 
Maxima & Minima
Maxima & MinimaMaxima & Minima
Maxima & MinimaArun Umrao
 
Maxima & Minima of Functions - Differential Calculus by Arun Umrao
Maxima & Minima of Functions - Differential Calculus by Arun UmraoMaxima & Minima of Functions - Differential Calculus by Arun Umrao
Maxima & Minima of Functions - Differential Calculus by Arun Umraossuserd6b1fd
 
MLHEP 2015: Introductory Lecture #1
MLHEP 2015: Introductory Lecture #1MLHEP 2015: Introductory Lecture #1
MLHEP 2015: Introductory Lecture #1arogozhnikov
 
On Application of Power Series Solution of Bessel Problems to the Problems of...
On Application of Power Series Solution of Bessel Problems to the Problems of...On Application of Power Series Solution of Bessel Problems to the Problems of...
On Application of Power Series Solution of Bessel Problems to the Problems of...BRNSS Publication Hub
 
Classical optimization theory Unconstrained Problem
Classical optimization theory Unconstrained ProblemClassical optimization theory Unconstrained Problem
Classical optimization theory Unconstrained ProblemSurya Teja
 
A Level Set Method For Multiobjective Combinatorial Optimization Application...
A Level Set Method For Multiobjective Combinatorial Optimization  Application...A Level Set Method For Multiobjective Combinatorial Optimization  Application...
A Level Set Method For Multiobjective Combinatorial Optimization Application...Scott Faria
 
Matlab lab manual
Matlab lab manualMatlab lab manual
Matlab lab manualnmahi96
 
Applied Algorithms and Structures week999
Applied Algorithms and Structures week999Applied Algorithms and Structures week999
Applied Algorithms and Structures week999fashiontrendzz20
 

Similar to opt_slides_ump.pdf (20)

Non linearequationsmatlab
Non linearequationsmatlabNon linearequationsmatlab
Non linearequationsmatlab
 
Solution of non-linear equations
Solution of non-linear equationsSolution of non-linear equations
Solution of non-linear equations
 
Non linearequationsmatlab
Non linearequationsmatlabNon linearequationsmatlab
Non linearequationsmatlab
 
algorithm Unit 4
algorithm Unit 4 algorithm Unit 4
algorithm Unit 4
 
Complexity Analysis of Recursive Function
Complexity Analysis of Recursive FunctionComplexity Analysis of Recursive Function
Complexity Analysis of Recursive Function
 
AJMS_389_22.pdf
AJMS_389_22.pdfAJMS_389_22.pdf
AJMS_389_22.pdf
 
AMS_502_13, 14,15,16 (1).pptx
AMS_502_13, 14,15,16 (1).pptxAMS_502_13, 14,15,16 (1).pptx
AMS_502_13, 14,15,16 (1).pptx
 
Delayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsDelayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithms
 
Fractional Newton-Raphson Method and Some Variants for the Solution of Nonlin...
Fractional Newton-Raphson Method and Some Variants for the Solution of Nonlin...Fractional Newton-Raphson Method and Some Variants for the Solution of Nonlin...
Fractional Newton-Raphson Method and Some Variants for the Solution of Nonlin...
 
Maxima & Minima
Maxima & MinimaMaxima & Minima
Maxima & Minima
 
Maxima & Minima of Functions - Differential Calculus by Arun Umrao
Maxima & Minima of Functions - Differential Calculus by Arun UmraoMaxima & Minima of Functions - Differential Calculus by Arun Umrao
Maxima & Minima of Functions - Differential Calculus by Arun Umrao
 
MLHEP 2015: Introductory Lecture #1
MLHEP 2015: Introductory Lecture #1MLHEP 2015: Introductory Lecture #1
MLHEP 2015: Introductory Lecture #1
 
On Application of Power Series Solution of Bessel Problems to the Problems of...
On Application of Power Series Solution of Bessel Problems to the Problems of...On Application of Power Series Solution of Bessel Problems to the Problems of...
On Application of Power Series Solution of Bessel Problems to the Problems of...
 
Classical optimization theory Unconstrained Problem
Classical optimization theory Unconstrained ProblemClassical optimization theory Unconstrained Problem
Classical optimization theory Unconstrained Problem
 
A Level Set Method For Multiobjective Combinatorial Optimization Application...
A Level Set Method For Multiobjective Combinatorial Optimization  Application...A Level Set Method For Multiobjective Combinatorial Optimization  Application...
A Level Set Method For Multiobjective Combinatorial Optimization Application...
 
03_AJMS_209_19_RA.pdf
03_AJMS_209_19_RA.pdf03_AJMS_209_19_RA.pdf
03_AJMS_209_19_RA.pdf
 
03_AJMS_209_19_RA.pdf
03_AJMS_209_19_RA.pdf03_AJMS_209_19_RA.pdf
03_AJMS_209_19_RA.pdf
 
Matlab lab manual
Matlab lab manualMatlab lab manual
Matlab lab manual
 
Applied Algorithms and Structures week999
Applied Algorithms and Structures week999Applied Algorithms and Structures week999
Applied Algorithms and Structures week999
 
BP106RMT.pdf
BP106RMT.pdfBP106RMT.pdf
BP106RMT.pdf
 

Recently uploaded

Call Girls Narol 7397865700 Independent Call Girls
Call Girls Narol 7397865700 Independent Call GirlsCall Girls Narol 7397865700 Independent Call Girls
Call Girls Narol 7397865700 Independent Call Girlsssuser7cb4ff
 
Comparative Analysis of Text Summarization Techniques
Comparative Analysis of Text Summarization TechniquesComparative Analysis of Text Summarization Techniques
Comparative Analysis of Text Summarization Techniquesugginaramesh
 
Electronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdfElectronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdfme23b1001
 
Artificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxArtificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxbritheesh05
 
Arduino_CSE ece ppt for working and principal of arduino.ppt
Arduino_CSE ece ppt for working and principal of arduino.pptArduino_CSE ece ppt for working and principal of arduino.ppt
Arduino_CSE ece ppt for working and principal of arduino.pptSAURABHKUMAR892774
 
TechTAC® CFD Report Summary: A Comparison of Two Types of Tubing Anchor Catchers
TechTAC® CFD Report Summary: A Comparison of Two Types of Tubing Anchor CatchersTechTAC® CFD Report Summary: A Comparison of Two Types of Tubing Anchor Catchers
TechTAC® CFD Report Summary: A Comparison of Two Types of Tubing Anchor Catcherssdickerson1
 
IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024Mark Billinghurst
 
Past, Present and Future of Generative AI
Past, Present and Future of Generative AIPast, Present and Future of Generative AI
Past, Present and Future of Generative AIabhishek36461
 
What are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxWhat are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxwendy cai
 
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)Dr SOUNDIRARAJ N
 
Work Experience-Dalton Park.pptxfvvvvvvv
Work Experience-Dalton Park.pptxfvvvvvvvWork Experience-Dalton Park.pptxfvvvvvvv
Work Experience-Dalton Park.pptxfvvvvvvvLewisJB
 
8251 universal synchronous asynchronous receiver transmitter
8251 universal synchronous asynchronous receiver transmitter8251 universal synchronous asynchronous receiver transmitter
8251 universal synchronous asynchronous receiver transmitterShivangiSharma879191
 
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)dollysharma2066
 
computer application and construction management
computer application and construction managementcomputer application and construction management
computer application and construction managementMariconPadriquez1
 
Call Girls Delhi {Jodhpur} 9711199012 high profile service
Call Girls Delhi {Jodhpur} 9711199012 high profile serviceCall Girls Delhi {Jodhpur} 9711199012 high profile service
Call Girls Delhi {Jodhpur} 9711199012 high profile servicerehmti665
 
An experimental study in using natural admixture as an alternative for chemic...
An experimental study in using natural admixture as an alternative for chemic...An experimental study in using natural admixture as an alternative for chemic...
An experimental study in using natural admixture as an alternative for chemic...Chandu841456
 

Recently uploaded (20)

Call Girls Narol 7397865700 Independent Call Girls
Call Girls Narol 7397865700 Independent Call GirlsCall Girls Narol 7397865700 Independent Call Girls
Call Girls Narol 7397865700 Independent Call Girls
 
Comparative Analysis of Text Summarization Techniques
Comparative Analysis of Text Summarization TechniquesComparative Analysis of Text Summarization Techniques
Comparative Analysis of Text Summarization Techniques
 
Electronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdfElectronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdf
 
Artificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxArtificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptx
 
Arduino_CSE ece ppt for working and principal of arduino.ppt
Arduino_CSE ece ppt for working and principal of arduino.pptArduino_CSE ece ppt for working and principal of arduino.ppt
Arduino_CSE ece ppt for working and principal of arduino.ppt
 
TechTAC® CFD Report Summary: A Comparison of Two Types of Tubing Anchor Catchers
TechTAC® CFD Report Summary: A Comparison of Two Types of Tubing Anchor CatchersTechTAC® CFD Report Summary: A Comparison of Two Types of Tubing Anchor Catchers
TechTAC® CFD Report Summary: A Comparison of Two Types of Tubing Anchor Catchers
 
IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024
 
Past, Present and Future of Generative AI
Past, Present and Future of Generative AIPast, Present and Future of Generative AI
Past, Present and Future of Generative AI
 
What are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxWhat are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptx
 
young call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Service
young call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Serviceyoung call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Service
young call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Service
 
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
 
Work Experience-Dalton Park.pptxfvvvvvvv
Work Experience-Dalton Park.pptxfvvvvvvvWork Experience-Dalton Park.pptxfvvvvvvv
Work Experience-Dalton Park.pptxfvvvvvvv
 
8251 universal synchronous asynchronous receiver transmitter
8251 universal synchronous asynchronous receiver transmitter8251 universal synchronous asynchronous receiver transmitter
8251 universal synchronous asynchronous receiver transmitter
 
Design and analysis of solar grass cutter.pdf
Design and analysis of solar grass cutter.pdfDesign and analysis of solar grass cutter.pdf
Design and analysis of solar grass cutter.pdf
 
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
 
computer application and construction management
computer application and construction managementcomputer application and construction management
computer application and construction management
 
young call girls in Green Park🔝 9953056974 🔝 escort Service
young call girls in Green Park🔝 9953056974 🔝 escort Serviceyoung call girls in Green Park🔝 9953056974 🔝 escort Service
young call girls in Green Park🔝 9953056974 🔝 escort Service
 
🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...
🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...
🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...
 
Call Girls Delhi {Jodhpur} 9711199012 high profile service
Call Girls Delhi {Jodhpur} 9711199012 high profile serviceCall Girls Delhi {Jodhpur} 9711199012 high profile service
Call Girls Delhi {Jodhpur} 9711199012 high profile service
 
An experimental study in using natural admixture as an alternative for chemic...
An experimental study in using natural admixture as an alternative for chemic...An experimental study in using natural admixture as an alternative for chemic...
An experimental study in using natural admixture as an alternative for chemic...
 

opt_slides_ump.pdf

  • 1. Optimization Optimization Dr Reshma Khemchandani Dr Reshma Khemchandani Optimization
  • 2. Optimization Unconstrained Optimization Problems OUTLINE 1 UNCONSTRAINED OPTIMIZATION PROBLEMS Dr Reshma Khemchandani Optimization
  • 3. Optimization Unconstrained Optimization Problems An optimization problem where decision vector x is allowed to take any value in Rn is called an unconstrained optimization problem (UMP). Aim is to study standard algorithms for unconstrained minimization of functions of one variable as well as functions of several variables. Consider the unconstrained minimization problem (UMP) Min x∈Rn f(x) to get a point x ∈ Rn such that f(x) ≤ f(x) for all x ∈ Rn . Unfortunately, finding local minimum for general function is very difficult task. Dr Reshma Khemchandani Optimization
  • 4. Optimization Unconstrained Optimization Problems Solution Set: Let f : Rn −→ R be continuously differentiable. Then the set Ω = { x ∈ Rn : ∇f(x) = 0 } is called the solution set of (1). Let f : Rn −→ R be differentiable convex function and x be a point of the solution set. Then x is a global min point of UMP. We now describe a common basic scheme of the form x(k+1) = x(k) + αk d(k) for solving the UMP where x(k) is the current solution, d(k) is the direction of movement from x(k) and αk > 0 (called the step size) is the distance upto which we move in the direction d(k) from the current point x(k) . Dr Reshma Khemchandani Optimization
  • 5. Optimization Unconstrained Optimization Problems Descent Property: An algorithm for solving the UMP is said to have the descent property if the objective function value decreases as we go through the sequence {x(k) }, i.e. f(x(k+1) ) < f(x(k) ) for all k. In other words, for the algorithm to possess the descent property, the objective function should decrease as we proceed. Quadratic Termination Property: An algorithm for UMP is said to have quadratic termination property (q.t.p) if the minimum of a positive definite quadratic form in n variables is reached in at most n iterations. The motivation for defining q.t.p stems from the fact that near a local min point, the function behaves like a strictly convex function(like a parabola in R or a positive definite quadratic form in Rn ) and therefore if the algorithm behaves well on such a function, it will ‘hopefully’ do well on the other functions as well. Dr Reshma Khemchandani Optimization
  • 6. Optimization Unconstrained Optimization Problems Globally Convergent: An algorithm for the UMP is said to be globally convergent if starting from any point x(0) ∈ Rn , the sequence {x(k) } always converges to a point of the solution set Ω. The property of ‘global convergence’ guarantees that no matter from any arbitrary point x(0) ∈ Rn we start, we are guaranteed to generate a sequence {x(k) } that converges to a point of the solution set. Here it must be noted that two different starting points will, in general, generate two different sequences of iterates but both converging to a point of solution set. Order of Convergence: Let the sequence {x(k) } converge to a point x and let x(k) 6= x for sufficiently large k. The quantity kx(k) − xk is called the error of the kth iterate x(k) . Dr Reshma Khemchandani Optimization
  • 7. Optimization Unconstrained Optimization Problems Suppose that there exists p and 0 < a < ∞ such that lim k→∞ ||x(k+1) − x|| ||x(k) − x||p = a (0 < a < ∞) then p is called the order of convergence of the sequence {x(k) }. Thus kx(k+1) − xk = a||x(k) − x||p asymptotically. If p = 1, the sequence {x(k) } is said to have linear convergence rate and for p = 2, it is said to have quadratic convergence rate. In case p = 1 but a = 0, then the sequence {x(k) } is said to have super linear convergence rate. Dr Reshma Khemchandani Optimization
  • 8. Optimization Unconstrained Optimization Problems LINE SEARCH METHODS FOR UNIMODAL FUNCTIONS Unimodal Min Function: The function f : [a, b] −→ R is said to be a unimodal (to be specific unimodal min) function if it has only one mode i.e. it has a single relative min, i.e. ∃ a ≤ α ≤ b such that (i) f is strictly decreasing (↓) in [a, α). (ii) f is strictly increasing (↑) in [α, b]. Similar definition hold for unimodal max function. Unimodal function may not be differentiable - in fact it may not even be continuous Dr Reshma Khemchandani Optimization
  • 9. Optimization Unconstrained Optimization Problems PLOT OF UNI-MODAL FUNCTIONS Dr Reshma Khemchandani Optimization
  • 10. Optimization Unconstrained Optimization Problems Basic Strategy Let us choose two distinct points (say) x1 and x2 in [a, b] i.e. a ≤ x1 < x2 ≤ b. Let xmin denote the point giving the minimum value of f(x) over [a, b]. Since f is uni-modal min function, it is clear that (i) f(x1) < f(x2) ⇒ xmin ∈ [a, x2] (ii) f(x1) > f(x2) ⇒ xmin ∈ [x1, b] (iii) f(x1) = f(x2) ⇒ xmin ∈ [x1, x2] . Dr Reshma Khemchandani Optimization
  • 11. Optimization Unconstrained Optimization Problems LINE SEARCH METHODS (I) The Golden Section Rule. (II) The Fibonacci Search Method. In discussing the aforesaid methods we shall make use of the following notations xL,k = lower limit of the search interval at the kth iteration (so xL,1 = a) xU,k = upper limit of the search interval at the kth iteration (so xU,1 = b) xp,k = 1st trial point at the kth iteration xq,k = 2nd trial point at the kth iteration Ep,k = f(xp,k ) value of the function at the 1st trial point Eq,k = f(xq,k ) value of the function at the 2nd trial point Ik = xU,k − xL,k = length of the search interval at the kth iteration (so I1 = (b - a)) IL k = length of the left part of the search interval Ik IR k = length of the right part of the search interval Ik . Dr Reshma Khemchandani Optimization
  • 12. Optimization Unconstrained Optimization Problems x x p,k L,k xq,k= xp,k+1 xq,k+1 x U,k I I Ik+1 k+1 k+2 I k+2 L L R R Dr Reshma Khemchandani Optimization
  • 13. Optimization Unconstrained Optimization Problems THE GOLDEN SECTION RULE Here the two trial points xp,k and xq,k are chosen as per the following criteria (I) IL k = IR k for all k , i.e. no undue advantage is given to left or right part. (II) xp,k+1 = xq,k and xq,k+1 is computed afresh for the interval IR k+1 ( or xq,k+1 = xp,k and xp,k+1 is computed afresh for the interval IL k+1), i.e. only one new trial point is computed at the kth iteration, the other trial point is obtained from the previous iteration. In view of this we have Ik = Ik+1 + Ik+2 for all k. (III) Ik Ik+1 = Ik+1 Ik+2 = c (constant) for all k. This is called the golden section criterion. Dr Reshma Khemchandani Optimization
  • 14. Optimization Unconstrained Optimization Problems Thus we wish to choose trial points xp,k and xq,k according to criterion (i), (ii) and (iii). From (ii) we have Ik = Ik+1 + Ik+2 i.e. Ik Ik+2 = Ik+1 Ik+2 + 1 i.e. Ik Ik+1 Ik+1 Ik+2 = Ik+1 Ik+2 + 1 (by (iii)) i.e. c2 = c + 1. Therefore c = 1± √ 5 2 Since c can not be negative, we get c = 1 + √ 5 2 = 1.618 (golden section ratio and 1 1.618 = 0.618 . Dr Reshma Khemchandani Optimization
  • 15. Optimization Unconstrained Optimization Problems STEPWISE DESCRIPTION Step 1 Input data- xL,1, xU,1, , f (Here 0 is the tolerance to be prescribed by the user). Step 2 Compute the first two trial points xp,1 and xq,1, where xp,1 = xU,1 − 0.618 (xU,1 − xL,1) xq,1 = xL,1 + 0.618 (xU,1 − xL,1). Set k = 1. Step 3 Evaluate the function f at the two trial points xp,k and xq,k . Let Ep,k = f(xp,k ) Eq,k = f(xq,k ) . Step 4 Test the interval which contains the minimum, i.e. if Ep,k ≤ Eq,k , go to Step 5 otherwise go to Step 6. Dr Reshma Khemchandani Optimization
  • 16. Optimization Unconstrained Optimization Problems Step 5 Use following relations to update the data xL,k+1 = xL,k xU,k+1 = xq,k xp,k+1 = xU,k+1 - 0.618 Ik+1 xq,k+1 = xp,k Ep,k+1 = f(xp,k+1) Eq,k+1 = f(xq,k+1) = Ep,k . Step 6 Use the following relations to update the data xL,k+1 = xp,k xU,k+1 = xU,k xp,k+1 = xq,k xq,k+1 = xL,k + 0.618 Ik+1 Ep,k+1 = f(xp,k+1) = Eq,k Eq,k+1 = f(xq,k+1) . Step 7 Test for the end of optimization , i.e. if Ik , go to Step 8, otherwise set k = (k+1) and go to Step 4. Dr Reshma Khemchandani Optimization
  • 17. Optimization Unconstrained Optimization Problems Step 8 Output xL,n, xU,n. Then xmin ∈ (xL,n, xU,n) and Emin ≤ Min(Ep,n−1, Eq,n−1). It can be noted that for the golden section rule, In I1 = 1 1.618 n−1 = (0.618)n−1 . Thus knowing I1 and In we can find out the number of iterations as well as the number of points at which the function is to be evaluated. Note that if we are going upto the 7th iteration (i.e. getting xL,7 and xU,7) then we shall be having n = 7 functional evaluations, namely, at xp,1, xq,1, and one each at 2nd , 3rd , 4th , 5th and 6th iteration. To stop at the 7th iteration we shall be computing xp,6 and xq,6 and Emin ≤ Min (xp,6, xq,6). Thus N functional evaluations will mean stopping at the Nth iteration and computing points upto xp,N−1 and xq,N−1. Also Emin ≤ Min (xp,N−1, xq,N−1) and there are only (N − 1) interval reductions. Dr Reshma Khemchandani Optimization
  • 18. Optimization Unconstrained Optimization Problems Find min x2 over [-5, 15] by the golden section rule. Take = 1.5. k xL,k xU,k xp,k xq,k Ep,k Eq,k L/R 1 - 5.0 15.00 2.64 7.36 6.96 54.1 L 2 -5.0 7.36 -0.2 2.64 0.077 6.96 L 3 -5.0 2.64 -2.08 -0.27 4.33 0.077 R . . 4 -2.08 2.64 -0.2 0.84 0.077 0.71 L 5 -2.08 0.84 -0.96 -0.27 0.92 0.077 R . . 6 -0.96 0.84 -0.27 0.15 0.077 0.023 R 7 -0.27 0.84 I7 = 1.11 1.5(= ), stop, xmin ∈ [-0.27, 0.84] Emin ≤ Min (0.077, 0.023) = 0.023. Dr Reshma Khemchandani Optimization
  • 19. Optimization Unconstrained Optimization Problems The Fibonacci Search Method Fibonacci Sequence: Let F0 = 1, F1 = 1 and Fi = Fi−1 + Fi−2 (i ≥ 2). Then {Fn} is called the sequence of Fibonacci numbers or in short, the Fibonacci sequence. Thus, the Fibonacci sequence is {1, 1, 2, 3, 5, 8, 13, 21, . . .} where F0 = 1, F1 = 1, F2 = 2, F3 = 3, F4 = 5, F5 = 8, F6 = 8, F7 = 21... and so on. (I) IL k = IR k for all k (II) Ik = Ik+1 + Ik+2 (III) Ik Ik+1 = Fn−k+1 Fn−k , where n denotes the number of iterations to be performed. Dr Reshma Khemchandani Optimization
  • 20. Optimization Unconstrained Optimization Problems We now give a justification of the condition (iii) given above. If we wish to stop at the nth iteration, then In+1 ' In = 1 In In = In = 1 In In−1 = In + In+1 = 2 In In−2 = In + In−1 = 3 In In−3 = 5 In In−4 = 8 In . . . I1 = FnIn. In the golden section rule Ik Ik+1 = c = 1.618 for all k and all n. In Fibonacci search method this ratio depends on both the current iteration k and also the total number of iterations n to be performed. Therefore instead of 0.618 (i.e. 1 c ) we shall be using the number Fn−k Fn−k+1 for determining the points xp,k and xq,k . Dr Reshma Khemchandani Optimization
  • 21. Optimization Unconstrained Optimization Problems STEPWISE APPROACH OF THE FIBONACCI SEARCH METHOD Step 1 Input data: xL,1, xU,1, n, f. Step 2 Compute Fibonacci numbers Fn and Fn−1 and find xp,1 = xU,1 - Fn−1 Fn (xU,1 − xL,1) xq,1 = xL,1 + Fn−1 Fn (xU,1 − xL,1) Set k = 1. Step 3 Evaluate Ep,k = f(xp,k ) Eq,k = f(xq,k ) and test the interval which contains the minimum. If Ep,k ≤ Eq,k , go to Step 4, otherwise go to Step 5. Dr Reshma Khemchandani Optimization
  • 22. Optimization Unconstrained Optimization Problems Step 4 Use the following relations to update the data xL,k+1 = xL,k xU,k+1 = xq,k xp,k+1 = xU,k+1 - Fn−k Fn−k+1 Ik+1 xq,k+1 = xp,k Ep,k+1 = f(xp,k+1) Eq,k+1 = f(xq,k+1) = Ep,k . Step 5 Use the following relations to update the data xL,k+1 = xp,k xU,k+1 = xU,k xp,k+1 = xq,k+1 xq,k+1 = xL,k + Fn−k Fn−k+1 Ik+1 Ep,k+1 = f(xp,k+1) = Eq,k Eq,k+1 = f(xq,k+1) . Dr Reshma Khemchandani Optimization
  • 23. Optimization Unconstrained Optimization Problems Step 6 If k ≤ (n − 2), then set k = k + 1 and go to step 3. If k = (n − 1) then go to Step 7. Step 7 As F0 = F1 = 1, it can be seen that at the (n − 1)th iteration, the two trial points xp,n−1 and xq,n−1 will come out to be the same , i.e. xp,n−1 = xq,n−1. To make them distinct we have to add / subtract a fixed positive number (say 0.01) to one of these. For this we look at the (n − 2)th iteration. If there (i.e. at the (n − 2)th iteration) we are searching in the left then is subtracted from xp,n−1 otherwise it is added to xq,n−1. We then determine new xp,n−1 and xq,n−1 and hence xL,n and xU,n. Then xmin ∈ (xL,n, xU,n) and fmin ≤ Min(Ep,n−1, Eq,n−1) As In I1 = 1 Fn , so given I1 and In, the value of n can be computed in advance. Dr Reshma Khemchandani Optimization
  • 24. Optimization Unconstrained Optimization Problems EXAMPLE With n = 7, find min x2 over [-5,15] via Fibonacci search method. k Fn−k /Fn−k+1 xL,k xU,k xp,k xq,k Ep,k Eq,k 1 13/21 -5.0 15.00 2.64 7.38 6.88 54.6 2 8/13 -5.0 7.38 -0.24 2.62 0.058 6.88 3 5/8 -5.0 2.62 -2.14 -0.24 4.67 0.058 . . 4 3/5 -2.14 2.62 -0.24 0.72 0.058 0.52 5 2/3 -2.14 0.72 -1.2 -0.24 1.44 0.058 . . 6 1/2 - 1.2 0.72 -0.24 -0.24+0.01 0.058 0.053 7 1 -0.24 0.72 Dr Reshma Khemchandani Optimization
  • 25. Optimization Unconstrained Optimization Problems From this table we observe that xmin ∈ [−0.24, 0.72] and fmin ≤ 0.053. In the given example, I1 = 20 and therefore if we take = 1.5, then In = 1.5. This gives I1 In = 20 1.5 = 13.3 which from the sequence of Fibonacci numbers determines n = 7 as F6 = 13 and F7 = 21. Also as explained in Step 7, the two trial points at the (n − 1)th iteration, i.e. xp,6 and xq,6 both becomes equal to 0.24. Therefore we look at the (n − 2)th iteration, i.e. the 5th iteration in our example. As there we are seeking the right part, we add 0.01 to xq,6 to get new xq,6 and then continue as explained. Dr Reshma Khemchandani Optimization
  • 26. Optimization Unconstrained Optimization Problems RELATION BETWEEN THE FIBONACCI SEARCH METHOD AND THE GOLDEN SECTION RULE For the constant c appearing in the golden section rule we have c = 1 + 1/c = 1.618. Fn ' cn+1 √ 5 (for large n). Let RF and RGS denote the reduction factors for the Fibonacci search method and the golden section rule respectively. RF = In I1 = 1 Fn = √ 5 cn+1 RGS = In I1 = 1 c n−1 = 1 cn−1 . Therefore RGS RF = 1 cn−1 × cn+1 √ 5 = c2 √ 5 ' 1.17 . Dr Reshma Khemchandani Optimization
  • 27. Optimization Unconstrained Optimization Problems THE STEEPEST DESCENT METHOD The steepest descent method is gradient based methods for solving UMP. Simple to implement and therefore used widely in various applications. The only drawback of the steepest descent method is its slow convergence (as its order of convergence is one ). The UMP: Min x∈Rn f(x), f has continuous first order partial derivatives on Rn . Then the basic scheme involved is x(k+1) = x(k) + αk d(k) where d(k) = −∇f(x(k) )/k∇f(x(k) )k, is the direction of movement and the step size αk ≥ 0 is chosen such that h(αk ) = Min αk ≥0 h(αk ). Here the function h(αk ) is given by h(αk ) = f(x(k) + αk d(k) ), and we stop when k∇f(x(k) )k . Dr Reshma Khemchandani Optimization
  • 28. Optimization Unconstrained Optimization Problems A stepwise description of the steepest descent method could be Step 1 Choose x(0) ∈ Rn and tolerance 0. Set k = 0. Step 2 Compute ∇f(x(k) ) and d(k) = −∇f(x(k) )/k∇f(x(k) )k. Step 3 Evaluate x(k+1) = x(k) + αk d(k) where αk 0 is chosen such that h(αk ) = Min αk ≥0 h(αk ) , h(αk ) = f(x(k) + αk d(k) ) . The Steps 2 and 3 above are repeated till k∇f(x(k) )k . In that case x(k) becomes a point of the solution set and therefore, if f is a convex function then it becomes an optimal solution. Dr Reshma Khemchandani Optimization
  • 29. Optimization Unconstrained Optimization Problems The steepest descent algorithm (I) has descent property (II) does not have quadratic termination property (III) is globally convergent and (IV) has order of convergence p = 1. Therefore if we are using the steepest descent method then we can start from any point x(0) and as we proceed, the objective function value will decrease, but the algorithm may take lot many iterations near the optimal solution. Also it may, in general, take more than n iterations to minimize a positive definite quadratic form of n variables. Dr Reshma Khemchandani Optimization
  • 30. Optimization Unconstrained Optimization Problems Example: Use the steepest descent method to minimize f(x1, x2) = 3x2 1 − 4x1x2 + 2x2 2 + 4x1 + 6 over (x1, x2) ∈ R2 . The given function is a convex function and so the steepest descent method will give a global optimal solution. Starting from x(0) = (0, 0)T and following Steps 1-3 described above, we get k x(k) ∇f(x(k) ) d(k) αk 1 (0, 0)T (4, 0)T (−1, 0)T 2/3 2 (−2/3, 0)T (0, 8/3)T (0, −1)T 2/3 3 (−2/3, −2/3)T (8/3, 0)T (−1, 0)T 1/6 4 (−10/9, −2/3)T (0, 16/9)T (0, −1)T 1/4 5 (−38/27, −10/9)T (16/9, 0)T (−1, 0)T 1/6 . . . . . . . . . . . . . . . Dr Reshma Khemchandani Optimization
  • 31. Optimization Unconstrained Optimization Problems The function f(x1, x2) is a positive definite quadratic form in two variables but its optimal solution has not been obtained in atmost two iterations. This illustrates that the method of steepest descent does not possess quadratic termination property. We observe that the directions d(k) are repeated alternately (−1, 0)T , (0, −1)T , (−1, 0)T , (0, −1)T etc. Any two consecutive directions d(k) and d(k+1) given by the steepest descent method are mutually orthogonal. Therefore in R2 , if the first two directions are d(1) and d(2) then d(3) has to be d(1) and d(4) has to be d(2) so on. But this repetition may not be in R3 and higher dimensional spaces because there if d(1) and d(2) are orthogonal then we may have d(3) , different from d(1) , which is orthogonal to d(2) . So the important thing is the orthogonality of consecutive directions and NOT their alternate repetition. Dr Reshma Khemchandani Optimization
  • 32. Optimization Unconstrained Optimization Problems x x2 1 FIGURE: −∇ f(x(k) ) which is normal to the surface f(x) = constant at x(k) , is tangent to the surface at the point x(k+1) . Dr Reshma Khemchandani Optimization
  • 33. Optimization Unconstrained Optimization Problems NEWTON’S METHOD Newton’s method: The basic scheme: yk+1 = yk − g(yk )/g0 (yk ) where yk is the current iterate or the current approximation, g(y) = 0, y ∈ R. Looking at the standard basic scheme of Newton’s method x(k+1) = x(k) − (Hf (x(k) ))−1 ∇f(x(k) ) for finding a solution of the system ∇f(x) = 0. Newton’s method for solving UMP’s has quadratic termination property (in fact here for minimizing a positive definite quadratic form of n variables it will take exactly one iteration). Order of convergence is 2 and it will have descent property, but it will not have the property of global convergence. Therefore except for the case when we are minimizing a positive definite quadratic form, we cannot start the method from an arbitrary point x(0) ∈ Rn . Dr Reshma Khemchandani Optimization
  • 34. Optimization Unconstrained Optimization Problems Use Newton’s method to minimize f(x1, x2) = 8x2 1 − 4x1x2 + 5x2 2 , (x1, x2) ∈ R2 . As the function f is a positive definite quadratic form we can start from any arbitrary point x(0) ∈ R2 . To be specific, let x(0) = (5, 2)T . Then ∇f(x(0) ) = 16x1 − 4x2 −4x1 + 10x2 (x1=5,x2=2) = 72 0 , Hf (x(0) ) = 16 −4 −4 10 , (Hf (x(0) ))−1 = 1 144 10 4 4 16 . x(1) = x(0) − (Hf (x(0) ))−1 ∇f(x(0) ) = 5 2 − 1 444 10 4 4 16 72 0 = 0 0 , giving x1 = 0, x2 = 0 as the minimizing point. Dr Reshma Khemchandani Optimization
  • 35. Optimization Unconstrained Optimization Problems THE CONJUGATE GRADIENT METHOD Conjugate based methods are better than the steepest descent method (in terms of order of convergence) and are simpler to implement than the modified Newton’s method. Conjugate Directions: Let Q be an (n × n) positive definite matrix. Any two non-zero vectors (directions) d(1) , d(2) ∈ Rn are said to be conjugate vectors or conjugate directions with respect to Q, if (d(1) )T Q d(2) = 0. For Q = I, conjugacy reduces to the usual concept of orthogonality. Therefore if d(1) and d(2) are conjugate with respect to Q, sometimes we also call them Q - orthogonal. A set {d(0) , . . . , d(k) } of (k + 1) vectors in Rn is said to be conjugate if every two of them are so, i.e. (d(i) )T Q d(j) = 0 (i 6= j). Dr Reshma Khemchandani Optimization
  • 36. Optimization Unconstrained Optimization Problems Let {d(0) , d(1) , . . . , d(k) } be a set of (k + 1) non-zero vectors which are conjugate with respect to a given positive definite matrix Q. Then the vectors d(0) , d(1) , . . . , d(k) are linearly independent. Consider the problem Min x∈Rn 1 2 xT Q x − bT x where Q is an (n × n) symmetric positive definite matrix and hence the problem is strictly convex and has unique minimizing point x ∈ Rn . Further the KKT conditions give ∇ 1 2 xT Qx − bT x = 0, i.e. Qx = b, which implies x = Q−1 b (note that as Q is positive definite, Q−1 exists), which is equivalent to finding the unique solution of the system of equations Qx = b. Dr Reshma Khemchandani Optimization
  • 37. Optimization Unconstrained Optimization Problems Let {d(0) , d(1) , . . . , d(n−1) } be a set of n non-zero vectors in Rn which are conjugate with respect to Q. Then x, the unique solution to the system Qx = b or equivalently, the unique minimizing point of problem, is given by x = n−1 X k=0 (d(k) )T b (d(k))T Qd(k) d(k) . Conjugate Direction Theorem: Let {d(0) , d(1) , . . . , d(n−1) } be a set of n non-zero vectors in Rn which are conjugate with respect to Q. For any x(0) ∈ Rn , the sequence {x(k) } is generated x(k+1) = x(k) + αk d(k) , αk = − (g(k) )T d(k) (d(k))T Q d(k) , g(k) = Qx(k) − b, converges to the unique solution x of the system Q x = b exactly after n steps, i.e. x(n) = x. Dr Reshma Khemchandani Optimization
  • 38. Optimization Unconstrained Optimization Problems CONJUGATE GRADIENT METHOD FOR THE QUADRATIC CASE Step 1 Choose x(0) ∈ Rn arbitrary. Define d(0) = −g(0) = b − Qx(0) . Set k = 0. Step 2 Use the scheme x(k+1) = x(k) + αk d(k) αk = −(g(k) )T d(k) (d(k))T Qd(k) d(k+1) = −g(k+1) + βk d(k) βk = (g(k+1) )T Qd(k) (d(k))T Qd(k) g(k) = Qx(k) − b . Step 3 Continue till we get x(n) . Then x(n) = x and hence stop. Dr Reshma Khemchandani Optimization
  • 39. Optimization Unconstrained Optimization Problems The main point to note here is that we do not need all directions d(0) , d(1) , . . . , d(k) at one go. We start from d(0) and generate subsequent conjugate directions as we proceed with the algorithm. As soon as we determine d(n−1) , the point x(n) is known and that is precisely the point x. Dr Reshma Khemchandani Optimization
  • 40. Optimization Unconstrained Optimization Problems EXAMPLE Use the conjugate gradient method to minimize f(x1, x2) = 3x2 1 − 4x1x2 + 2x2 2 + 4x1 + 6, (x1, x2) ∈ R2 . Express the given quadratic function in the form f(x) = 1 2 xT Qx − bT x. It is simple to check here that Q = 6 −4 −4 4 , b = −4 0 and x = x1 x2 . Q is positive definite. For the sake of illustration let x(0) = (0, 0)T . Dr Reshma Khemchandani Optimization
  • 41. Optimization Unconstrained Optimization Problems Step 1 For x(0) = (0, 0)T , g(0) = Qx(0) − b = 6 −4 −4 4 0 0 − −4 0 = 4 0 . Therefore, d(0) = −g(0) = (−4, 0)T . Step 2 Next x(1) = x(0) + α0d(0) where α0 = −(g(0) )T d(0) (d(0))T Qd(0) = −(4, 0) 4 0 (−4, 0) 6 −4 −4 4 −4 0 = 1/6 Therefore, x(1) = 0 0 + 1/6 −4 0 = −2/3 0 Dr Reshma Khemchandani Optimization
  • 42. Optimization Unconstrained Optimization Problems Step 3 Now, g(1) = Qx(1) − b = 6 −4 −4 4 −2/3 0 − −4 0 = 0 8/3 , β0 = (g(1) )T Qd(0) (d(0))T Qg(0) = (0, 8/3) 6 −4 −4 4 −4 0 96 = 4/9, and d(1) = −g(1) + β0d(0) = 0 −8/3 + 4 9 −4 0 = −16/9 −8/3 . Step 4 Now we obtain x(2) as x(2) = x(1) + α1d(1) , where α1 = −(g(1) )T d(1) (d(1))T Qd(1) = 3/4. x(2) = −2/3 0 + 3 4 −16/9 −8/3 = −2 −2 = x, i.e. the minimizing point of f(x1, x2) is x1 = −2, x2 = −2. Dr Reshma Khemchandani Optimization
  • 43. Optimization Unconstrained Optimization Problems For the conjugate gradient method, following are true (I) (g(k+1) )T d(k) = 0 (II) αk = (g(k) )T g(k) (d(k))T Q d(k) (III) βk = (g(k+1) )T g(k+1) (g(k))T g(k) (IV) (d(k) )T Q d(i) = 0, (i = 0, 1, . . . , k − 1). Here g(k) is the gradient of the objective function at x(k) i.e. g(k) = Q x(k) − b. Dr Reshma Khemchandani Optimization
  • 44. Optimization Unconstrained Optimization Problems FLETCHER AND REEVES’ METHOD Let us consider UMP Min x∈Rn f(x) and attempt to translate the steps of the conjugate gradient method for UMP. Step 1 Choose x(0) ∈ Rn arbitrary and take d(0) = −g(0) = −∇f(x(0) ). Set k = 0. Step 2 Obtain αk 0 such that h(αk ) = Min αk 0 h(αk ) , where h(αk ) = f(x(k) + αk d(k) ) . Step 3 Define x(k+1) = x(k) + αk d(k) and take d(k+1) = −g(k+1) + βk d(k) where g(k+1) = ∇f(x(k+1) ) and βk is given by βk = (g(k+1) )T g(k+1) (g(k))T g(k) . Dr Reshma Khemchandani Optimization
  • 45. Optimization Unconstrained Optimization Problems Step 4 Continue till we get a point x(p) of the solution set, i.e. k∇f(x(p) )k for some preassigned tolerance 0. Above method which has the descent property and also has the quadratic termination property, but is not globally convergent, i.e. we can not start the method from an arbitrary point x(0) ∈ Rn . To make this algorithm globally convergent we incorporate Powell’s correction in the above procedure. Dr Reshma Khemchandani Optimization
  • 46. Optimization Unconstrained Optimization Problems POWELL’S CORRECTION The procedure is as follows: Step 1 Choose x(0) ∈ Rn arbitrary and take d(0) = −g(0) = −∇f(x(0) ). Step 2 For k = 0, 1, . . . , (n − 1), (A) Set x(k+1) = x(k) + αk d(k) , where αk 0 is chosen such that h(αk ) = Min αk 0 h(αk ), h(αk ) = f(x(k) + αk d(k) ) . (B) Compute g(k+1) = ∇f(x(k+1) ) . (C) Unless k = (n − 1), set d(k+1) = −g(k+1) + βk d(k) , βk = (g(k+1) )T g(k+1) (g(k))T g(k) . Step 3 Replace x(0) by x(n) and go back to Step 1. Dr Reshma Khemchandani Optimization