3. Contents:
Meaning
Defination
What is dynamic programming used for
Technique used in it
Divide and Conquer Strategy
General Divide and Conquer recurrence
Common example
Approaches of dynamic programming
Elements of Dynamic Programming
Dynamic Programming and Chain Matrix Multiplication
Fibonacci Numbers
Steps for problem solving
Advantages and Disadvantages
Good bye :p :p
8. So Combine Meaning Of Dynamic Programming is :
Change and Schedule a Solution for a problem
9. Defination :
Dynamic Programming refers to a
very large class of algorithms. The
idea is to break a large problem
down (if possible) into incremental
steps so that, at any given stage,
optimal solutions are known to sub-
problems.
Ref :- Monash University
DEFINATION :
10. Use of dynamic programming
Dynamic programming is used for problems requiring a sequence of interrelated
decision. This means that to take another decision we have to depend on the
previous decision or solution formed.
13. The Divide and Conquer Algorithm
• Divide_Conquer(problem P)
• {
• if Small(P) return S(P);
• else {
• divide P into smaller instances P1, P2, …, P k , k1;
• Apply Divide_Conquer to each of these sub problems;
• return Combine(Divide_Conquer(P1),
Divide_Conquer(P2),…, Divide Conquer(P k));
• }
• }
14. Divide Conquer recurrence relation
The computing time of Divide Conquer is
T(n) is the time for Divide Conquer on any input size n.
g(n) is the time to compute the answer directly (for small inputs)
f(n) is the time for dividing P and combining the solutions.
14
)()(...)()(
)(
)(
21 nfnTnTnT
ng
nT
k
n small
otherwise
15. 15
General Divide and Conquer recurrence
The Master Theorem
T(n) = aT(n/b) + f (n), where f (n) ∈ Θ(n k)
1. a < bk T(n) ∈ Θ(n k)
2. a = bk T(n) ∈ Θ(n k lg n )
3. a > bk T(n) ∈ Θ(n log b a)
the time spent on solving a sub problem of size n/b.
the time spent on dividing the problem
into smaller ones and combining their solutions.
16. Difference between DP and Divide-and-
Conquer
Using Divide-and-Conquer to solve these problems is inefficient because
the same common sub problems have to be solved many times.
DP will solve each of them once and their answers are stored in a table
for future use.
27. It Contains Sequence of “n” decisions.
Each “n” corresponding to one of the decision.
Each stage of analysis is described by a set of elements
decision, input state, output state and return.
Then symbolic representation of n stages of analysis using
backward recursion so we can formalize the notation
Backward recursion
Cumulative return = Direct return + Cumulative return
through stage from stage through stage i-1
28. We use sb to denote the previous state
Tb determines the state that came before s when the decision made to reach state s is d
Db(s) is the set of decisions that can be used to enter state s
sb = Tb(s, d), //where d belongs to Db(s)
Backward recursion formula
29.
30. Approach takes a problem
Decomposed into a sequence of n stages
Analyzes the problem starting with the first stage
in the sequence
Working forward to the last stage
it is also known as deterministic probability
approach
Forward recursion
33. Elements of Dynamic Programming
i) Optimal substructure
ii) Overlapping sub problem
iii) Memoization
34. a problem is said to have optimal substructure if an optimal solution can be
constructed efficiently from optimal solutions of its sub problems. This property is
used to determine the usefulness of dynamic programming
ref : wikipedia
35.
36. Optimal Substructure
This is the necessary property , if this property is not present we cant use
dynamic programming.
a problem p , with sub problems p1,p2.
Solution of problem p is s , and s1 is the optimum solution of sub problem p1,and
s2 is the optimum solution of sub problem p2.
Claim “s” is the optimal solution ,if both solution of the sub problems are
optimal only then the final solution is optimal.
42. Dynamic Programming and
Chain Matrix Multiplication
In mathematics or computer science, Dynamic
Programming is a method for solving complex problems
by breaking them down into simpler sub-problems. So,
Matrix chain multiplication is an ideal example that
demonstrates utility of dynamic programming.
Engineering applications often have to multiply a large
chain of matrices with large dimensions, for example: 100
matrices of dimension 100×100. We can multiply this chain
of matrices in many different ways, but we will go for that
way which takes lower computations.
43. Dynamic Programming of Chain Matrix
Multiplication
For example, we are going to multiply 4 matrices:
M1 = 2 x 3
M2 = 3 x 4
M3 = 4 x 5
M4 = 5 x 7
And we have conditions for multiplying matrices:
• We can multiply only two matrices at a time.
• When we go to multiply 2 matrices, the number of columns of 1st matrix should
be same as the number of rows of 2nd matrix.
44. M1 = 2 x 3
M2 = 3 x 4
M3 = 4 x 5
M4 = 5 x 7
( M1 M2 )( M3 M4 ) = 220
(( M1 M2 ) M3 ) M4 = 134
M1 ( M2 ( M3 M4 ) = 266
( M1 ( M2 M3 ) M4 = 160
M1 (( M2 M3 ) M4 ) = 207
We can multiply the chain of matrices by following those
conditions in these ways:
Numbers of the rightmost side is number of total scalar
multiplication. So we have realized that we can reduce
the number of total multiplication and this reduced
time is a fact for a large chain of matrices.
45. Algorithm and Mechanism
Renaming matrices as Mi and dimensions as Pi - 1 x Pi , we have got:
M1 = P0 x P1
M2 = P1 x P2
M3 = P2 x P3
M4 = P3 x P4
| | |
Mi = Pi – 1 x Pi
46. We will use a formula:
Where C i, j means Mi to Mj .
i.e.: C 1, 4 means M1 to M4 .
And we will use a variable 'k' as follows:
M1 |k=1 M2 M3 M4
M1 M2 |k=2 M3 M4
M1 M2 M3 |k=3 M4
47. The thing we’re going to do is to apply above formula for every 'k' in the range 'i' to
'j' and pick the lowest value every step.
C 1 , 4 = min ( C1 , 1 + C2 , 4 + P0 * P1 * P4 , C1 , 2 + C3 , 4 + P0 * P2 * P4 , C1 , 3 + C4 , 4 + P0 *
P3 * P4 ) = min ( 207, 220, 134 ) = 134
C 2, 4 = min ( C2 , 2 + C3 , 4 + P1 * P2 * P4 , C2 , 3 + C4 , 4 + P1 * P3 * P4 ) = min ( 224, 165 )
= 165
C 1, 3 = min ( C1 , 1 + C2 , 3 + P0 * P1 * P3 , C1 , 2 + C3 , 3 + P0 * P2 * P3 ) = min ( 90, 64 ) =
64
C 1, 2 = P0 * P1 * P2 = 24
C 2, 3 = P1 * P2 * P3 = 60
C 3, 4 = P2 * P3 * P4 = 140
48. Pseudocode
1. int Chain( int i, int j )
2. {
3. int min = 10000, value, k;
4. if( i == j ){
5. return 0;
6. }
7. else{
8. for( k = i; k < j; k++ ){
9. value = (Chain(i, k) + Chain(k + 1, j) + (dimensions[i-1] *
dimensions[k] * dimensions[j]));
10. if( min > value ){
11. min = value;
12. mat[i][j] = k;
13. }
14. }
15. }
16. return min;
49. 1. int main(void)
2. {
3. int result, i;
4. printf("Enter number of matrices: ");
5. scanf("%d", &n);
6. printf("Enter dimensions : ");
7. for( i = 0; i <= n; i++ ){
8. scanf("%d", &dimensions[i]);
9. }
10. result = Chain(1, n);
11. printf("nTotal number of multiplications: %d andn", result);
12. printf("Multiplication order is: ");
13. PrintOrder( 1, n );
14. printf("n");
15. }
54. Fibonacci Numbers
How many summations are there? Using Golden Ratio
As you go farther and farther to the right in this sequence, the ratio of a term to the one
before it will get closer and closer to the Golden Ratio.
Our recursion tree has only 0s and 1s as leaves, thus we have 1.6n summations
Running time is exponential!
55. Fibonacci Numbers
We can calculate Fn in linear time by remembering solutions to the solved
subproblems – dynamic programming
Compute solution in a bottom-up fashion
In this case, only two values need to be remembered at any time
56.
57. Steps for problem solving
There are four steps of problem solving
Optimal Solution Structure
Recursive Solution
Optimal Solution Value
Optimal Solution
58. Problem Definition
Problem: Given all these costs, what stations should be chosen from line 1
and from line 2 for minimizing the total time for car assembly.
“Brute force” is to try all possibilities.
requires to examine Omega(2n) possibilities
Trying all 2n subsets is infeasible when n is large.
Simple example : 2 station (2n) possibilities =4
start end
59. Step 1: Optimal Solution Structure
optimal substructure : choosing the best path to Sij.
The structure of the fastest way through the factory (from
the starting point)
The fastest possible way to get through Si,1 (i = 1, 2)
Only one way: from entry starting point to Si,1
take time is entry time (ei)
60. Step 1: Optimal Solution Structure
The fastest possible way to get through Si,j (i = 1, 2) (j = 2, 3, ..., n).
Two choices:
Stay in the same line: Si,j-1 Si,j
Time is Ti,j-1 + ai,j
If the fastest way through Si,j is through Si,j-1, it must have taken a fastest way through Si,j-1
Transfer to other line: S3-i,j-1 Si,j
Time is T3-i,j-1 + t3-i,j-1 + ai,j
Same as above
61. Step 1: Optimal Solution Structure
An optimal solution to a problem
finding the fastest way to get through Si,j
contains within it an optimal solution to sub-problems
finding the fastest way to get through either Si,j-1 or S3-i,j-1
Fastest way from starting point to Si,j is either:
The fastest way from starting point to Si,j-1 and then directly from Si,j-1 to
Si,j
or
The fastest way from starting point to S3-i,j-1 then a transfer from line 3-i to
line i and finally to Si,j
Optimal Substructure.
65. Step 2: Recursive Solution
Define the value of an optimal solution recursively in terms of the
optimal solution to sub-problems
Sub-problem here
finding the fastest way through station j on both lines (i=1,2)
Let fi [j] be the fastest possible time to go from starting point through Si,j
The fastest time to go all the way through the factory: f*
x1 and x2 are the exit times from lines 1 and 2, respectively
66. Step 2: Recursive Solution
The fastest time to go through Si,j
e1 and e2 are the entry times for lines 1 and 2
69. Step 2: Recursive Solution
To help us keep track of how to construct an optimal solution, let us define
li[j ]: line # whose station j-1 is used in a fastest way through Si,j (i = 1, 2, and j =
2, 3,..., n)
we avoid defining li[1] because no station precedes station 1 on either lines.
We also define
l*: the line whose station n is used in a fastest way through the entire factory
70. Step 2: Recursive Solution
Using the values of l* and li[j] shown in Figure (b) in next slide, we would trace a
fastest way through the factory shown in part (a) as follows
The fastest total time comes from choosing stations
Line 1: 1, 3, & 6 Line 2: 2, 4, & 5
80. Step 4: Optimal Solution
Constructing the fastest way through the factory
81.
82. 1)`the process of breaking down a complex problem into
a series of interrelated sub problems often provides
insight into the nature of problem
2) Because dynamic programming is an approach to
optimization rather than a technique it has flexibility
that allows application to other types of mathematical
programming problems
3) The computational procedure in dynamic programming
allows for a built in form of sensitivity analysis based
on state variables and on variables represented by
stages
4)Dynamic programming achieves computational savings
over complete enumeration.
83. 1.)more expertise is required in solving dynamic programming problem then
using other methods
2.)lack of general algorithm like the simplex method. It restricts computer
codes necessary for inexpensive and widespread use
3.)the biggest problem is dimensionality. This problems occurs when a
particular application is characterized by multiple states. It creates lot of
problem for computers capabilities & is time consuming
Di