Analysis of Algorithms CS 477/677

Slides:



Advertisements
Similar presentations
Dynamic Programming.
Advertisements

Lecture 8: Dynamic Programming Shang-Hua Teng. Longest Common Subsequence Biologists need to measure how similar strands of DNA are to determine how closely.
15.Dynamic Programming Hsu, Lih-Hsing. Computer Theory Lab. Chapter 15P.2 Dynamic programming Dynamic programming is typically applied to optimization.
Overview What is Dynamic Programming? A Sequence of 4 Steps
CS Section 600 CS Section 002 Dr. Angela Guercio Spring 2010.
CS420 Lecture 9 Dynamic Programming. Optimization Problems In optimization problems a set of choices are to be made to arrive at an optimum, and sub problems.
Introduction to Algorithms
1 Dynamic Programming Jose Rolim University of Geneva.
Comp 122, Fall 2004 Dynamic Programming. dynprog - 2 Lin / Devi Comp 122, Spring 2004 Longest Common Subsequence  Problem: Given 2 sequences, X =  x.
Chapter 15 Dynamic Programming Lee, Hsiu-Hui Ack: This presentation is based on the lecture slides from Hsu, Lih-Hsing, as well as various materials from.
1 Dynamic Programming (DP) Like divide-and-conquer, solve problem by combining the solutions to sub-problems. Differences between divide-and-conquer and.
Dynamic Programming (pro-gram)
Dynamic Programming Part 1: intro and the assembly-line scheduling problem.
Dynamic Programming Lets begin by looking at the Fibonacci sequence.
Dynamic Programming Lecture 9 Asst. Prof. Dr. İlker Kocabaş 1.
Analysis of Algorithms CS 477/677 Instructor: Monica Nicolescu Lecture 11.
UMass Lowell Computer Science Analysis of Algorithms Prof. Karen Daniels Fall, 2006 Lecture 1 (Part 3) Design Patterns for Optimization Problems.
UMass Lowell Computer Science Analysis of Algorithms Prof. Karen Daniels Spring, 2002 Lecture 2 Tuesday, 2/5/02 Dynamic Programming.
Dynamic Programming Reading Material: Chapter 7..
Dynamic Programming CIS 606 Spring 2010.
UMass Lowell Computer Science Analysis of Algorithms Prof. Karen Daniels Spring, 2009 Design Patterns for Optimization Problems Dynamic Programming.
Dynamic Programming Code
UMass Lowell Computer Science Analysis of Algorithms Prof. Karen Daniels Fall, 2002 Lecture 1 (Part 3) Tuesday, 9/3/02 Design Patterns for Optimization.
CS3381 Des & Anal of Alg ( SemA) City Univ of HK / Dept of CS / Helena Wong 4. Dynamic Programming - 1 Dynamic.
CPSC 411 Design and Analysis of Algorithms Set 5: Dynamic Programming Prof. Jennifer Welch Spring 2011 CPSC 411, Spring 2011: Set 5 1.
Dynamic Programming1. 2 Outline and Reading Matrix Chain-Product (§5.3.1) The General Technique (§5.3.2) 0-1 Knapsack Problem (§5.3.3)
UNC Chapel Hill Lin/Manocha/Foskey Optimization Problems In which a set of choices must be made in order to arrive at an optimal (min/max) solution, subject.
Dynamic Programming Reading Material: Chapter 7 Sections and 6.
Analysis of Algorithms
11-1 Matrix-chain Multiplication Suppose we have a sequence or chain A 1, A 2, …, A n of n matrices to be multiplied –That is, we want to compute the product.
Lecture 7 Topics Dynamic Programming
Longest Common Subsequence
Dynamic Programming Introduction to Algorithms Dynamic Programming CSE 680 Prof. Roger Crawfis.
1 Dynamic Programming 2012/11/20. P.2 Dynamic Programming (DP) Dynamic programming Dynamic programming is typically applied to optimization problems.
First Ingredient of Dynamic Programming
Dynamic Programming UNC Chapel Hill Z. Guo.
October 21, Algorithms and Data Structures Lecture X Simonas Šaltenis Nykredit Center for Database Research Aalborg University
Dynamic Programming. Well known algorithm design techniques:. –Divide-and-conquer algorithms Another strategy for designing algorithms is dynamic programming.
CS 5243: Algorithms Dynamic Programming Dynamic Programming is applicable when sub-problems are dependent! In the case of Divide and Conquer they are.
COSC 3101A - Design and Analysis of Algorithms 7 Dynamic Programming Assembly-Line Scheduling Matrix-Chain Multiplication Elements of DP Many of these.
CS 8833 Algorithms Algorithms Dynamic Programming.
COSC 3101A - Design and Analysis of Algorithms 8 Elements of DP Memoization Longest Common Subsequence Greedy Algorithms Many of these slides are taken.
Algorithms: Design and Analysis Summer School 2013 at VIASM: Random Structures and Algorithms Lecture 4: Dynamic Programming Phan Th ị Hà D ươ ng 1.
Analysis of Algorithms CS 477/677 Instructor: Monica Nicolescu Lecture 14.
Algorithmics - Lecture 121 LECTURE 11: Dynamic programming - II -
Optimization Problems In which a set of choices must be made in order to arrive at an optimal (min/max) solution, subject to some constraints. (There may.
Dynamic Programming1. 2 Outline and Reading Matrix Chain-Product (§5.3.1) The General Technique (§5.3.2) 0-1 Knapsack Problem (§5.3.3)
Computer Sciences Department1.  Property 1: each node can have up to two successor nodes (children)  The predecessor node of a node is called its.
15.Dynamic Programming. Computer Theory Lab. Chapter 15P.2 Dynamic programming Dynamic programming is typically applied to optimization problems. In such.
Chapter 15 Dynamic Programming Lee, Hsiu-Hui Ack: This presentation is based on the lecture slides from Hsu, Lih-Hsing, as well as various materials from.
TU/e Algorithms (2IL15) – Lecture 3 1 DYNAMIC PROGRAMMING
TU/e Algorithms (2IL15) – Lecture 4 1 DYNAMIC PROGRAMMING II
1 Chapter 15-2: Dynamic Programming II. 2 Matrix Multiplication Let A be a matrix of dimension p x q and B be a matrix of dimension q x r Then, if we.
Dynamic Programming Csc 487/687 Computing for Bioinformatics.
Dynamic Programming Typically applied to optimization problems
Seminar on Dynamic Programming.
Advanced Design and Analysis Techniques
Dynamic Programming.
Analysis of Algorithms CS 477/677
Data Structure and Algorithms
Algorithms CSCI 235, Spring 2019 Lecture 28 Dynamic Programming III
Introduction to Algorithms: Dynamic Programming
DYNAMIC PROGRAMMING.
Algorithm Design Techniques Greedy Approach vs Dynamic Programming
Analysis of Algorithms CS 477/677
Matrix Chain Multiplication
Algorithms CSCI 235, Spring 2019 Lecture 27 Dynamic Programming II
Analysis of Algorithms CS 477/677
Seminar on Dynamic Programming.
Presentation transcript:

Analysis of Algorithms CS 477/677 Dynamic Programming Instructor: George Bebis (Chapter 15)

Dynamic Programming An algorithm design technique (like divide and conquer) Divide and conquer Partition the problem into independent subproblems Solve the subproblems recursively Combine the solutions to solve the original problem

Dynamic Programming Applicable when subproblems are not independent Subproblems share subsubproblems E.g.: Combinations: A divide and conquer approach would repeatedly solve the common subproblems Dynamic programming solves every subproblem just once and stores the answer in a table n k n-1 k-1 = + n 1 n =1 =1

Example: Combinations k n-1 k-1 = +

Dynamic Programming Used for optimization problems A set of choices must be made to get an optimal solution Find a solution with the optimal value (minimum or maximum) There may be many solutions that lead to an optimal value Our goal: find an optimal solution

Dynamic Programming Algorithm Characterize the structure of an optimal solution Recursively define the value of an optimal solution Compute the value of an optimal solution in a bottom-up fashion Construct an optimal solution from computed information (not always necessary)

Assembly Line Scheduling Automobile factory with two assembly lines Each line has n stations: S1,1, . . . , S1,n and S2,1, . . . , S2,n Corresponding stations S1, j and S2, j perform the same function but can take different amounts of time a1, j and a2, j Entry times are: e1 and e2; exit times are: x1 and x2

Assembly Line Scheduling After going through a station, can either: stay on same line at no cost, or transfer to other line: cost after Si,j is ti,j , j = 1, . . . , n - 1

Assembly Line Scheduling Problem: what stations should be chosen from line 1 and which from line 2 in order to minimize the total time through the factory for one car?

One Solution Brute force Solution: Enumerate all possibilities of selecting stations Compute how long it takes in each case and choose the best one Solution: There are 2n possible ways to choose stations Infeasible when n is large!! 1 1 if choosing line 1 at step j (= n) 2 3 4 n 0 if choosing line 2 at step j (= 3)

1. Structure of the Optimal Solution How do we compute the minimum time of going through a station?

1. Structure of the Optimal Solution Let’s consider all possible ways to get from the starting point through station S1,j We have two choices of how to get to S1, j: Through S1, j - 1, then directly to S1, j Through S2, j - 1, then transfer over to S1, j a1,j a1,j-1 a2,j-1 t2,j-1 S1,j S1,j-1 S2,j-1 Line 1 Line 2

1. Structure of the Optimal Solution Suppose that the fastest way through S1, j is through S1, j – 1 We must have taken a fastest way from entry through S1, j – 1 If there were a faster way through S1, j - 1, we would use it instead Similarly for S2, j – 1 a1,j a1,j-1 a2,j-1 t2,j-1 S1,j S1,j-1 S2,j-1 Line 1 Optimal Substructure Line 2

Optimal Substructure Generalization: an optimal solution to the problem “find the fastest way through S1, j” contains within it an optimal solution to subproblems: “find the fastest way through S1, j - 1 or S2, j – 1”. This is referred to as the optimal substructure property We use this property to construct an optimal solution to a problem from optimal solutions to subproblems

2. A Recursive Solution Define the value of an optimal solution in terms of the optimal solution to subproblems

2. A Recursive Solution (cont.) Definitions: f* : the fastest time to get through the entire factory fi[j] : the fastest time to get from the starting point through station Si,j f* = min (f1[n] + x1, f2[n] + x2)

2. A Recursive Solution (cont.) Base case: j = 1, i=1,2 (getting through station 1) f1[1] = e1 + a1,1 f2[1] = e2 + a2,1

2. A Recursive Solution (cont.) General Case: j = 2, 3, …,n, and i = 1, 2 Fastest way through S1, j is either: the way through S1, j - 1 then directly through S1, j, or f1[j - 1] + a1,j the way through S2, j - 1, transfer from line 2 to line 1, then through S1, j f2[j -1] + t2,j-1 + a1,j f1[j] = min(f1[j - 1] + a1,j ,f2[j -1] + t2,j-1 + a1,j) a1,j a1,j-1 a2,j-1 t2,j-1 S1,j S1,j-1 S2,j-1 Line 1 Line 2

2. A Recursive Solution (cont.) e1 + a1,1 if j = 1 f1[j] = min(f1[j - 1] + a1,j ,f2[j -1] + t2,j-1 + a1,j) if j ≥ 2 e2 + a2,1 if j = 1 f2[j] = min(f2[j - 1] + a2,j ,f1[j -1] + t1,j-1 + a2,j) if j ≥ 2

3. Computing the Optimal Solution f* = min (f1[n] + x1, f2[n] + x2) f1[j] = min(f1[j - 1] + a1,j ,f2[j -1] + t2,j-1 + a1,j) f2[j] = min(f2[j - 1] + a2,j ,f1[j -1] + t1,j-1 + a2,j) Solving top-down would result in exponential running time 1 2 3 4 5 f1[j] f1(1) f1(2) f1(3) f1(4) f1(5) f2[j] f2(1) f2(2) f2(3) f2(4) f2(5) 4 times 2 times

3. Computing the Optimal Solution For j ≥ 2, each value fi[j] depends only on the values of f1[j – 1] and f2[j - 1] Idea: compute the values of fi[j] as follows: Bottom-up approach First find optimal solutions to subproblems Find an optimal solution to the problem from the subproblems in increasing order of j 1 2 3 4 5 f1[j] f2[j]

Example e1 + a1,1, if j = 1 f1[j] = min(f1[j - 1] + a1,j ,f2[j -1] + t2,j-1 + a1,j) if j ≥ 2 1 2 3 4 5 f1[j] 9 18[1] 20[2] 24[1] 32[1] f* = 35[1] f2[j] 12 16[1] 22[2] 25[1] 30[2]

FASTEST-WAY(a, t, e, x, n) O(N) f1[1] ← e1 + a1,1 f2[1] ← e2 + a2,1 for j ← 2 to n do if f1[j - 1] + a1,j ≤ f2[j - 1] + t2, j-1 + a1, j then f1[j] ← f1[j - 1] + a1, j l1[j] ← 1 else f1[j] ← f2[j - 1] + t2, j-1 + a1, j l1[j] ← 2 if f2[j - 1] + a2, j ≤ f1[j - 1] + t1, j-1 + a2, j then f2[j] ← f2[j - 1] + a2, j l2[j] ← 2 else f2[j] ← f1[j - 1] + t1, j-1 + a2, j l2[j] ← 1 Compute initial values of f1 and f2 O(N) Compute the values of f1[j] and l1[j] Compute the values of f2[j] and l2[j]

FASTEST-WAY(a, t, e, x, n) (cont.) if f1[n] + x1 ≤ f2[n] + x2 then f* = f1[n] + x1 l* = 1 else f* = f2[n] + x2 l* = 2 Compute the values of the fastest time through the entire factory

4. Construct an Optimal Solution Alg.: PRINT-STATIONS(l, n) i ← l* print “line ” i “, station ” n for j ← n downto 2 do i ←li[j] print “line ” i “, station ” j - 1 1 2 3 4 5 f1[j]/l1[j] 9 18[1] 20[2] 24[1] 32[1] l* = 1 f2[j]/l2[j] 12 16[1] 22[2] 25[1] 30[2]

Matrix-Chain Multiplication Problem: given a sequence A1, A2, …, An, compute the product: A1  A2  An Matrix compatibility: C = A  B C=A1  A2  Ai  Ai+1  An colA = rowB coli = rowi+1 rowC = rowA rowC = rowA1 colC = colB colC = colAn

rows[A]  cols[A]  cols[B] MATRIX-MULTIPLY(A, B) if columns[A]  rows[B] then error “incompatible dimensions” else for i  1 to rows[A] do for j  1 to columns[B] do C[i, j] = 0 for k  1 to columns[A] do C[i, j]  C[i, j] + A[i, k] B[k, j] rows[A]  cols[A]  cols[B] multiplications k j cols[B] j cols[B] k i i * = A B rows[A] C rows[A]

Matrix-Chain Multiplication In what order should we multiply the matrices? A1  A2  An Parenthesize the product to get the order in which matrices are multiplied E.g.: A1  A2  A3 = ((A1  A2)  A3) = (A1  (A2  A3)) Which one of these orderings should we choose? The order in which we multiply the matrices has a significant impact on the cost of evaluating the product

Example A1  A2  A3 A1: 10 x 100 A2: 100 x 5 A3: 5 x 50 1. ((A1  A2)  A3): A1  A2 = 10 x 100 x 5 = 5,000 (10 x 5) ((A1  A2)  A3) = 10 x 5 x 50 = 2,500 Total: 7,500 scalar multiplications 2. (A1  (A2  A3)): A2  A3 = 100 x 5 x 50 = 25,000 (100 x 50) (A1  (A2  A3)) = 10 x 100 x 50 = 50,000 Total: 75,000 scalar multiplications one order of magnitude difference!!

Matrix-Chain Multiplication: Problem Statement Given a chain of matrices A1, A2, …, An, where Ai has dimensions pi-1x pi, fully parenthesize the product A1  A2  An in a way that minimizes the number of scalar multiplications. A1  A2  Ai  Ai+1  An p0 x p1 p1 x p2 pi-1 x pi pi x pi+1 pn-1 x pn

What is the number of possible parenthesizations? Exhaustively checking all possible parenthesizations is not efficient! It can be shown that the number of parenthesizations grows as Ω(4n/n3/2) (see page 333 in your textbook)

1. The Structure of an Optimal Parenthesization Notation: Ai…j = Ai Ai+1  Aj, i  j Suppose that an optimal parenthesization of Ai…j splits the product between Ak and Ak+1, where i  k < j Ai…j = Ai Ai+1  Aj = Ai Ai+1  Ak Ak+1  Aj = Ai…k Ak+1…j

Optimal Substructure Ai…j = Ai…k Ak+1…j The parenthesization of the “prefix” Ai…k must be an optimal parentesization If there were a less costly way to parenthesize Ai…k, we could substitute that one in the parenthesization of Ai…j and produce a parenthesization with a lower cost than the optimum  contradiction! An optimal solution to an instance of the matrix-chain multiplication contains within it optimal solutions to subproblems

2. A Recursive Solution Subproblem: determine the minimum cost of parenthesizing Ai…j = Ai Ai+1  Aj for 1  i  j  n Let m[i, j] = the minimum number of multiplications needed to compute Ai…j full problem (A1..n): m[1, n] i = j: Ai…i = Ai  m[i, i] = 0, for i = 1, 2, …, n

2. A Recursive Solution Consider the subproblem of parenthesizing Ai…j = Ai Ai+1  Aj for 1  i  j  n = Ai…k Ak+1…j for i  k < j Assume that the optimal parenthesization splits the product Ai Ai+1  Aj at k (i  k < j) m[i, j] = pi-1pkpj m[i, k] m[k+1,j] m[i, k] + m[k+1, j] + pi-1pkpj min # of multiplications to compute Ai…k min # of multiplications to compute Ak+1…j # of multiplications to compute Ai…kAk…j

2. A Recursive Solution (cont.) m[i, j] = m[i, k] + m[k+1, j] + pi-1pkpj We do not know the value of k There are j – i possible values for k: k = i, i+1, …, j-1 Minimizing the cost of parenthesizing the product Ai Ai+1  Aj becomes: 0 if i = j m[i, j] = min {m[i, k] + m[k+1, j] + pi-1pkpj} if i < j ik<j

3. Computing the Optimal Costs 0 if i = j m[i, j] = min {m[i, k] + m[k+1, j] + pi-1pkpj} if i < j ik<j Computing the optimal solution recursively takes exponential time! How many subproblems? Parenthesize Ai…j for 1  i  j  n One problem for each choice of i and j 1 2 3 n n  (n2) j 3 2 1 i

3. Computing the Optimal Costs (cont.) 0 if i = j m[i, j] = min {m[i, k] + m[k+1, j] + pi-1pkpj} if i < j ik<j How do we fill in the tables m[1..n, 1..n]? Determine which entries of the table are used in computing m[i, j] Ai…j = Ai…k Ak+1…j Subproblems’ size is one less than the original size Idea: fill in m such that it corresponds to solving problems of increasing length

3. Computing the Optimal Costs (cont.) 0 if i = j m[i, j] = min {m[i, k] + m[k+1, j] + pi-1pkpj} if i < j ik<j Length = 1: i = j, i = 1, 2, …, n Length = 2: j = i + 1, i = 1, 2, …, n-1 second first 1 2 3 n m[1, n] gives the optimal solution to the problem n j 3 Compute rows from bottom to top and from left to right 2 1 i

Example: min {m[i, k] + m[k+1, j] + pi-1pkpj} m[2, 2] + m[3, 5] + p1p2p5 m[2, 3] + m[4, 5] + p1p3p5 m[2, 4] + m[5, 5] + p1p4p5 k = 2 m[2, 5] = min k = 3 k = 4 1 2 3 4 5 6 6 5 Values m[i, j] depend only on values that have been previously computed 4 j 3 2 1 i

Example min {m[i, k] + m[k+1, j] + pi-1pkpj} 2 3 Compute A1  A2  A3 A1: 10 x 100 (p0 x p1) A2: 100 x 5 (p1 x p2) A3: 5 x 50 (p2 x p3) m[i, i] = 0 for i = 1, 2, 3 m[1, 2] = m[1, 1] + m[2, 2] + p0p1p2 (A1A2) = 0 + 0 + 10 *100* 5 = 5,000 m[2, 3] = m[2, 2] + m[3, 3] + p1p2p3 (A2A3) = 0 + 0 + 100 * 5 * 50 = 25,000 m[1, 3] = min m[1, 1] + m[2, 3] + p0p1p3 = 75,000 (A1(A2A3)) m[1, 2] + m[3, 3] + p0p2p3 = 7,500 ((A1A2)A3) 7500 2 25000 2 3 5000 1 2 1

Matrix-Chain-Order(p) O(N3)

4. Construct the Optimal Solution In a similar matrix s we keep the optimal values of k s[i, j] = a value of k such that an optimal parenthesization of Ai..j splits the product between Ak and Ak+1 1 2 3 n k n j 3 2 1

4. Construct the Optimal Solution s[1, n] is associated with the entire product A1..n The final matrix multiplication will be split at k = s[1, n] A1..n = A1..s[1, n]  As[1, n]+1..n For each subproduct recursively find the corresponding value of k that results in an optimal parenthesization 1 2 3 n n j 3 2 1

4. Construct the Optimal Solution s[i, j] = value of k such that the optimal parenthesization of Ai Ai+1  Aj splits the product between Ak and Ak+1 1 2 3 4 5 6 3 5 - 4 1 2 6 s[1, n] = 3  A1..6 = A1..3 A4..6 s[1, 3] = 1  A1..3 = A1..1 A2..3 s[4, 6] = 5  A4..6 = A4..5 A6..6 5 4 3 j 2 1 i

4. Construct the Optimal Solution (cont.) PRINT-OPT-PARENS(s, i, j) if i = j then print “A”i else print “(” PRINT-OPT-PARENS(s, i, s[i, j]) PRINT-OPT-PARENS(s, s[i, j] + 1, j) print “)” 1 2 3 4 5 6 3 5 - 4 1 2 6 5 4 j 3 2 1 i

Example: A1  A6 ( ( A1 ( A2 A3 ) ) ( ( A4 A5 ) A6 ) ) 3 5 - 4 1 2 PRINT-OPT-PARENS(s, i, j) if i = j then print “A”i else print “(” PRINT-OPT-PARENS(s, i, s[i, j]) PRINT-OPT-PARENS(s, s[i, j] + 1, j) print “)” s[1..6, 1..6] 1 2 3 4 5 6 3 5 - 4 1 2 6 5 4 j 3 2 P-O-P(s, 1, 6) s[1, 6] = 3 i = 1, j = 6 “(“ P-O-P (s, 1, 3) s[1, 3] = 1 i = 1, j = 3 “(“ P-O-P(s, 1, 1)  “A1” P-O-P(s, 2, 3) s[2, 3] = 2 i = 2, j = 3 “(“ P-O-P (s, 2, 2)  “A2” P-O-P (s, 3, 3)  “A3” “)” 1 i …

Memoization Top-down approach with the efficiency of typical dynamic programming approach Maintaining an entry in a table for the solution to each subproblem memoize the inefficient recursive algorithm When a subproblem is first encountered its solution is computed and stored in that table Subsequent “calls” to the subproblem simply look up that value

Memoized Matrix-Chain Alg.: MEMOIZED-MATRIX-CHAIN(p) n  length[p] – 1 for i  1 to n do for j  i to n do m[i, j]   return LOOKUP-CHAIN(p, 1, n) Initialize the m table with large values that indicate whether the values of m[i, j] have been computed Top-down approach

Memoized Matrix-Chain Alg.: LOOKUP-CHAIN(p, i, j) if m[i, j] <  then return m[i, j] if i = j then m[i, j]  0 else for k  i to j – 1 do q  LOOKUP-CHAIN(p, i, k) + LOOKUP-CHAIN(p, k+1, j) + pi-1pkpj if q < m[i, j] then m[i, j]  q return m[i, j] Running time is O(n3)

Dynamic Progamming vs. Memoization Advantages of dynamic programming vs. memoized algorithms No overhead for recursion, less overhead for maintaining the table The regular pattern of table accesses may be used to reduce time or space requirements Advantages of memoized algorithms vs. dynamic programming Some subproblems do not need to be solved

Matrix-Chain Multiplication - Summary Both the dynamic programming approach and the memoized algorithm can solve the matrix-chain multiplication problem in O(n3) Both methods take advantage of the overlapping subproblems property There are only (n2) different subproblems Solutions to these problems are computed only once Without memoization the natural recursive algorithm runs in exponential time

Elements of Dynamic Programming Optimal Substructure An optimal solution to a problem contains within it an optimal solution to subproblems Optimal solution to the entire problem is build in a bottom-up manner from optimal solutions to subproblems Overlapping Subproblems If a recursive algorithm revisits the same subproblems over and over  the problem has overlapping subproblems

Parameters of Optimal Substructure How many subproblems are used in an optimal solution for the original problem Assembly line: Matrix multiplication: How many choices we have in determining which subproblems to use in an optimal solution One subproblem (the line that gives best time) Two subproblems (subproducts Ai..k, Ak+1..j) Two choices (line 1 or line 2) j - i choices for k (splitting the product)

Parameters of Optimal Substructure Intuitively, the running time of a dynamic programming algorithm depends on two factors: Number of subproblems overall How many choices we look at for each subproblem Assembly line (n) subproblems (n stations) 2 choices for each subproblem Matrix multiplication: (n2) subproblems (1  i  j  n) At most n-1 choices (n) overall (n3) overall

Longest Common Subsequence Given two sequences X = x1, x2, …, xm Y = y1, y2, …, yn find a maximum length common subsequence (LCS) of X and Y E.g.: X = A, B, C, B, D, A, B Subsequences of X: A subset of elements in the sequence taken in order A, B, D, B, C, D, B, etc.

Example X = A, B, C, B, D, A, B X = A, B, C, B, D, A, B Y = B, D, C, A, B, A Y = B, D, C, A, B, A B, C, B, A and B, D, A, B are longest common subsequences of X and Y (length = 4) B, C, A, however is not a LCS of X and Y

Brute-Force Solution For every subsequence of X, check whether it’s a subsequence of Y There are 2m subsequences of X to check Each subsequence takes (n) time to check scan Y for first letter, from there scan for second, and so on Running time: (n2m)

Making the choice X = A, B, D, E Y = Z, B, E Choice: include one element into the common sequence (E) and solve the resulting subproblem X = A, B, D, G Y = Z, B, D Choice: exclude an element from a string and solve the resulting subproblem

Notations Given a sequence X = x1, x2, …, xm we define the i-th prefix of X, for i = 0, 1, 2, …, m Xi = x1, x2, …, xi c[i, j] = the length of a LCS of the sequences Xi = x1, x2, …, xi and Yj = y1, y2, …, yj

A Recursive Solution Case 1: xi = yj e.g.: Xi = A, B, D, E Yj = Z, B, E Append xi = yj to the LCS of Xi-1 and Yj-1 Must find a LCS of Xi-1 and Yj-1  optimal solution to a problem includes optimal solutions to subproblems c[i, j] = c[i - 1, j - 1] + 1

A Recursive Solution Case 2: xi  yj e.g.: Xi = A, B, D, G Yj = Z, B, D Must solve two problems find a LCS of Xi-1 and Yj: Xi-1 = A, B, D and Yj = Z, B, D find a LCS of Xi and Yj-1: Xi = A, B, D, G and Yj = Z, B Optimal solution to a problem includes optimal solutions to subproblems c[i, j] = max { c[i - 1, j], c[i, j-1] }

Overlapping Subproblems To find a LCS of X and Y we may need to find the LCS between X and Yn-1 and that of Xm-1 and Y Both the above subproblems has the subproblem of finding the LCS of Xm-1 and Yn-1 Subproblems share subsubproblems

3. Computing the Length of the LCS 0 if i = 0 or j = 0 c[i, j] = c[i-1, j-1] + 1 if xi = yj max(c[i, j-1], c[i-1, j]) if xi  yj 1 2 n yj: y1 y2 yn xi 1 x1 first 2 x2 second i m xm j

Additional Information A matrix b[i, j]: For a subproblem [i, j] it tells us what choice was made to obtain the optimal value If xi = yj b[i, j] = “ ” Else, if c[i - 1, j] ≥ c[i, j-1] b[i, j] = “  ” else b[i, j] = “  ” 0 if i,j = 0 c[i, j] = c[i-1, j-1] + 1 if xi = yj max(c[i, j-1], c[i-1, j]) if xi  yj 1 2 3 n b & c: yj: A C D F xi 1 A 2 B c[i,j-1] c[i-1,j] i 3 C m D j

LCS-LENGTH(X, Y, m, n) Running time: (mn) for i ← 1 to m do c[i, 0] ← 0 for j ← 0 to n do c[0, j] ← 0 do for j ← 1 to n do if xi = yj then c[i, j] ← c[i - 1, j - 1] + 1 b[i, j ] ← “ ” else if c[i - 1, j] ≥ c[i, j - 1] then c[i, j] ← c[i - 1, j] b[i, j] ← “↑” else c[i, j] ← c[i, j - 1] b[i, j] ← “←” return c and b The length of the LCS if one of the sequences is empty is zero Case 1: xi = yj Case 2: xi  yj Running time: (mn)

Example X = A, B, C, B, D, A Y = B, D, C, A, B, A If xi = yj 0 if i = 0 or j = 0 c[i, j] = c[i-1, j-1] + 1 if xi = yj max(c[i, j-1], c[i-1, j]) if xi  yj X = A, B, C, B, D, A Y = B, D, C, A, B, A 1 2 3 4 5 6 If xi = yj b[i, j] = “ ” Else if c[i - 1, j] ≥ c[i, j-1] b[i, j] = “  ” else b[i, j] = “  ” yj B D C A B A xi 1 A    1 1 1 2 B 1  1 2 1 1 2 3 C  1 2 2 4 B 1  2 3 3 5 D  1 2 3 6 A  1 2 3 4 7 B 1  2 3 4

4. Constructing a LCS Start at b[m, n] and follow the arrows When we encounter a “ “ in b[i, j]  xi = yj is an element of the LCS 1 2 3 4 5 6 yj B D C A B A xi 1 A    1 1 1 2 B 1  1 2 1 1 2 3 C  1 2 2 4 B 1  2 3 3 5 D  1 2 3 6 A  1 2 3 4 7 B 1  2 3 4

PRINT-LCS(b, X, i, j) if i = 0 or j = 0 then return if b[i, j] = “ ” then PRINT-LCS(b, X, i - 1, j - 1) print xi elseif b[i, j] = “↑” then PRINT-LCS(b, X, i - 1, j) else PRINT-LCS(b, X, i, j - 1) Initial call: PRINT-LCS(b, X, length[X], length[Y]) Running time: (m + n)

Improving the Code What can we say about how each entry c[i, j] is computed? It depends only on c[i -1, j - 1], c[i - 1, j], and c[i, j - 1] Eliminate table b and compute in O(1) which of the three values was used to compute c[i, j] We save (mn) space from table b However, we do not asymptotically decrease the auxiliary space requirements: still need table c

Improving the Code If we only need the length of the LCS LCS-LENGTH works only on two rows of c at a time The row being computed and the previous row We can reduce the asymptotic space requirements by storing only these two rows