Richard Anderson Lecture 20 Space Efficient LCS

Slides:



Advertisements
Similar presentations
CSEP 521 Applied Algorithms Richard Anderson Lecture 7 Dynamic Programming.
Advertisements

DYNAMIC PROGRAMMING. 2 Algorithmic Paradigms Greedy. Build up a solution incrementally, myopically optimizing some local criterion. Divide-and-conquer.
CSE 421 Algorithms Richard Anderson Lecture 19 Longest Common Subsequence.
Lecture 39 CSE 331 Dec 6, On Friday, Dec 10 hours-a-thon Atri: 2:00-3:30 (Bell 123) Jeff: 4:00-5:00 (Bell 224) Alex: 5:00-6:30 (Bell 242)
CSE 421 Algorithms Richard Anderson Lecture 19 Longest Common Subsequence.
CSE 421 Algorithms Richard Anderson Lecture 21 Shortest Paths.
CSCI-256 Data Structures & Algorithm Analysis Lecture Note: Some slides by Kevin Wayne. Copyright © 2005 Pearson-Addison Wesley. All rights reserved. 22.
Prof. Swarat Chaudhuri COMP 482: Design and Analysis of Algorithms Spring 2012 Lecture 17.
Sequence Alignment Tanya Berger-Wolf CS502: Algorithms in Computational Biology January 25, 2011.
CS38 Introduction to Algorithms Lecture 10 May 1, 2014.
Lecture 38 CSE 331 Dec 5, OHs today (only online 9:30-11)
CSEP 521 Applied Algorithms Richard Anderson Lecture 8 Network Flow.
CSE 421 Algorithms Richard Anderson Lecture 8 Optimal Caching Dijkstra’s algorithm.
CSE 421 Algorithms Richard Anderson Lecture 21 Shortest Paths and Network Flow.
CS502: Algorithms in Computational Biology
Lecture 32 CSE 331 Nov 16, 2016.
Richard Anderson Lecture 1
CSC317 Shortest path algorithms
Richard Anderson Lecture 29 NP-Completeness and course wrap-up
JinJu Lee & Beatrice Seifert CSE 5311 Fall 2005 Week 10 (Nov 1 & 3)
Greedy Algorithms / Minimum Spanning Tree Yin Tat Lee
Data Structures and Algorithms
Shortest Paths with Dynamic Programming
Lecture 7 All-Pairs Shortest Paths
Lecture 36 CSE 331 Nov 29, 2017.
Chapter 6 Dynamic Programming
Lecture 36 CSE 331 Nov 30, 2016.
Lecture 35 CSE 331 Nov 27, 2017.
Autumn 2016 Lecture 9 Dijkstra’s algorithm
Lecture 35 CSE 331 Nov 28, 2016.
Autumn 2015 Lecture 9 Dijkstra’s algorithm
Lecture 37 CSE 331 Dec 1, 2017.
Memory Efficient Longest Common Subsequence
Richard Anderson Lecture 20 LCS / Shortest Paths
Richard Anderson Lecture 9 Dijkstra’s algorithm
Richard Anderson Lecture 9 Dijkstra’s algorithm
Richard Anderson Lecture 19 Longest Common Subsequence
Autumn 2015 Lecture 10 Minimum Spanning Trees
Richard Anderson Lecture 21 Shortest Path Network Flow Introduction
Chapter 6 Dynamic Programming
Richard Anderson Lecture 21 Network Flow
Richard Anderson Lecture 30 NP-Completeness
Richard Anderson Lecture 21 Shortest Paths
Dynamic Programming Longest Path in a DAG, Yin Tat Lee
Lecture 32 CSE 331 Nov 15, 2017.
Lecture 33 CSE 331 Nov 14, 2014.
Richard Anderson Autumn 2016 Lecture 7
Lecture 8. Paradigm #6 Dynamic Programming
Fundamental Data Structures and Algorithms
Algorithms (2IL15) – Lecture 7
Lecture 14 Shortest Path (cont’d) Minimum Spanning Tree
Evaluation of the Course (Modified)
Lecture 36 CSE 331 Nov 28, 2011.
Lecture 36 CSE 331 Nov 30, 2012.
Longest Common Subsequence
Richard Anderson Lecture 19 Memory Efficient Dynamic Programming
Lecture 37 CSE 331 Dec 2, 2016.
Lecture 21 Network Flow, Part 1
Winter 2019 Lecture 9 Dijkstra’s algorithm
CSE 417: Algorithms and Computational Complexity
Richard Anderson Lecture 19 Longest Common Subsequence
Richard Anderson Lecture 20 Shortest Paths and Network Flow
Richard Anderson Lecture 19 Memory Efficient Dynamic Programming
Richard Anderson Lecture 18 Dynamic Programming
Lecture 13 Shortest Path (cont’d) Minimum Spanning Tree
Dynamic Programming.
Memory Efficient Dynamic Programming / Shortest Paths
Lecture 36 CSE 331 Nov 22, 2013.
Lecture 37 CSE 331 Dec 3, 2018.
Autumn 2019 Lecture 9 Dijkstra’s algorithm
Presentation transcript:

Richard Anderson Lecture 20 Space Efficient LCS CSE 421 Algorithms Richard Anderson Lecture 20 Space Efficient LCS

Longest Common Subsequence C=c1…cg is a subsequence of A=a1…am if C can be obtained by removing elements from A (but retaining order) LCS(A, B): A maximum length sequence that is a subsequence of both A and B Wednesday’s Result: O(mn) time, O(mn) space LCS Algorithm Today’s Result: O(mn) time, O(m+n) space LCS Algorithm

Digression: String Alignment Problem Align sequences with gaps Charge dx if character x is unmatched Charge gxy if character x is matched to character y Find alignment to minimize sum of costs CAT TGA AT CAGAT AGGA

Optimization Recurrence for the String Alignment Problem Charge dx if character x is unmatched Charge gxy if character x is matched to character y A = a1a2…am; B = b1b2…bn Opt[j, k] is the value of the minimum cost alignment a1a2…aj and b1b2…bk

Dynamic Programming Computation

Storing the path information A[1..m], B[1..n] for i := 1 to m Opt[i, 0] := 0; for j := 1 to n Opt[0,j] := 0; Opt[0,0] := 0; for i := 1 to m for j := 1 to n if A[i] = B[j] { Opt[i,j] := 1 + Opt[i-1,j-1]; Best[i,j] := Diag; } else if Opt[i-1, j] >= Opt[i, j-1] { Opt[i, j] := Opt[i-1, j], Best[i,j] := Left; } else { Opt[i, j] := Opt[i, j-1], Best[i,j] := Down; } b1…bn a1…am

Observations about the Algorithm The computation can be done in O(m+n) space if we only need one column of the Opt values or Best Values The algorithm can be run from either end of the strings

Divide and Conquer Algorithm Where does the best path cross the middle column? For a fixed i, and for each j, compute the LCS that has ai matched with bj

Constrained LCS LCSi,j(A,B): The LCS such that LCS4,3(abbacbb, cbbaa) a1,…,ai paired with elements of b1,…,bj ai+1,…am paired with elements of bj+1,…,bn LCS4,3(abbacbb, cbbaa)

A = RRSSRTTRTS B=RTSRRSTST Compute LCS5,0(A,B), LCS5,1(A,B), LCS5,2(A,B),…,LCS5,9(A,B)

A = RRSSRTTRTS B=RTSRRSTST Compute LCS5,0(A,B), LCS5,1(A,B), LCS5,2(A,B),…,LCS5,9(A,B) j left right 4 1 2 3 5 6 7 8 9

Computing the middle column From the left, compute LCS(a1…am/2,b1…bj) From the right, compute LCS(am/2+1…am,bj+1…bn) Add values for corresponding j’s Note – this is space efficient

Divide and Conquer A = a1,…,am B = b1,…,bn Find j such that Recurse LCS(a1…am/2, b1…bj) and LCS(am/2+1…am,bj+1…bn) yield optimal solution Recurse

Algorithm Analysis T(m,n) = T(m/2, j) + T(m/2, n-j) + cnm

Prove by induction that T(m,n) <= 2cmn

Shortest Path Problem Dijkstra’s Single Source Shortest Paths Algorithm O(mlog n) time, positive cost edges General case – handling negative edges If there exists a negative cost cycle, the shortest path is not defined Bellman-Ford Algorithm O(mn) time for graphs with negative cost edges

Lemma If a graph has no negative cost cycles, then the shortest paths are simple paths Shortest paths have at most n-1 edges

Shortest paths with a fixed number of edges Find the shortest path from v to w with exactly k edges

Express as a recurrence Optk(w) = minx [Optk-1(x) + cxw] Opt0(w) = 0 if v=w and infinity otherwise

Algorithm, Version 1 foreach w M[0, w] = infinity; M[0, v] = 0; for i = 1 to n-1 M[i, w] = minx(M[i-1,x] + cost[x,w]);

Algorithm, Version 2 foreach w M[0, w] = infinity; M[0, v] = 0; for i = 1 to n-1 M[i, w] = min(M[i-1, w], minx(M[i-1,x] + cost[x,w]))

Algorithm, Version 3 foreach w M[w] = infinity; M[v] = 0; for i = 1 to n-1 M[w] = min(M[w], minx(M[x] + cost[x,w]))

Correctness Proof for Algorithm 3 Key lemma – at the end of iteration i, for all w, M[w] <= M[i, w]; Reconstructing the path: Set P[w] = x, whenever M[w] is updated from vertex x

If the pointer graph has a cycle, then the graph has a negative cost cycle If P[w] = x then M[w] >= M[x] + cost(x,w) Equal after update, then M[x] could be reduced Let v1, v2,…vk be a cycle in the pointer graph with (vk,v1) the last edge added Just before the update M[vj] >= M[vj+1] + cost(vj+1, vj) for j < k M[vk] > M[v1] + cost(v1, vk) Adding everything up 0 > cost(v1,v2) + cost(v2,v3) + … + cost(vk, v1) v1 v4 v2 v3

Negative Cycles If the pointer graph has a cycle, then the graph has a negative cycle Therefore: if the graph has no negative cycles, then the pointer graph has no negative cycles

Finding negative cost cycles What if you want to find negative cost cycles?