Solving Recurrences Continued The Master Theorem

Slides:



Advertisements
Similar presentations
BY Lecturer: Aisha Dawood. Heapsort  O(n log n) worst case like merge sort.  Sorts in place like insertion sort.  Combines the best of both algorithms.
Advertisements

Analysis of Algorithms
David Luebke 1 5/20/2015 CS 332: Algorithms Quicksort.
September 19, Algorithms and Data Structures Lecture IV Simonas Šaltenis Nykredit Center for Database Research Aalborg University
Analysis of Algorithms CS 477/677 Instructor: Monica Nicolescu.
CS 253: Algorithms Chapter 6 Heapsort Appendix B.5 Credit: Dr. George Bebis.
Analysis of Algorithms CS 477/677
Comp 122, Spring 2004 Heapsort. heapsort - 2 Lin / Devi Comp 122 Heapsort  Combines the better attributes of merge sort and insertion sort. »Like merge.
Data Structures, Spring 2004 © L. Joskowicz 1 Data Structures – LECTURE 7 Heapsort and priority queues Motivation Heaps Building and maintaining heaps.
Data Structures Dynamic Sets Heaps Binary Trees & Sorting Hashing.
Heapsort Chapter 6. Heaps A data structure with  Nearly complete binary tree  Heap property: A[parent(i)] ≥ A[i] eg. Parent(i) { return } Left(i) {
UMass Lowell Computer Science Analysis of Algorithms Prof. Karen Daniels Spring, 2007 Heap Lecture Chapter 6 Use NOTES feature to see explanation.
David Luebke 1 7/2/2015 Merge Sort Solving Recurrences The Master Theorem.
Recurrences The expression: is a recurrence. –Recurrence: an equation that describes a function in terms of its value on smaller functions Analysis of.
UMass Lowell Computer Science Analysis of Algorithms Prof. Karen Daniels Fall, 2001 Heap Lecture 2 Chapter 7 Wed. 10/10/01 Use NOTES feature to.
Recurrences The expression: is a recurrence. –Recurrence: an equation that describes a function in terms of its value on smaller functions BIL741: Advanced.
David Luebke 1 9/8/2015 CS 332: Algorithms Review for Exam 1.
Sorting Algorithms (Part II) Slightly modified definition of the sorting problem: input: A collection of n data items where data item a i has a key, k.
2IL50 Data Structures Spring 2015 Lecture 3: Heaps.
David Luebke 1 10/3/2015 CS 332: Algorithms Solving Recurrences Continued The Master Theorem Introduction to heapsort.
Heaps, Heapsort, Priority Queues. Sorting So Far Heap: Data structure and associated algorithms, Not garbage collection context.
Binary Heap.
2IL50 Data Structures Fall 2015 Lecture 3: Heaps.
CS 2133:Data Structures Merge Sort Solving Recurrences The Master Theorem.
September 29, Algorithms and Data Structures Lecture V Simonas Šaltenis Aalborg University
Computer Sciences Department1. Sorting algorithm 3 Chapter 6 3Computer Sciences Department Sorting algorithm 1  insertion sort Sorting algorithm 2.
David Luebke 1 6/3/2016 CS 332: Algorithms Heapsort Priority Queues Quicksort.
CS 2133: Data Structures Quicksort. Review: Heaps l A heap is a “complete” binary tree, usually represented as an array:
1 Analysis of Algorithms Chapter - 03 Sorting Algorithms.
1 Algorithms CSCI 235, Fall 2015 Lecture 14 Analysis of Heap Sort.
David Luebke 1 12/23/2015 Heaps & Priority Queues.
Computer Algorithms Lecture 9 Heapsort Ch. 6, App. B.5 Some of these slides are courtesy of D. Plaisted et al, UNC and M. Nicolescu, UNR.
CPSC 311 Section 502 Analysis of Algorithm Fall 2002 Department of Computer Science Texas A&M University.
CSC 413/513: Intro to Algorithms Solving Recurrences Continued The Master Theorem Introduction to heapsort.
Chapter 6: Heapsort Combines the good qualities of insertion sort (sort in place) and merge sort (speed) Based on a data structure called a “binary heap”
CSE 5392 Fall 2005 Week 4 Devendra Patel Subhesh Pradhan.
1 Heap Sort. A Heap is a Binary Tree Height of tree = longest path from root to leaf =  (lgn) A heap is a binary tree satisfying the heap condition:
David Luebke 1 2/5/2016 CS 332: Algorithms Introduction to heapsort.
Heapsort A minimalist's approach Jeff Chastine. Heapsort Like M ERGE S ORT, it runs in O(n lg n) Unlike M ERGE S ORT, it sorts in place Based off of a.
David Luebke 1 2/19/2016 Priority Queues Quicksort.
Sept Heapsort What is a heap? Max-heap? Min-heap? Maintenance of Max-heaps -MaxHeapify -BuildMaxHeap Heapsort -Heapsort -Analysis Priority queues.
6.Heapsort. Computer Theory Lab. Chapter 6P.2 Why sorting 1. Sometimes the need to sort information is inherent in a application. 2. Algorithms often.
1 Algorithms CSCI 235, Fall 2015 Lecture 13 Heap Sort.
Lecture 2 Sorting.
Priority Queues A priority queue is an ADT where:
"Teachers open the door, but you must enter by yourself. "
Heaps, Heapsort, and Priority Queues
Heaps, Heap Sort and Priority Queues
Heapsort Chapter 6 Lee, Hsiu-Hui
Heapsort.
Heap Sort Example Qamar Abbas.
Introduction to Algorithms
T(n) = aT(n/b) + cn = a(aT(n/b/b) + cn/b) + cn 2
CS 583 Analysis of Algorithms
Heaps, Heapsort, and Priority Queues
CS200: Algorithm Analysis
Ch 6: Heapsort Ming-Te Chi
Counting (Pigeon Hole)
Heapsort.
Lecture 3 / 4 Algorithm Analysis
"Teachers open the door, but you must enter by yourself. "
Design and Analysis of Algorithms
Merge Sort Solving Recurrences The Master Theorem
Topic 5: Heap data structure heap sort Priority queue
HEAPS.
Heapsort Sorting in place
Algorithms CSCI 235, Spring 2019 Lecture 6 Recurrences
Computer Algorithms CISC4080 CIS, Fordham Univ.
Algorithms CSCI 235, Spring 2019 Lecture 14 Heap Sort Read: Ch. 6
Asst. Prof. Dr. İlker Kocabaş
Presentation transcript:

Solving Recurrences Continued The Master Theorem Introduction to heapsort

Review: Merge Sort MergeSort(A, left, right) { if (left < right) { mid = floor((left + right) / 2); MergeSort(A, left, mid); MergeSort(A, mid+1, right); Merge(A, left, mid, right); } // Merge() takes two sorted subarrays of A and // merges them into a single sorted subarray of A. // Code for this is in the book. It requires O(n) // time, and *does* require allocating O(n) space

Review: Analysis of Merge Sort Statement Effort So T(n) = (1) when n = 1, and 2T(n/2) + (n) when n > 1 This expression is a recurrence MergeSort(A, left, right) { T(n) if (left < right) { (1) mid = floor((left + right) / 2); (1) MergeSort(A, left, mid); T(n/2) MergeSort(A, mid+1, right); T(n/2) Merge(A, left, mid, right); (n) }

Review: Solving Recurrences Substitution method Iteration method Master method

Review: Solving Recurrences The substitution method A.k.a. the “making a good guess method” Guess the form of the answer, then use induction to find the constants and show that solution works Run an example: merge sort T(n) = 2T(n/2) + cn We guess that the answer is O(n lg n) Prove it by induction Can similarly show T(n) = Ω(n lg n), thus Θ(n lg n)

Review: Solving Recurrences The “iteration method” Expand the recurrence Work some algebra to express as a summation Evaluate the summation We showed several examples, were in the middle of:

T(n) = aT(n/b) + cn a(aT(n/b/b) + cn/b) + cn a2T(n/b2) + cna/b + cn a2T(n/b2) + cn(a/b + 1) a2(aT(n/b2/b) + cn/b2) + cn(a/b + 1) a3T(n/b3) + cn(a2/b2) + cn(a/b + 1) a3T(n/b3) + cn(a2/b2 + a/b + 1) … akT(n/bk) + cn(ak-1/bk-1 + ak-2/bk-2 + … + a2/b2 + a/b + 1)

So we have T(n) = akT(n/bk) + cn(ak-1/bk-1 + ... + a2/b2 + a/b + 1) For k = logb n n = bk T(n) = akT(1) + cn(ak-1/bk-1 + ... + a2/b2 + a/b + 1) = akc + cn(ak-1/bk-1 + ... + a2/b2 + a/b + 1) = cak + cn(ak-1/bk-1 + ... + a2/b2 + a/b + 1) = cnak /bk + cn(ak-1/bk-1 + ... + a2/b2 + a/b + 1) = cn(ak/bk + ... + a2/b2 + a/b + 1)

So with k = logb n What if a = b? T(n) = cn(ak/bk + ... + a2/b2 + a/b + 1) What if a = b? T(n) = cn(k + 1) = cn(logb n + 1) = (n log n)

So with k = logb n What if a < b? T(n) = cn(ak/bk + ... + a2/b2 + a/b + 1) What if a < b?

So with k = logb n What if a < b? T(n) = cn(ak/bk + ... + a2/b2 + a/b + 1) What if a < b? Recall that (xk + xk-1 + … + x + 1) = (xk+1 -1)/(x-1)

So with k = logb n What if a < b? T(n) = cn(ak/bk + ... + a2/b2 + a/b + 1) What if a < b? Recall that (xk + xk-1 + … + x + 1) = (xk+1 -1)/(x-1) So:

So with k = logb n What if a < b? T(n) = cn(ak/bk + ... + a2/b2 + a/b + 1) What if a < b? Recall that (xk + xk-1 + … + x + 1) = (xk+1 -1)/(x-1) So: T(n) = cn ·(1) = (n)

So with k = logb n What if a > b? T(n) = cn(ak/bk + ... + a2/b2 + a/b + 1) What if a > b?

So with k = logb n What if a > b? T(n) = cn(ak/bk + ... + a2/b2 + a/b + 1) What if a > b?

So with k = logb n What if a > b? T(n) = cn(ak/bk + ... + a2/b2 + a/b + 1) What if a > b? T(n) = cn · (ak / bk)

So with k = logb n What if a > b? T(n) = cn(ak/bk + ... + a2/b2 + a/b + 1) What if a > b? T(n) = cn · (ak / bk) = cn · (alog n / blog n) = cn · (alog n / n)

So with k = logb n What if a > b? T(n) = cn(ak/bk + ... + a2/b2 + a/b + 1) What if a > b? T(n) = cn · (ak / bk) = cn · (alog n / blog n) = cn · (alog n / n) recall logarithm fact: alog n = nlog a

So with k = logb n What if a > b? T(n) = cn(ak/bk + ... + a2/b2 + a/b + 1) What if a > b? T(n) = cn · (ak / bk) = cn · (alog n / blog n) = cn · (alog n / n) recall logarithm fact: alog n = nlog a = cn · (nlog a / n) = (cn · nlog a / n)

So with k = logb n What if a > b? T(n) = cn(ak/bk + ... + a2/b2 + a/b + 1) What if a > b? T(n) = cn · (ak / bk) = cn · (alog n / blog n) = cn · (alog n / n) recall logarithm fact: alog n = nlog a = cn · (nlog a / n) = (cn · nlog a / n) = (nlog a )

So…

The Master Theorem Given: a divide and conquer algorithm An algorithm that divides the problem of size n into a subproblems, each of size n/b Let the cost of each stage (i.e., the work to divide the problem + combine solved subproblems) be described by the function f(n) Then, the Master Theorem gives us a cookbook for the algorithm’s running time:

The Master Theorem if T(n) = aT(n/b) + f(n) then

Using The Master Method T(n) = 9T(n/3) + n a=9, b=3, f(n) = n nlogb a = nlog3 9 = (n2) Since f(n) = O(nlog3 9 - ), where =1, case 1 applies: Thus the solution is T(n) = (n2)

Sorting Revisited So far we’ve talked about two algorithms to sort an array of numbers What is the advantage of merge sort? What is the advantage of insertion sort? Next on the agenda: Heapsort Combines advantages of both previous algorithms

Heaps A heap can be seen as a complete binary tree: What makes a binary tree complete? Is the example above complete? 16 14 10 8 7 9 3 2 4 1

Heaps A heap can be seen as a complete binary tree: The book calls them “nearly complete” binary trees; can think of unfilled slots as null pointers 16 14 10 8 7 9 3 2 4 1 1 1 1 1 1

Heaps In practice, heaps are usually implemented as arrays: A = 16 14 10 8 7 9 3 2 4 1 A = 16 14 10 8 7 9 3 2 4 1 =

Heaps To represent a complete binary tree as an array: The root node is A[1] Node i is A[i] The parent of node i is A[i/2] (note: integer divide) The left child of node i is A[2i] The right child of node i is A[2i + 1] 16 14 10 8 7 9 3 2 4 1 A = 16 14 10 8 7 9 3 2 4 1 =

Referencing Heap Elements So… Parent(i) { return i/2; } Left(i) { return 2*i; } right(i) { return 2*i + 1; } An aside: How would you implement this most efficiently? Another aside: Really?

The Heap Property Heaps also satisfy the heap property: A[Parent(i)]  A[i] for all nodes i > 1 In other words, the value of a node is at most the value of its parent Where is the largest element in a heap stored? Definitions: The height of a node in the tree = the number of edges on the longest downward path to a leaf The height of a tree = the height of its root

Heap Height What is the height of an n-element heap? Why? This is nice: basic heap operations take at most time proportional to the height of the heap

Heap Operations: Heapify() Heapify(): maintain the heap property Given: a node i in the heap with children l and r Given: two subtrees rooted at l and r, assumed to be heaps Problem: The subtree rooted at i may violate the heap property (How?) Action: let the value of the parent node “float down” so subtree at i satisfies the heap property What do you suppose will be the basic operation between i, l, and r?

Heap Operations: Heapify() Heapify(A, i) { l = Left(i); r = Right(i); if (l <= heap_size(A) && A[l] > A[i]) largest = l; else largest = i; if (r <= heap_size(A) && A[r] > A[largest]) largest = r; if (largest != i) Swap(A, i, largest); Heapify(A, largest); }

Heapify() Example 16 4 10 14 7 9 3 2 8 1 A = 16 4 10 14 7 9 3 2 8 1

Heapify() Example 16 4 10 14 7 9 3 2 8 1 A = 16 4 10 14 7 9 3 2 8 1

Heapify() Example 16 4 10 14 7 9 3 2 8 1 A = 16 4 10 14 7 9 3 2 8 1

Heapify() Example 16 14 10 4 7 9 3 2 8 1 A = 16 14 10 4 7 9 3 2 8 1

Heapify() Example 16 14 10 4 7 9 3 2 8 1 A = 16 14 10 4 7 9 3 2 8 1

Heapify() Example 16 14 10 4 7 9 3 2 8 1 A = 16 14 10 4 7 9 3 2 8 1

Heapify() Example 16 14 10 8 7 9 3 2 4 1 A = 16 14 10 8 7 9 3 2 4 1

Heapify() Example 16 14 10 8 7 9 3 2 4 1 A = 16 14 10 8 7 9 3 2 4 1

Heapify() Example 16 14 10 8 7 9 3 2 4 1 A = 16 14 10 8 7 9 3 2 4 1

Analyzing Heapify(): Informal Aside from the recursive call, what is the running time of Heapify()? How many times can Heapify() recursively call itself? What is the worst-case running time of Heapify() on a heap of size n?

Analyzing Heapify(): Formal Fixing up relationships between i, l, and r takes (1) time If the heap at i has n elements, how many elements can the subtrees at l or r have? Draw it Answer: 2n/3 (worst case: bottom row 1/2 full) So time taken by Heapify() is given by T(n)  T(2n/3) + (1)

Analyzing Heapify(): Formal So we have T(n)  T(2n/3) + (1) By case 2 of the Master Theorem, T(n) = O(lg n) Thus, Heapify() takes linear time

Heap Operations: BuildHeap() We can build a heap in a bottom-up manner by running Heapify() on successive subarrays Fact: for array of length n, all elements in range A[n/2 + 1 .. n] are heaps (Why?) So: Walk backwards through the array from n/2 to 1, calling Heapify() on each node. Order of processing guarantees that the children of node i are heaps when i is processed

BuildHeap() // given an unsorted array A, make A a heap BuildHeap(A) { heap_size(A) = length(A); for (i = length[A]/2 downto 1) Heapify(A, i); }

BuildHeap() Example Work through example A = {4, 1, 3, 2, 16, 9, 10, 14, 8, 7} 4 1 3 2 16 9 10 14 8 7

Analyzing BuildHeap() Each call to Heapify() takes O(lg n) time There are O(n) such calls (specifically, n/2) Thus the running time is O(n lg n) Is this a correct asymptotic upper bound? Is this an asymptotically tight bound? A tighter bound is O(n) How can this be? Is there a flaw in the above reasoning?

Analyzing BuildHeap(): Tight To Heapify() a subtree takes O(h) time where h is the height of the subtree h = O(lg m), m = # nodes in subtree The height of most subtrees is small Fact: an n-element heap has at most n/2h+1 nodes of height h CLR 7.3 uses this fact to prove that BuildHeap() takes O(n) time

Heapsort Given BuildHeap(), an in-place sorting algorithm is easily constructed: Maximum element is at A[1] Discard by swapping with element at A[n] Decrement heap_size[A] A[n] now contains correct value Restore heap property at A[1] by calling Heapify() Repeat, always swapping A[1] for A[heap_size(A)]

Heapsort Heapsort(A) { BuildHeap(A); for (i = length(A) downto 2) Swap(A[1], A[i]); heap_size(A) -= 1; Heapify(A, 1); }

Analyzing Heapsort The call to BuildHeap() takes O(n) time Each of the n - 1 calls to Heapify() takes O(lg n) time Thus the total time taken by HeapSort() = O(n) + (n - 1) O(lg n) = O(n) + O(n lg n) = O(n lg n)

Priority Queues Heapsort is a nice algorithm, but in practice Quicksort (coming up) usually wins But the heap data structure is incredibly useful for implementing priority queues A data structure for maintaining a set S of elements, each with an associated value or key Supports the operations Insert(), Maximum(), and ExtractMax() What might a priority queue be useful for?

Priority Queue Operations Insert(S, x) inserts the element x into set S Maximum(S) returns the element of S with the maximum key ExtractMax(S) removes and returns the element of S with the maximum key How could we implement these operations using a heap?