Lecture No 6 Advance Analysis of Institute of Southern Punjab Multan

Slides:



Advertisements
Similar presentations
Analysis of Algorithms CS 477/677 Instructor: Monica Nicolescu Lecture 6.
Advertisements

Comp 122, Spring 2004 Divide and Conquer (Merge Sort)
1 Divide & Conquer Algorithms. 2 Recursion Review A function that calls itself either directly or indirectly through another function Recursive solutions.
September 12, Algorithms and Data Structures Lecture III Simonas Šaltenis Nykredit Center for Database Research Aalborg University
Analysis of Algorithms CS 477/677 Instructor: Monica Nicolescu Lecture 5.
Analysis of Algorithms CS 477/677 Sorting – Part B Instructor: George Bebis (Chapter 7)
Spring 2015 Lecture 5: QuickSort & Selection
Analysis of Algorithms CS 477/677 Randomizing Quicksort Instructor: George Bebis (Appendix C.2, Appendix C.3) (Chapter 5, Chapter 7)
Quicksort Ack: Several slides from Prof. Jim Anderson’s COMP 750 notes. UNC Chapel Hill1.
Quicksort Many of the slides are from Prof. Plaisted’s resources at University of North Carolina at Chapel Hill.
Lecture 8 Jianjun Hu Department of Computer Science and Engineering University of South Carolina CSCE350 Algorithms and Data Structure.
CS Section 600 CS Section 002 Dr. Angela Guercio Spring 2010.
CS 253: Algorithms Chapter 7 Mergesort Quicksort Credit: Dr. George Bebis.
Ch. 7 - QuickSort Quick but not Guaranteed. Ch.7 - QuickSort Another Divide-and-Conquer sorting algorithm… As it turns out, MERGESORT and HEAPSORT, although.
Analysis of Algorithms CS 477/677 Instructor: Monica Nicolescu.
Analysis of Algorithms CS 477/677 Instructor: Monica Nicolescu.
Lecture 2: Divide and Conquer I: Merge-Sort and Master Theorem Shang-Hua Teng.
Analysis of Algorithms CS 477/677 Midterm Exam Review Instructor: George Bebis.
Quicksort CIS 606 Spring Quicksort Worst-case running time: Θ(n 2 ). Expected running time: Θ(n lg n). Constants hidden in Θ(n lg n) are small.
Analysis of Algorithms CS 477/677
1 QuickSort Worst time:  (n 2 ) Expected time:  (nlgn) – Constants in the expected time are small Sorts in place.
Computer Algorithms Lecture 10 Quicksort Ch. 7 Some of these slides are courtesy of D. Plaisted et al, UNC and M. Nicolescu, UNR.
10 Algorithms in 20th Century Science, Vol. 287, No. 5454, p. 799, February 2000 Computing in Science & Engineering, January/February : The Metropolis.
10/14/ Algorithms1 Algorithms - Ch2 - Sorting.
Algorithms IS 320 Spring 2015 Sorting. 2 The Sorting Problem Input: –A sequence of n numbers a 1, a 2,..., a n Output: –A permutation (reordering) a 1.
QuickSort (Ch. 7) Like Merge-Sort, based on the three-step process of divide- and-conquer. Input: An array A[1…n] of comparable elements, the starting.
COSC 3101A - Design and Analysis of Algorithms 2 Asymptotic Notations Continued Proof of Correctness: Loop Invariant Designing Algorithms: Divide and Conquer.
1 Computer Algorithms Lecture 8 Sorting Algorithms Some of these slides are courtesy of D. Plaisted, UNC and M. Nicolescu, UNR.
COSC 3101A - Design and Analysis of Algorithms 4 Quicksort Medians and Order Statistics Many of these slides are taken from Monica Nicolescu, Univ. of.
Lecture # 6 1 Advance Analysis of Algorithms. Divide-and-Conquer Divide the problem into a number of subproblems Similar sub-problems of smaller size.
Sorting. 2 The Sorting Problem Input: A sequence of n numbers a 1, a 2,..., a n Output: A permutation (reordering) a 1 ’, a 2 ’,..., a n ’ of the input.
BITS Pilani Pilani Campus Data Structure and Algorithms Design Dr. Maheswari Karthikeyan Lecture3.
Analysis of Algorithms CS 477/677 Instructor: Monica Nicolescu Lecture 4.
Algorithms Sorting – Part 3.
Advanced Sorting.
CMPT 438 Algorithms.
Analysis of Algorithms CS 477/677
Subject Name: Design and Analysis of Algorithm Subject Code: 10CS43
Quick Sort Divide: Partition the array into two sub-arrays
Algorithm Design & Analysis
Advance Analysis of Algorithms
CSC 413/513: Intro to Algorithms
Divide-and-Conquer The most-well known algorithm design strategy:
Growth Functions Algorithms Lecture 8
Analysis of Algorithms CS 477/677
CO 303 Algorithm Analysis And Design Quicksort
Ch 7: Quicksort Ming-Te Chi
Divide and Conquer (Merge Sort)
Algorithms and Data Structures Lecture III
Lecture 3 / 4 Algorithm Analysis
Divide-and-Conquer The most-well known algorithm design strategy:
CS200: Algorithms Analysis
CS 583 Analysis of Algorithms
Design and Analysis of Algorithms
Divide and Conquer (Merge Sort)
CS 3343: Analysis of Algorithms
CS 332: Algorithms Quicksort David Luebke /9/2019.
Algorithms: Design and Analysis
CSC 380: Design and Analysis of Algorithms
CSC 380: Design and Analysis of Algorithms
Ack: Several slides from Prof. Jim Anderson’s COMP 202 notes.
Divide & Conquer Algorithms
Analysis of Algorithms
CS200: Algorithm Analysis
Algorithms Sorting.
The Selection Problem.
Design and Analysis of Algorithms
Quicksort Quick sort Correctness of partition - loop invariant
Divide and Conquer Merge sort and quick sort Binary search
Presentation transcript:

Lecture No 6 Advance Analysis of Institute of Southern Punjab Multan Department of Computer Science

MergeSort (Example) - 1

MergeSort (Example) - 2

MergeSort (Example) - 3

MergeSort (Example) - 4

MergeSort (Example) - 5

MergeSort (Example) - 6

MergeSort (Example) - 7

MergeSort (Example) - 8

MergeSort (Example) - 9

MergeSort (Example) - 10

MergeSort (Example) - 11

MergeSort (Example) - 12

MergeSort (Example) - 13

MergeSort (Example) - 14

MergeSort (Example) - 15

MergeSort (Example) - 16

MergeSort (Example) - 17

MergeSort (Example) - 18

MergeSort (Example) - 19

MergeSort (Example) - 20

MergeSort (Example) - 21

MergeSort (Example) - 22

14 23 45 98 6 33 42 67

14 23 45 98 6 33 42 67 Merge

14 23 45 98 6 33 42 67 6 Merge

14 23 45 98 6 33 42 67 6 14 Merge

14 23 45 98 6 33 42 67 6 14 23 Merge

14 23 45 98 6 33 42 67 6 14 23 33 Merge

14 23 45 98 6 33 42 67 6 14 23 33 42 Merge

14 23 45 98 6 33 42 67 6 14 23 33 42 45 Merge

14 23 45 98 6 33 42 67 6 14 23 33 42 45 67 Merge

14 23 45 98 6 33 42 67 6 14 23 33 42 45 67 98 Merge

Divide-and-Conquer Divide the problem into a number of sub-problems Similar sub-problems of smaller size Conquer the sub-problems Solve the sub-problems recursively Sub-problem size small enough  solve the problems in straightforward manner Combine the solutions of the sub-problems Obtain the solution for the original problem

Merge Sort Approach To sort an array A[p . . r]: Divide Conquer Divide the n-element sequence to be sorted into two subsequences of n/2 elements each Conquer Sort the subsequences recursively using merge sort When the size of the sequences is 1 there is nothing more to do Combine Merge the two sorted subsequences

Merge Sort Alg.: MERGE-SORT(A, p, r) if p < r Check for base case q r Alg.: MERGE-SORT(A, p, r) if p < r Check for base case then q ← (p + r)/2 Divide MERGE-SORT(A, p, q) Conquer MERGE-SORT(A, q + 1, r) Conquer MERGE(A, p, q, r) Combine Initial call: MERGE-SORT(A, 1, n) 1 2 3 4 5 6 7 8 5 2 4 7 1 3 2 6

Example – n Power of 2 Divide 7 5 5 7 5 4 7 q = 4 1 2 3 4 5 6 7 8 1 2

Example – n Power of 2 Conquer and Merge 7 5 5 7 5 4 7 1 2 3 4 5 6 7 8

Example – n Not a Power of 2 6 2 5 3 7 4 1 8 9 10 11 q = 6 Divide 4 1 6 2 7 3 5 8 9 10 11 q = 9 q = 3 2 7 4 1 3 6 5 8 9 10 11 7 4 1 2 3 6 5 8 9 10 11 4 1 7 2 6 5 3 8

Example – n Not a Power of 2 7 6 5 4 3 2 1 8 9 10 11 Conquer and Merge 7 6 4 2 1 3 5 8 9 10 11 7 4 2 1 3 6 5 8 9 10 11 2 3 4 6 5 9 10 11 1 7 8 7 4 1 2 6 5 3 8

Merging Input: Array A and indices p, q, r such that p ≤ q < r 1 2 3 4 5 6 7 8 p r q Input: Array A and indices p, q, r such that p ≤ q < r Subarrays A[p . . q] and A[q + 1 . . r] are sorted Output: One single sorted subarray A[p . . r]

Merging Idea for merging: Two piles of sorted cards 1 2 3 4 5 6 7 8 p r q Idea for merging: Two piles of sorted cards Choose the smaller of the two top cards Remove it and place it in the output pile Repeat the process until one pile is empty Take the remaining input pile and place it face-down onto the output pile A1 A[p, q] A[p, r] A2 A[q+1, r]

Example: MERGE(A, 9, 12, 16) p r q

Example: MERGE(A, 9, 12, 16)

Example (cont.)

Example (cont.)

Example (cont.) Done!

Merge - Pseudocode Alg.: MERGE(A, p, q, r) Compute n1 and n2 3 4 5 6 7 8 p r q n1 n2 Alg.: MERGE(A, p, q, r) Compute n1 and n2 Copy the first n1 elements into L[1 . . n1 + 1] and the next n2 elements into R[1 . . n2 + 1] L[n1 + 1] ← ; R[n2 + 1] ←  i ← 1; j ← 1 for k ← p to r do if L[ i ] ≤ R[ j ] then A[k] ← L[ i ] i ←i + 1 else A[k] ← R[ j ] j ← j + 1 p q 7 5 4 2 6 3 1 r q + 1 L R 

Running Time of Merge (assume last for loop) Initialization (copying into temporary arrays): (n1 + n2) = (n) Adding the elements to the final array: - n iterations, each taking constant time  (n) Total time for Merge: (n)

Analyzing Divide-and Conquer Algorithms The recurrence is based on the three steps of the paradigm: T(n) – running time on a problem of size n Divide the problem into a subproblems, each of size n/b: takes D(n) Conquer (solve) the subproblems aT(n/b) Combine the solutions C(n) (1) if n ≤ c T(n) = aT(n/b) + D(n) + C(n) otherwise

MERGE-SORT Running Time Divide: compute q as the average of p and r: D(n) = (1) Conquer: recursively solve 2 subproblems, each of size n/2  2T (n/2) Combine: MERGE on an n-element subarray takes (n) time  C(n) = (n) (1) if n =1 T(n) = 2T(n/2) + (n) if n > 1

Solve the Recurrence T(n) = c if n = 1 2T(n/2) + cn if n > 1 Use Master’s Theorem: Compare n with f(n) = cn Case 2: T(n) = Θ(nlgn)

Merge Sort - Discussion Running time insensitive of the input Advantages: Guaranteed to run in (nlgn) Disadvantage Requires extra space N

Quicksort Sort an array A[p…r] Divide A[p…q] A[q+1…r] ≤ Sort an array A[p…r] Divide Partition the array A into 2 subarrays A[p..q] and A[q+1..r], such that each element of A[p..q] is smaller than or equal to each element in A[q+1..r] Need to find index q to partition the array

Quicksort Conquer Combine A[p…q] A[q+1…r] ≤ Conquer Recursively sort A[p..q] and A[q+1..r] using Quicksort Combine Trivial: the arrays are sorted in place No additional work is required to combine them The entire array is now sorted

QUICKSORT Alg.: QUICKSORT(A, p, r) Initially: p=1, r=n if p < r then q  PARTITION(A, p, r) QUICKSORT (A, p, q) QUICKSORT (A, q+1, r) Initially: p=1, r=n Recurrence: T(n) = T(q) + T(n – q) + f(n) PARTITION())

Partitioning the Array Choosing PARTITION() There are different ways to do this Each has its own advantages/disadvantages Hoare partition (see prob. 7-1, page 159) Select a pivot element x around which to partition Grows two regions A[p…i]  x x  A[j…r] A[p…i]  x x  A[j…r] i j

Example 7 3 1 4 6 2 5 i j A[p…r] pivot x=5 7 3 1 4 6 2 5 i j 7 5 1 4 6 A[p…q] A[q+1…r] 7 5 6 4 1 2 3 i j

Example

Partitioning the Array Alg. PARTITION (A, p, r) x  A[p] i  p – 1 j  r + 1 while TRUE do repeat j  j – 1 until A[j] ≤ x do repeat i  i + 1 until A[i] ≥ x if i < j then exchange A[i]  A[j] else return j p r A: 7 3 1 4 6 2 5 i j ar ap i j=q A: A[p…q] A[q+1…r] ≤ Each element is visited once! Running time: (n) n = r – p + 1

Recurrence Alg.: QUICKSORT(A, p, r) Initially: p=1, r=n if p < r then q  PARTITION(A, p, r) QUICKSORT (A, p, q) QUICKSORT (A, q+1, r) Initially: p=1, r=n Recurrence: T(n) = T(q) + T(n – q) + n

Worst Case Partitioning One region has one element and the other has n – 1 elements Maximally unbalanced Recurrence: q=1 T(n) = T(1) + T(n – 1) + n, T(1) = (1) T(n) = T(n – 1) + n = n n - 1 n - 2 n - 3 2 1 3 (n2) When does the worst case happen?

Best Case Partitioning Partitioning produces two regions of size n/2 Recurrence: q=n/2 T(n) = 2T(n/2) + (n) T(n) = (nlgn) (Master theorem)

Case Between Worst and Best 9-to-1 proportional split Q(n) = Q(9n/10) + Q(n/10) + n

How does partition affect performance?

How does partition affect performance?

Performance of Quicksort Average case All permutations of the input numbers are equally likely On a random input array, we will have a mix of well balanced and unbalanced splits Good and bad splits are randomly distributed across throughout the tree partitioning cost: n = (n) n n - 1 1 combined partitioning cost: 2n-1 = (n) n (n – 1)/2 (n – 1)/2 + 1 (n – 1)/2 Alternate of a good and a bad split Nearly well balanced split Running time of Quicksort when levels alternate between good and bad splits is O(nlgn)

Selection Sort (Ex. 2.2-2, page 27) Idea: Find the smallest element in the array Exchange it with the element in the first position Find the second smallest element and exchange it with the element in the second position Continue until the array is sorted Disadvantage: Running time depends only slightly on the amount of order in the file

Example 1 3 2 9 6 4 8 8 6 9 4 3 2 1 8 3 2 9 6 4 1 8 9 6 4 3 2 1 8 3 4 9 6 2 1 9 8 6 4 3 2 1 8 6 4 9 3 2 1 9 8 6 4 3 2 1

Selection Sort Alg.: SELECTION-SORT(A) n ← length[A] for j ← 1 to n - 1 do smallest ← j for i ← j + 1 to n do if A[i] < A[smallest] then smallest ← i exchange A[j] ↔ A[smallest] 1 3 2 9 6 4 8

Analysis of Selection Sort Alg.: SELECTION-SORT(A) n ← length[A] for j ← 1 to n - 1 do smallest ← j for i ← j + 1 to n do if A[i] < A[smallest] then smallest ← i exchange A[j] ↔ A[smallest] cost times c1 1 c2 n c3 n-1 c4 c5 c6 c7 n-1 n2/2 comparisons n exchanges

Sorting Challenge 1 Problem: Sort a file of huge records with tiny keys Example application: Reorganize your MP-3 files Which method to use? merge sort, guaranteed to run in time NlgN selection sort bubble sort a custom algorithm for huge records/tiny keys insertion sort

Sorting Files with Huge Records and Small Keys Insertion sort or bubble sort? NO, too many exchanges Selection sort? YES, it takes linear time for exchanges Merge sort or custom method? Probably not: selection sort simpler, does less swaps

Sorting Challenge 2 Problem: Sort a huge randomly-ordered file of small records Application: Process transaction record for a phone company Which sorting method to use? Bubble sort Selection sort Mergesort guaranteed to run in time NlgN Insertion sort

Sorting Huge, Randomly - Ordered Files Selection sort? NO, always takes quadratic time Bubble sort? NO, quadratic time for randomly-ordered keys Insertion sort? Mergesort? YES, it is designed for this problem

Sorting Challenge 3 Problem: sort a file that is already almost in order Applications: Re-sort a huge database after a few changes Doublecheck that someone else sorted a file Which sorting method to use? Mergesort, guaranteed to run in time NlgN Selection sort Bubble sort A custom algorithm for almost in-order files Insertion sort

Sorting Files That are Almost in Order Selection sort? NO, always takes quadratic time Bubble sort? NO, bad for some definitions of “almost in order” Ex: B C D E F G H I J K L M N O P Q R S T U V W X Y Z A Insertion sort? YES, takes linear time for most definitions of “almost in order” Mergesort or custom method? Probably not: insertion sort simpler and faster

Sorting Insertion sort Bubble Sort Design approach: Sorts in place: Best case: Worst case: Bubble Sort Running time: incremental Yes (n) (n2) incremental Yes (n2)

Sorting Selection sort Merge Sort Design approach: Sorts in place: Running time: Merge Sort incremental Yes (n2) divide and conquer No Let’s see!!