Sorting Fun1 Chapter 4: Sorting 7 4 9 6 2  2 4 6 7 9 4 2  2 47 9  7 9 2  29  9.

Slides:



Advertisements
Similar presentations
Bucket-Sort and Radix-Sort B 1, c7, d7, g3, b3, a7, e 
Advertisements

Goodrich, Tamassia Sorting, Sets and Selection1 Merge Sort 7 2  9 4   2  2 79  4   72  29  94  4.
Data Structures Lecture 9 Fang Yu Department of Management Information Systems National Chengchi University Fall 2010.
Merge Sort 7 2  9 4   2  2 79  4   72  29  94  4.
Merge Sort 4/15/2017 4:30 PM Merge Sort 7 2   7  2  2 7
Quicksort Quicksort     29  9.
Insertion Sort. Selection Sort. Bubble Sort. Heap Sort. Merge-sort. Quick-sort. 2 CPSC 3200 University of Tennessee at Chattanooga – Summer 2013 © 2010.
© 2004 Goodrich, Tamassia Quick-Sort     29  9.
© 2004 Goodrich, Tamassia QuickSort1 Quick-Sort     29  9.
1 Sorting Problem: Given a sequence of elements, find a permutation such that the resulting sequence is sorted in some order. We have already seen: –Insertion.
Quick-Sort     29  9.
© 2004 Goodrich, Tamassia Quick-Sort     29  9.
© 2004 Goodrich, Tamassia Sorting Lower Bound1. © 2004 Goodrich, Tamassia Sorting Lower Bound2 Comparison-Based Sorting (§ 10.3) Many sorting algorithms.
© 2004 Goodrich, Tamassia Sorting Lower Bound1. © 2004 Goodrich, Tamassia Sorting Lower Bound2 Comparison-Based Sorting (§ 10.3) Many sorting algorithms.
© 2004 Goodrich, Tamassia Selection1. © 2004 Goodrich, Tamassia Selection2 The Selection Problem Given an integer k and n elements x 1, x 2, …, x n, taken.
TTIT33 Algorithms and Optimization – Dalg Lecture 2 HT TTIT33 Algorithms and optimization Lecture 2 Algorithms Sorting [GT] 3.1.2, 11 [LD] ,
© 2004 Goodrich, Tamassia Merge Sort1 Quick-Sort     29  9.
Divide-and-Conquer1 7 2  9 4   2  2 79  4   72  29  94  4 Modified by: Daniel Gomez-Prado, University of Massachusetts Amherst.
1 CSC401 – Analysis of Algorithms Lecture Notes 9 Radix Sort and Selection Objectives  Introduce no-comparison-based sorting algorithms: Bucket-sort and.
Selection1. 2 The Selection Problem Given an integer k and n elements x 1, x 2, …, x n, taken from a total order, find the k-th smallest element in this.
© 2004 Goodrich, Tamassia Merge Sort1 Chapter 8 Sorting Topics Basics Merge sort ( 合併排序 ) Quick sort Bucket sort ( 分籃排序 ) Radix sort ( 基數排序 ) Summary.
Sorting Lower Bound1. 2 Comparison-Based Sorting (§ 4.4) Many sorting algorithms are comparison based. They sort by making comparisons between pairs of.
Sorting.
CSE 373 Data Structures Lecture 15
Ch. 8 & 9 – Linear Sorting and Order Statistics What do you trade for speed?
1 More Sorting radix sort bucket sort in-place sorting how fast can we sort?
CSC 213 Lecture 12: Quick Sort & Radix/Bucket Sort.
The Selection Problem. 2 Median and Order Statistics In this section, we will study algorithms for finding the i th smallest element in a set of n elements.
Analysis of Algorithms CS 477/677
© 2004 Goodrich, Tamassia Bucket-Sort and Radix-Sort B 1, c7, d7, g3, b3, a7, e 
Randomized Algorithms CSc 4520/6520 Design & Analysis of Algorithms Fall 2013 Slides adopted from Dmitri Kaznachey, George Mason University and Maciej.
© 2004 Goodrich, Tamassia Quick-Sort     29  9.
1 Merge Sort 7 2  9 4   2  2 79  4   72  29  94  4.
Towers of Hanoi Move n (4) disks from pole A to pole B such that a larger disk is never put on a smaller disk A BC ABC.
QuickSort by Dr. Bun Yue Professor of Computer Science CSCI 3333 Data.
1 COMP9024: Data Structures and Algorithms Week Eight: Sortings and Sets Hui Wu Session 1, 2016
Advanced Sorting 7 2  9 4   2   4   7
Chapter 11 Sorting Acknowledgement: These slides are adapted from slides provided with Data Structures and Algorithms in C++, Goodrich, Tamassia and Mount.
Merge Sort 7 2  9 4   2   4   7 2  2 9  9 4  4.
Merge Sort 1/12/2018 9:39 AM Presentation for use with the textbook Data Structures and Algorithms in Java, 6th edition, by M. T. Goodrich, R. Tamassia,
Quick-Sort 2/18/2018 3:56 AM Selection Selection.
Bucket-Sort and Radix-Sort
COMP9024: Data Structures and Algorithms
COMP9024: Data Structures and Algorithms
Quick-Sort 9/12/2018 3:26 PM Presentation for use with the textbook Data Structures and Algorithms in Java, 6th edition, by M. T. Goodrich, R. Tamassia,
Quick-Sort 9/13/2018 1:15 AM Quick-Sort     2
Selection Selection 1 Quick-Sort Quick-Sort 10/30/16 13:52
Radish-Sort 11/11/ :01 AM Quick-Sort     2 9  9
Objectives Introduce different known sorting algorithms
Bucket-Sort and Radix-Sort
Bucket-Sort and Radix-Sort
Quick-Sort 11/14/2018 2:17 PM Chapter 4: Sorting    7 9
Bucket-Sort and Radix-Sort
Quick-Sort 11/19/ :46 AM Chapter 4: Sorting    7 9
Bucket-Sort and Radix-Sort
Merge Sort 12/4/ :32 AM Merge Sort 7 2   7  2   4  4 9
(2,4) Trees 12/4/2018 1:20 PM Sorting Lower Bound Sorting Lower Bound.
Sorting Recap & More about Sorting
Sorting, Sets and Selection
Quick-Sort 2/23/2019 1:48 AM Chapter 4: Sorting    7 9
Quick-Sort 2/25/2019 2:22 AM Quick-Sort     2
Copyright © Aiman Hanna All rights reserved
Quick-Sort 4/8/ :20 AM Quick-Sort     2 9  9
Bucket-Sort and Radix-Sort
Quick-Sort 4/25/2019 8:10 AM Quick-Sort     2
Quick-Sort 5/7/2019 6:43 PM Selection Selection.
Quick-Sort 5/25/2019 6:16 PM Selection Selection.
CS203 Lecture 15.
Bucket-Sort and Radix-Sort
Divide-and-Conquer 7 2  9 4   2   4   7
Presentation transcript:

Sorting Fun1 Chapter 4: Sorting     29  9

Sorting Fun2 What We’ll Do Quick Sort Lower bound on runtimes for comparison based sort Radix and Bucket sort

Sorting Fun3 Quick-Sort     29  9

Sorting Fun4 Quick-Sort Quick-sort is a randomized sorting algorithm based on the divide-and-conquer paradigm: Divide: pick a random element x (called pivot) and partition S into  L elements less than x  E elements equal x  G elements greater than x Recur: sort L and G Conquer: join L, E and G x x L G E x

Sorting Fun5 Partition We partition an input sequence as follows: We remove, in turn, each element y from S and We insert y into L, E or G, depending on the result of the comparison with the pivot x Each insertion and removal is at the beginning or at the end of a sequence, and hence takes O(1) time Thus, the partition step of quick-sort takes O(n) time Algorithm partition(S, p) Input sequence S, position p of pivot Output subsequences L, E, G of the elements of S less than, equal to, or greater than the pivot, resp. L, E, G  empty sequences x  S.remove(p) while  S.isEmpty() y  S.remove(S.first()) if y < x L.insertLast(y) else if y = x E.insertLast(y) else { y > x } G.insertLast(y) return L, E, G

Sorting Fun6 Quick-Sort Tree An execution of quick-sort is depicted by a binary tree Each node represents a recursive call of quick-sort and stores  Unsorted sequence before the execution and its pivot  Sorted sequence at the end of the execution The root is the initial call The leaves are calls on subsequences of size 0 or     29  9

Sorting Fun7 Execution Example Pivot selection      38    94  4

Sorting Fun8 Execution Example (cont.) Partition, recursive call, pivot selection    94     38  8 2  2

Sorting Fun9 Execution Example (cont.) Partition, recursive call, base case   11    94      38  8

Sorting Fun10 Execution Example (cont.) Recursive call, …, base case, join   38     11  14 3   94  44  4

Sorting Fun11 Execution Example (cont.) Recursive call, pivot selection      11  14 3   94  44  4

Sorting Fun12 Execution Example (cont.) Partition, …, recursive call, base case      11  14 3   94  44  4 9  99  9

Sorting Fun13 Execution Example (cont.) Join, join      11  14 3   94  44  4 9  99  9

Sorting Fun14 Worst-case Running Time The worst case for quick-sort occurs when the pivot is the unique minimum or maximum element One of L and G has size n  1 and the other has size 0 The running time is proportional to the sum n  (n  1)  …  2  Thus, the worst-case running time of quick-sort is O(n 2 ) depthtime 0n 1 n  1 …… 1 … number of comparisons in partition step

Sorting Fun15 Expected Running Time Consider a recursive call of quick-sort on a sequence of size s Good call: the sizes of L and G are each less than 3s  4 Bad call: one of L and G has size greater than 3s  4 A call is good with probability 1  2 1/2 of the possible pivots cause good calls:  Good callBad call Good pivotsBad pivots

Sorting Fun16 Expected Running Time, Part 2 Probabilistic Fact: The expected number of coin tosses required in order to get k heads is 2k For a node of depth i, we expect i  2 ancestors are good calls The size of the input sequence for the current call is at most ( 3  4 ) i  2 n  Since each good call shrinks size to ¾ or less than previous size Therefore, we have For a node of depth 2log 4  3 n, the expected input size is one The expected height of the quick-sort tree is O(log n) The amount or work done at the nodes of the same depth is O(n) Thus, the expected running time of quick-sort is O(n log n)

Sorting Fun17 Sorting Lower Bound

Sorting Fun18 Comparison-Based Sorting (§ 4.4) Many sorting algorithms are comparison based. They sort by making comparisons between pairs of objects Examples: bubble-sort, selection-sort, insertion-sort, heap-sort, merge-sort, quick-sort,... Let us therefore derive a lower bound on the running time of any algorithm that uses comparisons to sort a set S of n elements, x 1, x 2, …, x n. Is x i < x j ? yes no Assume that the x i are distinct, which is not a restriction

Sorting Fun19 Counting Comparisons Let us just count comparisons then. First, we can map any comparison based sorting algorithm to a decision tree as follows: Let the root node of the tree correspond to the first comparison, (is x i < x j ?), that occurs in the algorithm. The outcome of the comparison is either yes or no. If yes we proceed to another comparison, say x a < x b ? We let this comparison correspond to the left child of the root. If no we proceed to the comparison x c < x d ? We let this comparison correspond to the right child of the root. Each of those comparisons can be either yes or no…

Sorting Fun20 The Decision Tree Each possible permutation of the set S will cause the sorting algorithm to execute a sequence of comparison, effectively traversing a path in the tree from the root to some external node

Sorting Fun21 Paths Represent Permutations Fact: Each external node v in the tree can represent the sequence of comparisons for exactly one permutation of S If P 1 and P 2 different permutations, then there is at least one pair x i, x j with x i before x j in P 1 and x i after x j in P 2 For both P 1 and P 2 to end up at v, this means every decision made along the way resulted in the exact same outcome. This cannot occur if the sorting algorithm behaves correctly, because in one permutation x i started before x j and in the other their order was reversed (remember, they cannot be equal)

Sorting Fun22 Decision Tree Height The height of this decision tree is a lower bound on the running time Every possible input permutation must lead to a separate leaf output (by previous slide). There are n! permutations, so there are n! leaves. Since there are n!=1*2*…*n leaves, the height is at least log (n!)

Sorting Fun23 The Lower Bound Any comparison-based sorting algorithms takes at least log (n!) time Therefore, any such algorithm takes time at least Since there are at least n/2 terms larger than n/2 in n! That is, any comparison-based sorting algorithm must run in Ω(n log n) time.

Sorting Fun24 Bucket-Sort and Radix-Sort B 1, c7, d7, g3, b3, a7, e 

Sorting Fun25 Bucket-Sort (§ 4.5.1) Let be S be a sequence of n (key, element) items with keys in the range [0, N  1] Bucket-sort uses the keys as indices into an auxiliary array B of sequences (buckets) Phase 1: Empty sequence S by moving each item (k, o) into its bucket B[k] Phase 2: For i  0, …, N  1, move the items of bucket B[i] to the end of sequence S Analysis: Phase 1 takes O(n) time Phase 2 takes O(n  N) time Bucket-sort takes O(n  N) time Algorithm bucketSort(S, N) Input sequence S of (key, element) items with keys in the range [0, N  1] Output sequence S sorted by increasing keys B  array of N empty sequences while  S.isEmpty() f  S.first() (k, o)  S.remove(f) B[k].insertLast((k, o)) for i  0 to N  1 while  B[i].isEmpty() f  B[i].first() (k, o)  B[i].remove(f) S.insertLast((k, o))

Sorting Fun26 Example Key range [0, 9] 7, d1, c3, a7, g3, b7, e1, c3, a3, b7, d7, g7, e Phase 1 Phase B 1, c7, d7, g3, b3, a7, e 

Sorting Fun27 Properties and Extensions Key-type Property The keys are used as indices into an array and cannot be arbitrary objects No external comparator Stable Sort Property The relative order of any two items with the same key is preserved after the execution of the algorithm Extensions Integer keys in the range [a, b]  Put item (k, o) into bucket B[k  a] String keys from a set D of possible strings, where D has constant size (e.g., names of the 50 U.S. states)  Sort D and compute the rank r(k) of each string k of D in the sorted sequence  Put item (k, o) into bucket B[r(k)]

Sorting Fun28 Lexicographic Order A d- tuple is a sequence of d keys (k 1, k 2, …, k d ), where key k i is said to be the i- th dimension of the tuple Example: The Cartesian coordinates of a point in space are a 3-tuple The lexicographic order of two d- tuples is recursively defined as follows (x 1, x 2, …, x d )  (y 1, y 2, …, y d )  x 1  y 1  x 1   y 1  (x 2, …, x d )  (y 2, …, y d ) I.e., the tuples are compared by the first dimension, then by the second dimension, etc.

Sorting Fun29 Lexicographic-Sort Let C i be the comparator that compares two tuples by their i- th dimension Let stableSort(S, C) be a stable sorting algorithm that uses comparator C Lexicographic-sort sorts a sequence of d- tuples in lexicographic order by executing d times algorithm stableSort, one per dimension Lexicographic-sort runs in O(dT(n)) time, where T(n) is the running time of stableSort Algorithm lexicographicSort(S) Input sequence S of d-tuples Output sequence S sorted in lexicographic order for i  d downto 1 stableSort(S, C i ) Example: (7,4,6) (5,1,5) (2,4,6) (2, 1, 4) (3, 2, 4) (2, 1, 4) (3, 2, 4) (5,1,5) (7,4,6) (2,4,6) (2, 1, 4) (5,1,5) (3, 2, 4) (7,4,6) (2,4,6) (2, 1, 4) (2,4,6) (3, 2, 4) (5,1,5) (7,4,6)

Sorting Fun30 Radix-Sort (§ 4.5.2) Radix-sort is a specialization of lexicographic-sort that uses bucket-sort as the stable sorting algorithm in each dimension Radix-sort is applicable to tuples where the keys in each dimension i are integers in the range [0, N  1] Radix-sort runs in time O(d( n  N)) Algorithm radixSort(S, N) Input sequence S of d-tuples such that (0, …, 0)  (x 1, …, x d ) and (x 1, …, x d )  (N  1, …, N  1) for each tuple (x 1, …, x d ) in S Output sequence S sorted in lexicographic order for i  d downto 1 bucketSort(S, N)

Sorting Fun31 Radix-Sort for Binary Numbers Consider a sequence of n b -bit integers x  x b  … x 1 x 0 We represent each element as a b -tuple of integers in the range [0, 1] and apply radix-sort with N  2 This application of the radix-sort algorithm runs in O(bn) time For example, we can sort a sequence of 32-bit integers in linear time Algorithm binaryRadixSort(S) Input sequence S of b-bit integers Output sequence S sorted replace each element x of S with the item (0, x) for i  0 to b  1 replace the key k of each item (k, x) of S with bit x i of x bucketSort(S, 2)

Sorting Fun32 Example Sorting a sequence of 4-bit integers