Section 5: More Parallel Algorithms

Slides:



Advertisements
Similar presentations
2. Getting Started Heejin Park College of Information and Communications Hanyang University.
Advertisements

Data Structures Through C
Chapter 9 continued: Quicksort
13.1 Vis_2003 Data Visualization Lecture 13 Visualization of Very Large Datasets.
Comp 122, Spring 2004 Order Statistics. order - 2 Lin / Devi Comp 122 Order Statistic i th order statistic: i th smallest element of a set of n elements.
David Luebke 1 6/1/2014 CS 332: Algorithms Medians and Order Statistics Structures for Dynamic Sets.
CS16: Introduction to Data Structures & Algorithms
Ack: Several slides from Prof. Jim Anderson’s COMP 202 notes.
Parallel List Ranking Advanced Algorithms & Data Structures Lecture Theme 17 Prof. Dr. Th. Ottmann Summer Semester 2006.
Abstract Data Types and Algorithms
Topic 14 Searching and Simple Sorts "There's nothing in your head the sorting hat can't see. So try me on and I will tell you where you ought to be." -The.
Briana B. Morrison Adapted from William Collins
1 Symbol Tables Chapter Sedgewick. 2 Symbol Tables Searching Searching is a fundamental element of many computational tasks looking up a name.
Introduction to Algorithms Quicksort
Squares and Square Root WALK. Solve each problem REVIEW:
ITEC200 Week10 Sorting. pdp 2 Learning Objectives – Week10 Sorting (Chapter10) By working through this chapter, students should: Learn.
Addition 1’s to 20.
25 seconds left…...
Week 1.
Binary Search Trees Ravi Chugh March 28, Review: Linked Lists Goal: Program that keeps track of friends Problem: Arrays have fixed length Solution:
18-Dec-14 Pruning. 2 Exponential growth How many leaves are there in a complete binary tree of depth N? This is easy to demonstrate: Count “going left”
Chapter 12 Binary Search Trees
CSE Lecture 17 – Balanced trees
Topic 16 Sorting Using ADTs to Implement Sorting Algorithms.
Foundations of Data Structures Practical Session #11 Sort properties, Quicksort algorithm.
CS Divide and Conquer/Recurrence Relations1 Divide and Conquer.
Transform and Conquer Chapter 6. Transform and Conquer Solve problem by transforming into: a more convenient instance of the same problem (instance simplification)
Quicksort Quicksort     29  9.
Sorting Chapter Sorting Consider list x 1, x 2, x 3, … x n We seek to arrange the elements of the list in order –Ascending or descending Some O(n.
© 2004 Goodrich, Tamassia Quick-Sort     29  9.
CS 206 Introduction to Computer Science II 04 / 27 / 2009 Instructor: Michael Eckmann.
A Sophomoric Introduction to Shared-Memory Parallelism and Concurrency Lecture 3 Parallel Prefix, Pack, and Sorting Dan Grossman Last Updated: November.
CSE332: Data Abstractions Lecture 20: Parallel Prefix and Parallel Sorting Dan Grossman Spring 2010.
CSE332: Data Abstractions Lecture 21: Parallel Prefix and Parallel Sorting Tyler Robison Summer
1 Lecture 11 Sorting Parallel Computing Fall 2008.
CS 206 Introduction to Computer Science II 12 / 03 / 2008 Instructor: Michael Eckmann.
Sorting CS-212 Dick Steflik. Exchange Sorting Method : make n-1 passes across the data, on each pass compare adjacent items, swapping as necessary (n-1.
MergeSort Source: Gibbs & Tamassia. 2 MergeSort MergeSort is a divide and conquer method of sorting.
A Sophomoric Introduction to Shared-Memory Parallelism and Concurrency Lecture 3 Parallel Prefix, Pack, and Sorting Steve Wolfman, based on work by Dan.
Computer Science and Software Engineering University of Wisconsin - Platteville 12. Heap Yan Shi CS/SE 2630 Lecture Notes Partially adopted from C++ Plus.
CSCE 3110 Data Structures & Algorithm Analysis Sorting (I) Reading: Chap.7, Weiss.
C++ Programming: Program Design Including Data Structures, Fourth Edition Chapter 19: Searching and Sorting Algorithms.
C++ Programming: From Problem Analysis to Program Design, Second Edition Chapter 19: Searching and Sorting.
1 Joe Meehean.  Problem arrange comparable items in list into sorted order  Most sorting algorithms involve comparing item values  We assume items.
Adapted from instructor resource slides Nyhoff, ADTs, Data Structures and Problem Solving with C++, Second Edition, © 2005 Pearson Education, Inc. All.
Heapsort. Heapsort is a comparison-based sorting algorithm, and is part of the selection sort family. Although somewhat slower in practice on most machines.
Sorting Chapter Sorting Consider list x 1, x 2, x 3, … x n We seek to arrange the elements of the list in order –Ascending or descending Some O(n.
CS 206 Introduction to Computer Science II 04 / 22 / 2009 Instructor: Michael Eckmann.
Chapter 18: Searching and Sorting Algorithms. Objectives In this chapter, you will: Learn the various search algorithms Implement sequential and binary.
Week 13 - Friday.  What did we talk about last time?  Sorting  Insertion sort  Merge sort  Started quicksort.
Asymptotic Behavior Algorithm : Design & Analysis [2]
Tree Data Structures. Heaps for searching Search in a heap? Search in a heap? Would have to look at root Would have to look at root If search item smaller.
Chapter 4, Part II Sorting Algorithms. 2 Heap Details A heap is a tree structure where for each subtree the value stored at the root is larger than all.
PREVIOUS SORTING ALGORITHMS  BUBBLE SORT –Time Complexity: O(n 2 ) For each item, make (n –1) comparisons Gives: Comparisons = (n –1) + (n – 2)
Week 13 - Wednesday.  What did we talk about last time?  NP-completeness.
Chapter 9: Sorting1 Sorting & Searching Ch. # 9. Chapter 9: Sorting2 Chapter Outline  What is sorting and complexity of sorting  Different types of.
CSCI-455/552 Introduction to High Performance Computing Lecture 21.
A Sophomoric Introduction to Shared-Memory Parallelism and Concurrency Lecture 3 Parallel Prefix, Pack, and Sorting Steve Wolfman, based on work by Dan.
Data Structures and Algorithms Instructor: Tesfaye Guta [M.Sc.] Haramaya University.
Parallel Prefix, Pack, and Sorting. Outline Done: –Simple ways to use parallelism for counting, summing, finding –Analysis of running time and implications.
Mergesort example: Merge as we return from recursive calls Merge Divide 1 element 829.
Chapter 11 Sorting Acknowledgement: These slides are adapted from slides provided with Data Structures and Algorithms in C++, Goodrich, Tamassia and Mount.
Quick-Sort 9/12/2018 3:26 PM Presentation for use with the textbook Data Structures and Algorithms in Java, 6th edition, by M. T. Goodrich, R. Tamassia,
Quick-Sort 9/13/2018 1:15 AM Quick-Sort     2
Parallel Sorting Algorithms
CSE332: Data Abstractions Lecture 20: Parallel Prefix, Pack, and Sorting Dan Grossman Spring 2012.
Parallel Sorting Algorithms
Quick-Sort 4/25/2019 8:10 AM Quick-Sort     2
CSE 332: Parallel Algorithms
Presentation transcript:

Section 5: More Parallel Algorithms Michelle Kuttel mkuttel@cs.uct.ac.za

The prefix-sum problem Given int[] input, produce int[] output where output[i] is the sum of input[0]+input[1]+…+input[i] Sequential can be a CS1 exam problem: int[] prefix_sum(int[] input){ int[] output = new int[input.length]; output[0] = input[0]; for(int i=1; i < input.length; i++) output[i] = output[i-1]+input[i]; return output; } Does not seem parallelizable Work: O(n), Span: O(n) This algorithm is sequential, but a different algorithm has Work: O(n), Span: O(log n) slide adapted from: Sophomoric Parallelism and Concurrency, Lecture 3

Parallel prefix-sum The parallel-prefix algorithm does two passes Each pass has O(n) work and O(log n) span So in total there is O(n) work and O(log n) span So just like with array summing, the parallelism is n/log n, an exponential speedup The first pass builds a tree bottom-up: the “up” pass The second pass traverses the tree top-down: the “down” pass Historical note: Original algorithm due to R. Ladner and M. Fischer at the University of Washington in 1977 slide adapted from: Sophomoric Parallelism and Concurrency, Lecture 3

Example input 6 4 16 10 16 14 2 8 output range 0,8 sum fromleft 76 36 40 range 0,2 sum fromleft range 2,4 sum fromleft range 4,6 sum fromleft range 6,8 sum fromleft 10 26 30 10 r 0,1 s f r 1,2 s f r 2,3 s f r 3,4 s f r 4,5 s f r 5,6 s f r 6,7 s f r 7.8 s f 6 4 16 10 16 14 2 8 input 6 4 16 10 16 14 2 8 output slide from: Sophomoric Parallelism and Concurrency, Lecture 3

Example range 0,8 sum fromleft 76 range 0,4 sum fromleft range 4,8 sum fromleft 36 40 36 range 0,2 sum fromleft range 2,4 sum fromleft range 4,6 sum fromleft range 6,8 sum fromleft 10 26 30 10 10 36 66 r 0,1 s f r 1,2 s f r 2,3 s f r 3,4 s f r 4,5 s f r 5,6 s f r 6,7 s f r 7.8 s f 6 4 16 10 16 14 2 8 6 10 26 36 52 66 68 input 6 4 16 10 16 14 2 8 output 6 10 26 36 52 66 68 76 slide from: Sophomoric Parallelism and Concurrency, Lecture 3

The algorithm, part 1 Up: Build a binary tree where Root has sum of the range [x,y) If a node has sum of [lo,hi) and hi>lo, Left child has sum of [lo,middle) Right child has sum of [middle,hi) A leaf has sum of [i,i+1), i.e., input[i] This is an easy fork-join computation: combine results by actually building a binary tree with all the range-sums Tree built bottom-up in parallel Could be more clever with an array, as with heaps Analysis: O(n) work, O(log n) span slide adapted from: Sophomoric Parallelism and Concurrency, Lecture 3

The algorithm, part 2 Down: Pass down a value fromLeft Root given a fromLeft of 0 Node takes its fromLeft value and Passes its left child the same fromLeft Passes its right child its fromLeft plus its left child’s sum (as stored in part 1) At the leaf for array position i, output[i]=fromLeft+input[i] This is an easy fork-join computation: traverse the tree built in step 1 and produce no result Leaves assign to output Invariant: fromLeft is sum of elements left of the node’s range Analysis: O(n) work, O(log n) span slide from: Sophomoric Parallelism and Concurrency, Lecture 3

Sequential cut-off Adding a sequential cut-off is easy as always: Up: just a sum, have leaf node hold the sum of a range Down: output[lo] = fromLeft + input[lo]; for(i=lo+1; i < hi; i++) output[i] = output[i-1] + input[i] slide from: Sophomoric Parallelism and Concurrency, Lecture 3

Parallel prefix, generalized Just as sum-array was the simplest example of a pattern that matches many, many problems, so is prefix-sum Minimum, maximum of all elements to the left of i Is there an element to the left of i satisfying some property? Count of elements to the left of i satisfying some property This last one is perfect for an efficient parallel pack… Perfect for building on top of the “parallel prefix trick” We did an inclusive sum, but exclusive is just as easy slide from: Sophomoric Parallelism and Concurrency, Lecture 3

Pack [Non-standard terminology] Given an array input, produce an array output containing only elements such that f(elt) is true in the same order… Example: input [17, 4, 6, 8, 11, 5, 13, 19, 0, 24] f: is elt > 10 output [17, 11, 13, 19, 24] Parallelizable? Finding elements for the output is easy But getting them in the right place seems hard slide from: Sophomoric Parallelism and Concurrency, Lecture 3

Parallel prefix to the rescue Parallel map to compute a bit-vector for true elements input [17, 4, 6, 8, 11, 5, 13, 19, 0, 24] bits [1, 0, 0, 0, 1, 0, 1, 1, 0, 1] Parallel-prefix sum on the bit-vector bitsum [1, 1, 1, 1, 2, 2, 3, 4, 4, 5] Parallel map to produce the output output [17, 11, 13, 19, 24] output = new array of size bitsum[n-1] FORALL(i=0; i < input.length; i++){ if(bits[i]==1) output[bitsum[i]-1] = input[i]; } slide from: Sophomoric Parallelism and Concurrency, Lecture 3

Pack comments First two steps can be combined into one pass Just using a different base case for the prefix sum No effect on asymptotic complexity Can also combine third step into the down pass of the prefix sum Again no effect on asymptotic complexity Analysis: O(n) work, O(log n) span 2 or 3 passes, but 3 is a constant Parallelized packs will help us parallelize quicksort… slide from: Sophomoric Parallelism and Concurrency, Lecture 3

Quicksort review Very popular sequential sorting algorithm that performs well with an average sequential time complexity of O(nlogn). First list divided into two sublists. All the numbers in one sublist arranged to be smaller than all the numbers in the other sublist. Achieved by first selecting one number, called a pivot, against which every other number is compared. If the number is less than the pivot, it is placed in one sublist. Otherwise, it is placed in the other sublist.

Quicksort review sequential, in-place, expected time O(n log n) Best / expected case work Pick a pivot element O(1) Partition all the data into: O(n) The elements less than the pivot The pivot The elements greater than the pivot Recursively sort A and C 2T(n/2) How should we parallelize this? slide adapted from: Sophomoric Parallelism and Concurrency, Lecture 3

Quicksort Best / expected case work Pick a pivot element O(1) Partition all the data into: O(n) The elements less than the pivot The pivot The elements greater than the pivot Recursively sort A and C 2T(n/2) Easy: Do the two recursive calls in parallel Work: unchanged, of course, O(n log n) Span: Now T(n) = O(n) + 1T(n/2) = O(n) So parallelism (i.e., work / span) is O(log n) slide adapted from: Sophomoric Parallelism and Concurrency, Lecture 3

Naïve Parallelization of Quicksort

Parallelizing Quicksort With the pivot being withheld in processes:

Analysis Fundamental problem with all tree constructions – initial division done by a single thread, which will seriously limit speed. Tree in quicksort will not, in general, be perfectly balanced Pivot selection very important to make quicksort operate fast.

Doing better O(log n) speed-up with an infinite number of processors is okay, but a bit underwhelming Sort 109 elements 30 times faster Google searches strongly suggest quicksort cannot do better because the partition cannot be parallelized The Internet has been known to be wrong  But we need auxiliary storage (no longer in place) In practice, constant factors may make it not worth it, but remember Amdahl’s Law Already have everything we need to parallelize the partition… slide from: Sophomoric Parallelism and Concurrency, Lecture 3

Parallel partition (not in place) Partition all the data into: The elements less than the pivot The pivot The elements greater than the pivot This is just two packs! We know a pack is O(n) work, O(log n) span Pack elements less than pivot into left side of aux array Pack elements greater than pivot into right size of aux array Put pivot between them and recursively sort With a little more cleverness, can do both packs at once but no effect on asymptotic complexity With O(log n) span for partition, the total span for quicksort is T(n) = O(log n) + 1T(n/2) = O(log2 n) Hence the available parallelism is proportional to n log n/log2 n = n/ log n an exponential speed-up. slide from: Sophomoric Parallelism and Concurrency, Lecture 3

Example Step 1: pick pivot as median of three 8 1 4 9 3 5 2 7 6 3 5 2 7 6 Steps 2a and 2c (combinable): pack less than, then pack greater than into a second array Fancy parallel prefix to pull this off not shown 1 4 3 5 2 1 4 3 5 2 6 8 9 7 Step 3: Two recursive sorts in parallel Can sort back into original array (like in mergesort) slide from: Sophomoric Parallelism and Concurrency, Lecture 3