Download presentation
Presentation is loading. Please wait.
Published byVanessa Long Modified over 9 years ago
2
CPSC 404, Laks V.S. Lakshmanan1 External Sorting Chapter 13 (Sec. 13-1-13.5): Ramakrishnan & Gehrke and Chapter 11 (Sec. 11.4-11.5): G-M et al. (R2) OR Chapter 2 (Sec. 2.4-2.5): Garcia-et Molina al. (R1) NOTE: sorting using B-trees – to be assigned for reading after we cover B-trees.
3
CPSC 404, Laks V.S. Lakshmanan2 What you will learn from this set of lectures v How to efficiently sort a file so large it won’t all fit in memory? v What knobs are available for me to tune the performance of such a sort algorithm & how do I tune them?
4
CPSC 404, Laks V.S. Lakshmanan3 Motivation v A classic problem in computer science! v Data requested in sorted order – e.g., arrange movie records in increase order of ratings. v Why is it important? –is first step in bulk loading (i.e., creating) B+ tree index. –useful also for eliminating duplicate copies in a collection of records & for aggregation (Why?) [can you think of any other strategy for these tasks?] –(later) for efficient joins of relations, i.e., Sort-merge algo. v Problem: sort 1GB of data with 50MB of RAM.
5
CPSC 404, Laks V.S. Lakshmanan4 I/O Computation Model v disk I/O (read/write a block) very expensive compared to processing it once it’s in memory. v random block accesses very common. v reasonable model for computation using secondary storage count only disk I/O’s.
6
CPSC 404, Laks V.S. Lakshmanan5 Desiderata for good DBMS algorithms v overlap I/O and cpu processing as much as possible. v use as much data from each block read as possible; depends on record clustering. v cluster records that are accessed together in consecutive blocks/pages. v buffer frequently accessed blocks in memory.
7
CPSC 404, Laks V.S. Lakshmanan6 Merge Sort Overview Idea: given B buffer pages of main memory v (Sort phase) –read in B pages of data each time and sort internally –suppose entire data stored in m * B pages –sort phase produces m sorted runs (sublists) of B pages each
8
CPSC 404, Laks V.S. Lakshmanan7 Merge Sort Overview v (Merge phase) –repeatedly merge two sorted runs (Pass 0: the sort phase) Pass 1 produces m/2 sorted runs of 2B pages each Pass 2 produces m/4 sorted runs of 4B pages each continue until one sorted run of m * B pages produced 2-way merge can be optimized to k-way merge, where k can be as large as B-1 (see Ch. 13.2 and (R1 OR R2) chapters).
9
CPSC 404, Laks V.S. Lakshmanan8 Sorting Example Setup: v 10M records of 100 bytes = 1GB file. –Stored on Megatron 747 disk, with 4KB blocks, each holding 40 records + header information. –Suppose each cylinder = 1MB. entire file takes up 1000 cylinders 250 x 1000 blocks. v 50M available main memory = 50 x 2^20/(4 x 2^10) = 12,800 blocks = 1/20th of file. v Sort by primary key field. => ==>=>==>
10
CPSC 404, Laks V.S. Lakshmanan9 Merge Sort v Common mainmemory sorting algorithms don't optimize disk I/O's. Variants of Merge Sort do better. v Merge = take two sorted lists and repeatedly choose the smaller of the ``heads'' of the lists (head = first among the unchosen). v Example: merge 1,3,4,8 with 2,5,7,9 = 1,2,3,4,5,7,8,9. v Merge Sort based on recursive algorithm: divide list of records into two parts; recursively mergesort the parts, and merge the resulting lists.
11
CPSC 404, Laks V.S. Lakshmanan10 TwoPhase, Multiway Merge Sort v “Vanilla” Merge Sort still not very good in disk I/O model. v log 2 n passes, so each record is read/written from disk log 2 n times. v 2PMMS: 2 reads + 2 writes per block. Phase 1 v 1. Fill buffer with records. v 2. Sort using favorite mainmemory sort. v 3. Write sorted sublist to disk. v 4. Repeat until all records have been put into one of the sorted sublists. v Sorted sublist = SSL = sorted run.
12
CPSC 404, Laks V.S. Lakshmanan11 TwoPhase, Multiway Merge Sort
13
CPSC 404, Laks V.S. Lakshmanan12 TwoPhase, Multiway Merge Sort
14
CPSC 404, Laks V.S. Lakshmanan13 TwoPhase, Multiway Merge Sort
15
CPSC 404, Laks V.S. Lakshmanan14 TwoPhase, Multiway Merge Sort
16
CPSC 404, Laks V.S. Lakshmanan15 2PMMS (contd.) Phase 2 v Use one buffer for each of the sorted sublists and one buffer for output. –i.e., split available buffer into several parts (logical buffers) and allocate for various purposes. Available Memory …
17
CPSC 404, Laks V.S. Lakshmanan16 2PMMS (contd.) Phase 2 v Use one buffer for each of the sorted sublists and one buffer for output. Pointers to first unchosen Select the smallest unchosen Output buffer
18
CPSC 404, Laks V.S. Lakshmanan17 2PMMS (contd.) Phase 2 v Use one buffer for each of the sorted sublists and one buffer for an output block. Pointers to first unchosen Select the smallest unchosen Output buffer 6 5 8 5
19
CPSC 404, Laks V.S. Lakshmanan18 2PMMS (contd.) v Initially load input buffers with the first blocks of their respective sorted lists. v Repeatedly run a competition among the first unchosen records of each of the buffered blocks. –Move the record with the least key to the output block; it is now ``chosen.'' v Manage the buffers as needed: –If an input block is exhausted, get the next block from the same list (i.e., file). –If the output block is full, write it to disk. (assumes each buffer holds just one block. In practice,you design for each buffer to hold as many blocks as possible.)
20
CPSC 404, Laks V.S. Lakshmanan19 Analysis of Naïve Implementation Recall problem parameters. v Assume blocks are stored at random, so average access time includes seek + rotational delay + transfer, and is about 15 ms. v File stored on 250,000 blocks, read and written once in each phase. v 1,000,000 disk I/O's * 0.015 seconds = 15,000 seconds = 250 minutes = 4+ hours. Problem and Interlude: How many records can you sort with 2PMMS, under our assumptions about records, main memory, and the disk? What would you do if there were more?
21
CPSC 404, Laks V.S. Lakshmanan20 Improving the Running Time of 2PMMS Here are some techniques that sometimes make secondarymemory algorithms more efficient: 1. Group blocks by cylinder (“cylindrification”). 2. One big disk several smaller disks. 3. Mirror disks = multiple copies of the same data. 4. ``Prefetching'' or ``double buffering.'' 5. Disk scheduling; the ``elevator'' algorithm.
22
CPSC 404, Laks V.S. Lakshmanan21 Cylindrification If we are going to read or write blocks in a known order, place them by cylinder, so once we reach that cylinder we can read block after block, with no seek time or rotational latency. Application to Phase 1 of 2PMMS: Recall problem parameters.Recall problem parameters. 1. Initially, records on 1000 consecutive cylinders. 2. Load main memory from 50 consecutive cylinders. v Order of blocks read is unimportant, so only time besides transfer time is one random seek and 49 1cylinder seeks (neglect). v Time to transfer 12,800 blocks at 0.5 ms/block = 6.4 sec. 3. Write each sorted sublist onto 50 consecutive cylinders, so write time is also about 6.4 sec. What about for the whole file? But in Phase 2: Blocks are read from the 20 sublists in a datadependent order, so we cannot take much advantage of cylinders. (why 20 sublists? ‘cos there is 1000 cylinders worth of data). Single sorted sublist perspective
23
CPSC 404, Laks V.S. Lakshmanan22 Prefetching If we have extra space for mainmemory buffers, consider loading buffers in advance of need. Similarly, keep output blocks buffered in main memory until it is convenient to write them to disk. Example: Phase 2 of 2PMMS With 50MB of main memory, we can afford to buffer two cylinders for each sublist and for the output (recall: one cylinder = 1MB). v Consume one cylinder for each sublist while the other is being loaded from disk. v Similarly, write one output cylinder while the other is being constructed (via merge).
24
CPSC 404, Laks V.S. Lakshmanan23 Prefetching v Thus, seek and rotational latency are made negligible for Phase I (thanks to cylindrification) and are minimized for phase II (thanks to prefetching). v So, the total time for phase 2 is approx. one read plus one write of whole file, but one cylinderful at a time.
25
CPSC 404, Laks V.S. Lakshmanan24 Prefetching/Double buffering Illustrated Alternately process different halves of each input buffer while the other part is being filled from disk. Similar idea for output buffer. i/p buffer o/p buffer
26
CPSC 404, Laks V.S. Lakshmanan25 Using B+ Trees for Sorting v Scenario: Table to be sorted has B+ tree index on sorting column(s). v Idea: Can retrieve records in order by traversing leaf pages. v Is this a good idea? v Cases to consider: – B+ tree is clusteredGood idea! – B+ tree is not clusteredCould be a very bad idea!
27
CPSC 404, Laks V.S. Lakshmanan26 Clustered B+ Tree Used for Sorting v Cost: root to the left- most leaf, then retrieve all leaf pages (Alternative 1) v If Alternative 2 is used? Additional cost of retrieving data records: each page fetched just once. (Direct search) Data Records Index Data Entries ("Sequence set")
28
CPSC 404, Laks V.S. Lakshmanan27 Unclustered B+ Tree Used for Sorting v Alternative (2) for data entries; each data entry contains rid of a data record. In general, one I/O per data record! (Directs search) Data Records Index Data Entries ("Sequence set")
29
CPSC 404, Laks V.S. Lakshmanan28 Conclusion: Sorting Records! v Sorting has become a blood sport! – Parallel sorting is the name of the game... v Datamation: Sort 1M records of size 100 bytes – Typical DBMS: 15 minutes – World record: 3.5 seconds –12-CPU SGI machine, 96 disks, 2GB of RAM v New benchmarks proposed: – Minute Sort: How many can you sort in 1 minute? – Dollar Sort: How many can you sort for $1.00?
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.