CS224 Spring 2011 Computer Organization CS224 Chapter 6A: Disk Systems With thanks to M.J. Irwin, D. Patterson, and J. Hennessy for some lecture slide.

Slides:



Advertisements
Similar presentations
1 Lecture 22: I/O, Disk Systems Todays topics: I/O overview Disk basics RAID Reminder: Assignment 8 due Tue 11/21.
Advertisements

IT253: Computer Organization
Disk Arrays COEN 180. Large Storage Systems Collection of disks to store large amount of data. Performance advantage: Each drive can satisfy only so many.
RAID Redundant Arrays of Independent Disks Courtesy of Satya, Fall 99.
A CASE FOR REDUNDANT ARRAYS OF INEXPENSIVE DISKS (RAID) D. A. Patterson, G. A. Gibson, R. H. Katz University of California, Berkeley.
I/O Chapter 8. Outline Introduction Disk Storage and Dependability – 8.2 Buses and other connectors – 8.4 I/O performance measures – 8.6.
Lecture 4: A Case for RAID (Part 2) Prof. Shahram Ghandeharizadeh Computer Science Department University of Southern California.
RAID Redundant Array of Inexpensive Disks Presented by Greg Briggs.
Copyright © 2009 EMC Corporation. Do not Copy - All Rights Reserved.
I/O Management and Disk Scheduling
1 Lecture 18: RAID n I/O bottleneck n JBOD and SLED n striping and mirroring n classic RAID levels: 1 – 5 n additional RAID levels: 6, 0+1, 10 n RAID usage.
IT 344: Operating Systems Winter 2007 Module 18 Redundant Arrays of Inexpensive Disks (RAID) Chia-Chi Teng CTB 265.
1 Jason Drown Mark Rodden (Redundant Array of Inexpensive Disks) RAID.
Faculty of Information Technology Department of Computer Science Computer Organization Chapter 7 External Memory Mohammad Sharaf.
RAID (Redundant Arrays of Independent Disks). Disk organization technique that manages a large number of disks, providing a view of a single disk of High.
CSE431 Chapter 6A.1Irwin, PSU, 2008 Chapter 6A: Disk Systems Mary Jane Irwin ( ) [Adapted from Computer Organization.
Princess Sumaya Univ. Computer Engineering Dept. Chapter 6:
RAID Redundant Array of Independent Disks
Storage and Disks.
CS 6560: Operating Systems Design
CSCE430/830 Computer Architecture
CPE 442 io.1 Introduction To Computer Architecture CpE 442 I/O Systems.
RAID- Redundant Array of Inexpensive Drives. Purpose Provide faster data access and larger storage Provide data redundancy.
RAID Redundant Arrays of Inexpensive Disks –Using lots of disk drives improves: Performance Reliability –Alternative: Specialized, high-performance hardware.
Lecture 36: Chapter 6 Today’s topic –RAID 1. RAID Redundant Array of Inexpensive (Independent) Disks –Use multiple smaller disks (c.f. one large disk)
CSCE 212 Chapter 8 Storage, Networks, and Other Peripherals Instructor: Jason D. Bakos.
Computer ArchitectureFall 2007 © November 28, 2007 Karem A. Sakallah Lecture 24 Disk IO and RAID CS : Computer Architecture.
1 Lecture 26: Storage Systems Topics: Storage Systems (Chapter 6), other innovations Final exam stats:  Highest: 95  Mean: 70, Median: 73  Toughest.
Lecture 3: A Case for RAID (Part 1) Prof. Shahram Ghandeharizadeh Computer Science Department University of Southern California.
1 Lecture 27: Disks, Reliability, SSDs, Processors Topics: HDDs, SSDs, RAID, Intel and IBM case studies Final exam stats:  Highest 91, 18 scores of 82+
Computer ArchitectureFall 2008 © November 12, 2007 Nael Abu-Ghazaleh Lecture 24 Disk IO.
S.1 Review: Major Components of a Computer Processor Control Datapath Memory Devices Input Output Cache Main Memory Secondary Memory (Disk)
RAID Systems CS Introduction to Operating Systems.
Lecture 11: Storage Systems Disk, RAID, Dependability Kai Bu
ICOM 6005 – Database Management Systems Design Dr. Manuel Rodríguez-Martínez Electrical and Computer Engineering Department Lecture 6 – RAID ©Manuel Rodriguez.
Lecture 11: Storage Systems Disk, RAID, Dependability Kai Bu
Chapter 6 RAID. Chapter 6 — Storage and Other I/O Topics — 2 RAID Redundant Array of Inexpensive (Independent) Disks Use multiple smaller disks (c.f.
Eng. Mohammed Timraz Electronics & Communication Engineer University of Palestine Faculty of Engineering and Urban planning Software Engineering Department.
Storage & Peripherals Disks, Networks, and Other Devices.
CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Storage Systems.
CSE 321b Computer Organization (2) تنظيم الحاسب (2) 3 rd year, Computer Engineering Winter 2015 Lecture #4 Dr. Hazem Ibrahim Shehata Dept. of Computer.
L/O/G/O External Memory Chapter 3 (C) CS.216 Computer Architecture and Organization.
CSE431 Chapter 6A.1Irwin, PSU, 2008 CSE 431 Computer Architecture Fall 2008 Chapter 6A: Disk Systems Mary Jane Irwin ( )
CSE431 Chapter 6A.1Irwin, PSU, 2008 Chapter 6A: Disk Systems Mary Jane Irwin ( ) [Adapted from Computer Organization.
I/O – Chapter 8 Introduction Disk Storage and Dependability – 8.2 Buses and other connectors – 8.4 I/O performance measures – 8.6.
Lecture 9 of Advanced Databases Storage and File Structure (Part II) Instructor: Mr.Ahmed Al Astal.
Redundant Array of Inexpensive Disks aka Redundant Array of Independent Disks (RAID) Modified from CCT slides.
I/O Computer Organization II 1 Introduction I/O devices can be characterized by – Behavior: input, output, storage – Partner: human or machine – Data rate:
1 Lecture 27: Disks Today’s topics:  Disk basics  RAID  Research topics.
1 Lecture 23: Storage Systems Topics: disk access, bus design, evaluation metrics, RAID (Sections )
COSC 6340: Disks 1 Disks and Files DBMS stores information on (“hard”) disks. This has major implications for DBMS design! » READ: transfer data from disk.
Mohamed Younis CMCS 411, Computer Architecture 1 CMCS Computer Architecture Lecture 25 I/O Systems May 2,
LECTURE 13 I/O. I/O CANNOT BE IGNORED Assume a program requires 100 seconds, 90 seconds for main memory, 10 seconds for I/O. Assume main memory access.
RAID TECHNOLOGY RASHMI ACHARYA CSE(A) RG NO
Network-Attached Storage. Network-attached storage devices Attached to a local area network, generally an Ethernet-based network environment.
Lecture 11: Storage Systems Disk, RAID, Dependability Kai Bu
CMSC 611: Advanced Computer Architecture I/O & Storage Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted.
I/O Errors 1 Computer Organization II © McQuain RAID Redundant Array of Inexpensive (Independent) Disks – Use multiple smaller disks (c.f.
CS Introduction to Operating Systems
Multiple Platters.
IT 251 Computer Organization and Architecture
Introduction I/O devices can be characterized by I/O bus connections
Lecture 13 I/O.
RAID RAID Mukesh N Tekwani
ICOM 6005 – Database Management Systems Design
Lecture 28: Reliability Today’s topics: GPU wrap-up Disk basics RAID
CSE 451: Operating Systems Winter 2007 Module 18 Redundant Arrays of Inexpensive Disks (RAID) Ed Lazowska Allen Center 570.
RAID RAID Mukesh N Tekwani April 23, 2019
IT 344: Operating Systems Winter 2007 Module 18 Redundant Arrays of Inexpensive Disks (RAID) Chia-Chi Teng CTB
Presentation transcript:

CS224 Spring 2011 Computer Organization CS224 Chapter 6A: Disk Systems With thanks to M.J. Irwin, D. Patterson, and J. Hennessy for some lecture slide contents Spring 2011

CS224 Spring 2011 Review: Major Components of a Computer Processor Control Datapath Memory Devices Input Output Cache Main Memory Secondary Memory (Disk)

CS224 Spring 2011 Magnetic Disk Purpose Long term, nonvolatile storage Lowest level in the memory hierarchy -slow, large, inexpensive General structure A rotating platter coated with a magnetic surface A moveable read/write head to access the information on the disk Typical numbers 1 to 4 platters (each with 2 recordable surfaces) per disk of 1 to 3.5 in diameter Rotational speeds of 5,400 to 15,000 RPM 10,000 to 50,000 tracks per surface -cylinder - all the tracks under the head at a given point on all surfaces 100 to 500 sectors per track -the smallest unit that can be read/written (typically 512B) Track Sector

CS224 Spring 2011 Magnetic Disk Characteristic Disk read/write components 1. Seek time: position the head over the proper track (3 to 13 ms avg) -due to locality of disk references the actual average seek time may be only 25% to 33% of the advertised number 2. Rotational latency: wait for the desired sector to rotate under the head (½ of 1 rotation/RPM, converted to ms) -0.5/5400RPM = 5.6 ms to 0.5/15000RPM = 2.0 ms 3. Transfer time: transfer a block of bits (one or more sectors) under the head to the disk controllers cache (70 to 125 MB/s are typical disk transfer rates in 2008) -disk controllers cache takes advantage of spatial locality in disk accesses. Cache transfer rates are much faster (e.g.375 MB/s) 4. Controller time: the overhead the disk controller imposes in performing a disk I/O access (typically <0.2 ms) Sector Track Cylinder Head Platter Controller + Cache

CS224 Spring 2011 Typical Disk Access Time The average time to read or write a 512B sector for a disk rotating at 15,000 RPM with average seek time of 4 ms, a 100MB/sec transfer rate, and a 0.2 ms controller overhead If the measured average seek time is 25% of the advertised average seek time, then The rotational latency is usually the largest component of the access time

CS224 Spring 2011 Typical Disk Access Time If the measured average seek time is 25% of the advertised average seek time, then Avg disk read/write = 4.0 ms + 0.5/(15,000RPM/(60sec/min))+ 0.5KB/(100MB/sec) ms = = 6.2 ms Avg disk read/write = = 3.2 ms The average time to read or write a 512B sector for a disk rotating at 15,000 RPM with average seek time of 4 ms, a 100MB/sec transfer rate, and a 0.2 ms controller overhead The rotational latency is usually the largest component of the access time!

CS224 Spring 2011 Disk Interface Standards Higher-level disk interfaces have a microprocessor disk controller that can lead to performance optimizations ATA (Advanced Technology Attachment ) – An interface standard for the connection of storage devices such as hard disks, solid-state drives, and CD-ROM drives. Parallel ATA has been largely replaced by serial ATA. Widely used in PCs. SCSI (Small Computer Systems Interface) – A set of standards (commands, protocols, and electrical and optical interfaces) for physically connecting and transferring data between computers and peripheral devices. Most commonly used for hard disks and tape drives. In particular, disk controllers have SRAM disk caches which support fast access to data that was recently read and often also include prefetch algorithms to try to anticipate demand.

CS224 Spring 2011 Magnetic Disk Examples ( FeatureSeagate ST NS Seagate ST973451SS Seagate ST AS Disk diameter (inches) Capacity (GB) # of surfaces (heads)422 Rotation speed (RPM)7,20015,0005,400 Transfer rate (MB/sec) Minimum seek (ms)0.8r-1.0w0.2r-0.4w1.5r-2.0w Average seek (ms)8.5r-9.5w2.9r-3.3w12.5r-13.0w MTTF o C)1,200,0001,600,000?? Dim (inches), Weight (lbs) 1x4x5.8, x2.8x3.9, x2.8x3.9, 0.2 GB/cu.inch, GB/watt43, 9111, 937, 84 Power: op/idle/sb (watts)11/8/18/5.8/-1.9/0.6/0.2 Price in 2008, $/GB~$0.3/GB~$5/GB~$0.6/GB

CS224 Spring 2011 Magnetic Disk Examples ( FeatureSeagate ST37Seagate ST32Seagate ST94 Disk diameter (inches) Capacity (GB) # of surfaces (heads)842 Rotation speed (RPM)15,0007,2005,400 Transfer rate (MB/sec) Minimum seek (ms)0.2r-0.4w1.0r-1.2w1.5r-2.0w Average seek (ms)3.6r-3.9w8.5r-9.5w12r-14w MTTF o C)1,200,000600,000330,000 Dimensions (inches)1x4x x2.7x3.9 GB/cu.inch3910 Power: op/idle/sb (watts)20?/12/-12/8/12.4/1/0.4 GB/watt41617 Weight (pounds)

CS224 Spring 2011 Disk Latency & Bandwidth Milestones Disk latency is one average seek time plus the rotational latency. Disk bandwidth is the peak transfer time of formatted data from the media (not from the cache). CDC WrenSG ST41SG ST15SG ST39SG ST37 RSpeed (RPM) Year Capacity (Gbytes) Diameter (inches) InterfaceST-412SCSI Bandwidth (MB/s) Latency (msec) Patterson, CACM Vol 47, #10, 2004

CS224 Spring 2011 Latency & Bandwidth Improvements In the time that the disk bandwidth doubles the latency improves by a factor of only 1.2 to 1.4

CS224 Spring 2011 Aside: Media Bandwidth/Latency Demands Bandwidth requirements High quality video -Digital data = (30 frames/s) × (640 x 480 pixels) × (24-b color/pixel) = 221 Mb/s ( MB/s) High quality audio -Digital data = (44,100 audio samples/s) × (16-b audio samples) × (2 audio channels for stereo) = 1.4 Mb/s (0.175 MB/s) Compression reduces the bandwidth requirements considerably Latency issues How sensitive is your eye (ear) to variations in video (audio) rates? How can you ensure a constant rate of delivery? How important is synchronizing the audio and video streams? -15 to 20 ms early to 30 to 40 ms late is tolerable

CS224 Spring 2011 Flash Storage Flash memory is the first credible challenger to disks. It is semiconductor memory that is nonvolatile like disks, but has latency 100 to 1000 times faster than disk and is smaller, more power efficient, and more shock resistant. In 2008, the price of flash is $4 to $10 per GB or about 2 to 10 times higher than disk and 5 to 10 times lower than DRAM. Flash memory bits wear out (unlike disks and DRAMs), but wear leveling can make it unlikely that the write limits of the flash will be exceeded FeatureKingstonTransendRiDATA Capacity (GB)81632 Bytes/sector512 Transfer rates (MB/sec)420r-18w68r-50w MTTF>1,000,000 >4,000,000 Price (2008)~ $30~ $70~ $300

CS224 Spring 2011 Dependability, Reliability, Availability Reliability – measured by the mean time to failure (MTTF). Service interruption is measured by mean time to repair (MTTR) Availability – a measure of service accomplishment Availability = MTTF/(MTTF + MTTR) To increase MTTF, either improve the quality of the components or design the system to continue operating in the presence of faulty components 1. Fault avoidance: preventing fault occurrence by construction 2. Fault tolerance: using redundancy to correct or bypass faulty components (hardware) Fault detection versus fault correction Permanent faults versus transient faults

CS224 Spring 2011 RAIDs: Disk Arrays Arrays of small and inexpensive disks Increase potential throughput by having many disk drives -Data is spread over multiple disk -Multiple accesses are made to several disks at a time Reliability is lower than a single disk But availability can be improved by adding redundant disks (RAID) Lost information can be reconstructed from redundant information MTTR: mean time to repair is in the order of hours MTTF: mean time to failure of disks is tens of years Redundant Array of Inexpensive Disks

CS224 Spring 2011 RAID: Level 0 (No Redundancy, but Striping) Multiple smaller disks as opposed to one big disk Spreading the sector over multiple disks – striping – means that multiple blocks can be accessed in parallel increasing the performance (i.e throughput) -this 4 disk system gives four times the throughput of a 1 disk system Same cost as one big disk – assuming 4 small disks cost the same as one big disk No redundancy, so what if one disk fails? Failure of one or more disks is more likely as the number of disks in the system increases sec1sec3sec2sec4 sec1,b0sec1,b2sec1,b1sec1,b3

CS224 Spring 2011 RAID: Level 1 (Redundancy via Mirroring) Uses twice as many disks as RAID 0 (e.g., 8 smaller disks with the second set of 4 duplicating the first set) so there are always two copies of the data # redundant disks = # of data disks so reliability doubles the cost (reliable, but most expensive) -writes have to be made to both sets of disks, so writes would only have ½ the performance of a RAID 0 with 8 disks What if one disk fails? If a disk fails, the system just goes to the mirror for the data redundant (check) data sec1sec3sec2sec4sec1sec3sec2sec4

CS224 Spring 2011 RAID: Level 0+1 (Striping with Mirroring) Combines the best of RAID 0 and RAID 1, data is striped across four disks and mirrored to four disks Four times the throughput (due to striping) # redundant disks = # of data disks, so reliability doubles the cost -writes have to be made to both sets of disks, so writes would be only 1/2 the performance of a RAID 0 with 8 disks What if one disk fails? If a disk fails, the system just goes to the mirror for the data sec1blk3blk2blk4blk1blk2blk3blk4 redundant (check) data sec1,b0sec1,b2sec1,b1sec1,b3sec1,b0sec1,b2sec1,b1sec1,b3

CS224 Spring 2011 RAID: Level 2 (Redundancy via ECC) ECC disks contain the parity of data on a set of distinct overlapping disks # redundant disks ~= log (total # of data disks) + 1, so reliability almost doubles the cost -writes require computing parity to write to the ECC disks -reads require reading ECC disks and confirming parity Can tolerate limited disk failure, since the data can be reconstructed (similar to memory ECC systems) sec1,b0sec1,b2sec1,b1sec1,b3 Checks 4,5,6,7 Checks 2,3,6,7 Checks 1,3,5, ECC disks (even parity) 0 ECC disks 4 and 2 point to either data disk 6 or 7, but ECC disk 1 says disk 7 is okay, so disk 6 must be in error 1

CS224 Spring 2011 RAID: Level 3 (Bit-Interleaved Parity) Cost of higher availability is reduced to 1/N where N is the number of disks in a protection group # redundant disks = 1 × # of protection groups Reads and writes must access all disks writing new data to the data disk as well as computing and writing the parity disk, means reading the other disks, so that the parity disk can be updated Can tolerate limited disk failure, since the data can be reconstructed reads require reading all the operational data disks as well as the parity disk to calculate the missing data stored on the failed disk sec1,b0sec1,b2sec1,b1sec1,b (odd) bit parity disk

CS224 Spring 2011 RAID: Level 3 (Bit-Interleaved Parity) Cost of higher availability is reduced to 1/N where N is the number of disks in a protection group # redundant disks = 1 × # of protection groups Reads and writes must access all disks writing new data to the data disk as well as computing and writing the parity disk, means reading the other disks, so that the parity disk can be updated Can tolerate limited (single) disk failure, since the data can be reconstructed reads require reading all the operational data disks as well as the parity disk to calculate the missing data stored on the failed disk sec1,b0sec1,b2sec1,b1sec1,b (odd) bit parity disk disk fails 1

CS224 Spring 2011 RAID: Level 4 (Block-Interleaved Parity) Cost of higher availability still only 1/N but the parity is stored as blocks associated with sets of data blocks Four times the throughput (blocks are striped) # redundant disks = 1 × # of protection groups Supports small reads and small writes (reads and writes that go to just one (or a few) data disk in a protection group, not to all) -by watching which bits change when writing new information, need only to change the corresponding bits on the parity disk (read-modify-write) -the parity disk must be updated on every write, so it is a bottleneck for back-to-back writes Can tolerate limited disk failure, since the data can be reconstructed block parity disk sec1sec2sec3sec4

CS224 Spring 2011 Small Writes RAID 3 writes New D1 data D1D2D3D4P D1D2D3D4P 3 reads and 2 writes involving all the disks RAID 4 small writes New D1 data D1D2D3D4P D1D2D3D4P 2 reads and 2 writes involving just two disks

CS224 Spring 2011 RAID: Level 5 (Distributed Block-Interleaved Parity) Cost of higher availability still only 1/N but the parity block can be located on any of the disks so there is no single bottleneck for writes Still four times the throughput (block striping) # redundant disks = 1 × # of protection groups Supports small reads and small writes (reads and writes that go to just one (or a few) data disk in a protection group, not to all) Allows multiple simultaneous writes as long as the accompanying parity blocks are not located on the same disk Can tolerate limited disk failure, since the data can be reconstructed one of these assigned as the block parity disk

CS224 Spring 2011 Distributing Parity Blocks By distributing parity blocks to all disks, some small writes can be performed in parallel P P P P3 RAID 4RAID P P P P Time Can be done in parallel

CS224 Spring 2011 Summary Four components of disk access time: Seek Time: advertised to be 3 to 14 ms but lower in real systems Rotational Latency: 5.6 ms at 5400 RPM and 2.0 ms at RPM Transfer Time: 30 to 80 MB/s Controller Time: typically less than.2 ms RAIDS can be used to improve availability RAID 1 and RAID 5 – widely used in servers, one estimate is that 80% of disks in servers are RAIDs RAID 0+1 (mirroring) – EMC, HP/Tandem, IBM RAID 3 – Storage Concepts RAID 4 – Network Appliance RAIDS have enough redundancy to allow continuous operation, but hot swapping (replacement while system is running) is challenging