Module: Storage Systems

Slides:



Advertisements
Similar presentations
Faculty of Information Technology Department of Computer Science Computer Organization Chapter 7 External Memory Mohammad Sharaf.
Advertisements

- Dr. Kalpakis CMSC Dr. Kalpakis 1 Outline In implementing DBMS we need to answer How should the system store and manage very large amounts of data?
CSCE430/830 Computer Architecture
Operating Systems ECE344 Ashvin Goel ECE University of Toronto Disks and RAID.
Lecture 36: Chapter 6 Today’s topic –RAID 1. RAID Redundant Array of Inexpensive (Independent) Disks –Use multiple smaller disks (c.f. one large disk)
CS 430 – Computer Architecture Disks
CS61C L40 I/O: Disks (1) Ho, Fall 2004 © UCB TA Casey Ho inst.eecs.berkeley.edu/~cs61c CS61C : Machine Structures Lecture 39 I/O : Disks Microsoft.
Disk Storage SystemsCSCE430/830 Disk Storage Systems CSCE430/830 Computer Architecture Lecturer: Prof. Hong Jiang Courtesy of Yifeng Zhu (U. Maine) Fall,
Computer ArchitectureFall 2007 © November 28, 2007 Karem A. Sakallah Lecture 24 Disk IO and RAID CS : Computer Architecture.
1 Lecture 26: Storage Systems Topics: Storage Systems (Chapter 6), other innovations Final exam stats:  Highest: 95  Mean: 70, Median: 73  Toughest.
Lecture 3: A Case for RAID (Part 1) Prof. Shahram Ghandeharizadeh Computer Science Department University of Southern California.
CS61C L40 I/O: Disks (1) Garcia, Fall 2004 © UCB Lecturer PSOE Dan Garcia inst.eecs.berkeley.edu/~cs61c CS61C : Machine Structures.
1 Storage Hierarchy Cache Main Memory Virtual Memory File System Tertiary Storage Programs DBMS Capacity & Cost Secondary Storage.
Computer ArchitectureFall 2008 © November 12, 2007 Nael Abu-Ghazaleh Lecture 24 Disk IO.
Disk Technologies. Magnetic Disks Purpose: – Long-term, nonvolatile, inexpensive storage for files – Large, inexpensive, slow level in the memory hierarchy.
S.1 Review: Major Components of a Computer Processor Control Datapath Memory Devices Input Output Cache Main Memory Secondary Memory (Disk)
12/3/2004EE 42 fall 2004 lecture 391 Lecture #39: Magnetic memory storage Last lecture: –Dynamic Ram –E 2 memory This lecture: –Future memory technologies.
Secondary Storage CSCI 444/544 Operating Systems Fall 2008.
1 CS222: Principles of Database Management Fall 2010 Professor Chen Li Department of Computer Science University of California, Irvine Notes 01.
CS 61C L41 I/O Disks (1) Garcia, Spring 2004 © UCB Lecturer PSOE Dan Garcia inst.eecs.berkeley.edu/~cs61c CS61C : Machine.
1 COMP 206: Computer Architecture and Implementation Montek Singh Wed., Nov. 20, 2002 Topic: Storage Systems (Disk Technology)
CPSC 231 Secondary storage (D.H.)1 Learning Objectives Understanding disk organization. Sectors, clusters and extents. Fragmentation. Disk access time.
Introduction to Database Systems 1 The Storage Hierarchy and Magnetic Disks Storage Technology: Topic 1.
Secondary Storage Unit 013: Systems Architecture Workbook: Secondary Storage 1G.
Storage & Peripherals Disks, Networks, and Other Devices.
CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Storage Systems.
L/O/G/O External Memory Chapter 3 (C) CS.216 Computer Architecture and Organization.
CSE431 Chapter 6A.1Irwin, PSU, 2008 Chapter 6A: Disk Systems Mary Jane Irwin ( ) [Adapted from Computer Organization.
I/O – Chapter 8 Introduction Disk Storage and Dependability – 8.2 Buses and other connectors – 8.4 I/O performance measures – 8.6.
1 Chapter 7: Storage Systems Introduction Magnetic disks Buses RAID: Redundant Arrays of Inexpensive Disks.
Disks Chapter 5 Thursday, April 5, Today’s Schedule Input/Output – Disks (Chapter 5.4)  Magnetic vs. Optical Disks  RAID levels and functions.
Disk Storage SystemsCSCE430/830 Disk Storage Systems CSCE430/830 Computer Architecture Lecturer: Prof. Hong Jiang Courtesy of Yifeng Zhu (U. Maine) Fall,
Auxiliary Memory Magnetic Disk:
Csci 136 Computer Architecture II – IO and Storage Systems Xiuzhen Cheng
1 Lecture 27: Disks Today’s topics:  Disk basics  RAID  Research topics.
1 Lecture 23: Storage Systems Topics: disk access, bus design, evaluation metrics, RAID (Sections )
Abstract Increases in CPU and memory will be wasted if not matched by similar performance in I/O SLED vs. RAID 5 levels of RAID and respective cost/performance.
1 Lecture: Storage, GPUs Topics: disks, RAID, reliability, GPUs (Appendix D, Ch 4)
CPSC 231 Secondary storage (D.H.)1 Learning Objectives Understanding disk organization. Sectors, clusters and extents. Fragmentation. Disk access time.
COSC 6340: Disks 1 Disks and Files DBMS stores information on (“hard”) disks. This has major implications for DBMS design! » READ: transfer data from disk.
Mohamed Younis CMCS 411, Computer Architecture 1 CMCS Computer Architecture Lecture 25 I/O Systems May 2,
W4118 Operating Systems Instructor: Junfeng Yang.
LECTURE 13 I/O. I/O CANNOT BE IGNORED Assume a program requires 100 seconds, 90 seconds for main memory, 10 seconds for I/O. Assume main memory access.
CS422 Principles of Database Systems Disk Access Chengyu Sun California State University, Los Angeles.
CMSC 611: Advanced Computer Architecture I/O & Storage Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted.
Magnetic Disks Have cylinders, sectors platters, tracks, heads virtual and real disk blocks (x cylinders, y heads, z sectors per track) Relatively slow,
A Case for Redundant Arrays of Inexpensive Disks (RAID) -1988
A Case for Redundant Arrays of Inexpensive Disks (RAID)
Virtual Memory Lecture notes from MKP and S. Yalamanchili.
Multiple Platters.
Disks and RAID.
RAID Non-Redundant (RAID Level 0) has the lowest cost of any RAID
Vladimir Stojanovic & Nicholas Weaver
IT 251 Computer Organization and Architecture
CS 554: Advanced Database System Notes 02: Hardware
Part V Memory System Design
Disks and Files DBMS stores information on (“hard”) disks.
Introduction I/O devices can be characterized by I/O bus connections
Lecture 13 I/O.
Sanuja Dabade & Eilbroun Benjamin CS 257 – Dr. TY Lin
Lecture 21: Storage Systems
ICOM 6005 – Database Management Systems Design
Input-output I/O is very much architecture/system dependent
Lecture 28: Reliability Today’s topics: GPU wrap-up Disk basics RAID
Mark Zbikowski and Gary Kimura
Persistence: hard disk drive
Mass-Storage Systems.
Persistence: I/O devices
CS 245: Database System Principles Notes 02: Hardware
Presentation transcript:

Module: Storage Systems ECE 4100/6100 Fall 2004

Reading Storage technologies and trends Section 7.2 http://www.storagereview.com/map/lm.cgi/areal_density Interfacing storage devices to the CPU Section 7.3 Redundant arrays of inexpensive disks (RAID) Section 7.5 A very good reference for an introductory explanation of the basics a of RAID systems http://www.ecs.umass.edu/ece/koren/architecture/Raid/raidhome.html Fall 2004

Motivation: Who Cares About I/O? CPU Performance has been doubling every 18 months I/O system performance grows much more slowly limited by speed of mechanical components Disk seek speeds have improved on the order of 10%/yr Amdahl's Law: system speed-up limited by the slowest part! Time spent in I/O determines application speedup Speedup limited by 1/s, where s is fraction spent in I/O Faster CPUs do not lead to corresponding reductions in execution time Ancestor of Java had no I/O CPU vs. Peripheral Primary vs. Secondary What maks portable, PDA exciting? Fall 2004

The I/O Subsystem Main Memory Fall 2004 Processor Memory Bus I/O Bus L1 Cache L2 Cache Processor Memory Bus Bridge I/O Bus I/O Controller Graphics Network Interface Fall 2004

Storage Technology Drivers Continued migration of data into electronic form Businesses Financial, operational, matrketing Personal Consumer electronics, home computing Dependability More important than more performance Cannot lose data! Fall 2004

Price/Capacity Rate of price decline accelerates in the recent past Disk capacity increase of 4000 in 18 years Smaller disks playing a bigger role More efficient to spin smaller mass, smaller diameter disks Fall 2004

Price/Gigabyte Price per gigabyte has dropped by a factor of over 10,000 since 1983! Making personal, storage applications affordable! Media technologies Fall 2004

Cost/Access Time Several orders of magnitude gap in cost and access time between disk and semiconductor memory Performance gain of semiconductor memory relative costs a factor of 100 more per gigabyte Fall 2004

Disk Drive Terminology Platter Head Arm Actuator Sector Platters Track Data is recorded on concentric tracks on both sides of a platter Tracks are organized as fixed size (bytes) sectors Corresponding tracks on all platters form a cylinder Data is addressed by three coordinates: cylinder, platter, and sector Actuator moves (seek) the correct read/write head over the correct sector Under the control of the controller Fall 2004

Disk Performance Head Arm Actuator Platters Disk latency = controller overhead + seek time + rotational delay + transfer delay Seek time and rotational delay are limited by mechanical parts Fall 2004

Disk Performance Seek time determined by the current position of the head, i.e., what track is it covering, and the new position of the head Average rotational delay is time for 0.5 revolutions Transfer rate is a function of bit density Fall 2004

Areal Density Density in a track is measured in bits per inch (BPI) Track density is measured as tracks per inch (TPI) Areal density is bit density per unit area and is BPI x TPI Check out http://www.storagereview.com/map/lm.cgi/areal_density Fall 2004

Physical Layout Older designs had the same number of sectors on all tracks Outer tracks had lower bit density Constant bit density recording Outer tracks have more sectors than the inner tracks Bit density is not quite constant with inner tracks having higher recording density Fall 2004

Disk Performance Model /Trends Capacity More recently growth accelerated to 100%/year Transfer rate 40%/yr Rotation + Seek time Closer to 10%/yr MB/$ See Figure 7.4 Fall 2004

State of the Art: Barracuda 7200.7 200 GB, 3.5 inch disk 16,383 cylinders (def) 7,200 RPM 8.5 ms average seek time 100 Mbytes/sec max external transfer rate Recording density 671,500 bpi 7.5 watts (idle) – 12.5 (seek) Source: www.seagate.com 47 to 26 MB/s (external) Fall 2004

Microdrives Hitachi Microdrive 2004 2006 IBM (speculation)? 56.5 Gbits/sq in 3600 rpm 8.3 ms average seek 4.2 – 7.3 Mbits/sec data rate 2006 IBM (speculation)? 6 GB for 1 in drives 100 Gbits/sq in http://news.com.com/2100-1040275494.html?legacy=cnet Fall 2004

Disk Drive Performance Published seek times are based on uniformly distributed requests Reality is that layout is optimized particularly when large data sets are involved Published data rates include reading error correction bits and therefore “actual” data rates are lower Fall 2004

Improving Disk I/O Performance Single disk performance is limited by the drive mechanics Multiple disks operating together can offer a single, faster, larger, logical disk that can offer Higher aggregate data transfer rates on large data sets Higher aggregate I/O request service rate on small accesses that are distributed across the disks Fall 2004

Arrays of Inexpensive Disks: Throughput CPU read request Block 0 Block 1 Block 2 Block 3 Data is striped across all disks Visible performance overhead of drive mechanics is amortized across multiple accesses Scientific workloads are well suited to such oranizations Fall 2004

Arrays of Inexpensive Disks: Request Rate Multiple CPU read requests Consider multiple read requests for small blocks of data Several I/O requests can be serviced concurrently Fall 2004

Reliability of Disk Arrays Redundant information The reliability of an array of N disks is lower than the reliability of a single disk Any single disk failure will cause the array to fail The array is N times more likely to fail Use redundant disks to recover from failures Similar to use of error correcting codes Overhead Bandwidth and cost Fall 2004

Redundant Arrays of Small Disks (RAID) What size disks should we use? Smaller disks have the advantages of Lower cost/MB Higher MB/volume Higher MB/watt Classic paper “Disk System Architectures for High Performance Computing,” R. Katz, G. Gibson, D. Patterson, Proceedings of the IEEE, vol. 77, no. 12, December 1989 Fall 2004

RAID Level 0 RAID 0 corresponds to use of striping with no redundancy 1 2 3 4 5 6 7 RAID 0 corresponds to use of striping with no redundancy Provides the highest performance Provides the lowest reliability Frequently used in scientific and supercomputing applications where data throughput is important Fall 2004

RAID Level 1 mirrors The disk array is “mirrored” or “shadowed” in its entirety Reads can be optimized Pick the array with smaller queuing and seek times Performance sacrifice on writes – to both arrays Fall 2004

RAID Level 10 Use a combination of mirroring and striping mirrors 1 2 3 4 5 6 7 mirrors 1 2 3 4 5 6 7 Use a combination of mirroring and striping Fall 2004

RAID Level 3 Bit level parity 1 1 Parity Disk RAID 2 evolved to bit-level striping and use of error correcting codes Key observation Failed disk can always be recognized Gave way to RAID level 3 with the use of parity The contents of the failed disk can be reconstructed Fall 2004

RAID Level 4 Block level parity Block 0 Block 1 Block 2 Block 3 Parity Block 4 Block 5 Block 6 Block 7 Parity Parity Disk Data is interleaved in blocks, referred to as the striping unit and striping width Small reads can access subset of the disks A write to a single disk requires 4 accesses read old block, write new block, read and write parity disk Parity disk can become a bottleneck Fall 2004

The Small Write Problem 4 1 B1-New Ex-OR 2 Ex-OR 3 Two disk read operations followed by two disk write operations Fall 2004

RAID 4 Logical View Small reads can take place concurrently B0 B1 B2 B2 parity B4 B5 B6 B7 parity B8 B9 B10 B11 parity B12 B13 B14 B15 parity Small reads can take place concurrently Large reads/writes use multiple disks and (for writes) the parity disk Small writes take four accesses Parity disk is a bottleneck Fall 2004

RAID 5 Logical View B0 B1 B2 B2 parity B4 B5 B6 parity B7 B8 B9 parity B11 B10 B12 parity B14 B15 B13 Block interleaved distributed parity organization distributes write traffic amongst disks Best combination of small read and small write performance Fall 2004

Summary Storage: Systems Extraordinary advances in the densities and packaging Raw performance limited by drive mechanics System level organizations such as RAID evolved to continue the growth in performance Fall 2004