Bridging the Information Gap in Storage Protocol Stacks Timothy E. Denehy, Andrea C. Arpaci-Dusseau, and Remzi H. Arpaci-Dusseau University of Wisconsin,

Slides:



Advertisements
Similar presentations
A Case for Redundant Arrays Of Inexpensive Disks Paper By David A Patterson Garth Gibson Randy H Katz University of California Berkeley.
Advertisements

Redundant Array of Independent Disks (RAID) Striping of data across multiple media for expansion, performance and reliability.
A CASE FOR REDUNDANT ARRAYS OF INEXPENSIVE DISKS (RAID) D. A. Patterson, G. A. Gibson, R. H. Katz University of California, Berkeley.
By Rashid Khan Lesson 6-A Place for Everything: Storage Management.
More on File Management
The Zebra Striped Network Filesystem. Approach Increase throughput, reliability by striping file data across multiple servers Data from each client is.
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
The Zebra Striped Network File System Presentation by Joseph Thompson.
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
1 Magnetic Disks 1956: IBM (RAMAC) first disk drive 5 Mb – Mb/in $/year 9 Kb/sec 1980: SEAGATE first 5.25’’ disk drive 5 Mb – 1.96 Mb/in2 625.
Chapter 5: Server Hardware and Availability. Hardware Reliability and LAN The more reliable a component, the more expensive it is. Server hardware is.
RAID Technology CS350 Computer Organization Section 2 Larkin Young Rob Deaderick Amos Painter Josh Ellis.
Log-Structured Memory for DRAM-Based Storage Stephen Rumble, Ankita Kejriwal, and John Ousterhout Stanford University.
G Robert Grimm New York University Sprite LFS or Let’s Log Everything.
1 File Management in Representative Operating Systems.
Cse Feb-001 CSE 451 Section February 24, 2000 Project 3 – VM.
RAID-x: A New Distributed Disk Array for I/O-Centric Cluster Computing Kai Hwang, Hai Jin, and Roy Ho.
Storage System: RAID Questions answered in this lecture: What is RAID? How does one trade-off between: performance, capacity, and reliability? What is.
Frangipani: A Scalable Distributed File System C. A. Thekkath, T. Mann, and E. K. Lee Systems Research Center Digital Equipment Corporation.
Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung Google∗
PETAL: DISTRIBUTED VIRTUAL DISKS E. K. Lee C. A. Thekkath DEC SRC.
ICOM 6005 – Database Management Systems Design Dr. Manuel Rodríguez-Martínez Electrical and Computer Engineering Department Lecture 6 – RAID ©Manuel Rodriguez.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
1 The Google File System Reporter: You-Wei Zhang.
Page 19/4/2015 CSE 30341: Operating Systems Principles Raid storage  Raid – 0: Striping  Good I/O performance if spread across disks (equivalent to n.
CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Storage Systems.
Lecture 9 of Advanced Databases Storage and File Structure (Part II) Instructor: Mr.Ahmed Al Astal.
Parity Logging O vercoming the Small Write Problem in Redundant Disk Arrays Daniel Stodolsky Garth Gibson Mark Holland.
Buffalo Data Storage Expansion June As organizations grow the amount of data storage capacity required to support it grows as well Increased data.
CS 153 Design of Operating Systems Spring 2015 Final Review.
Topics covered: Memory subsystem CSE243: Introduction to Computer Architecture and Hardware/Software Interface.
Module 9: Configuring Storage
MCTS Guide to Microsoft Windows Vista Chapter 4 Managing Disks.
Data Tagging Architecture for System Monitoring in Dynamic Environments Bharat Krishnamurthy, Anindya Neogi, Bikram Sengupta, Raghavendra Singh (IBM Research.
X-RAY: A Non-Invasive Exclusive Caching Mechanism for RAIDs Lakshmi N. Bairavasundaram Muthian Sivathanu Andrea C. Arpaci-Dusseau Remzi H. Arpaci-Dusseau.
Journal-guided Resynchronization for Software RAID
The Design and Implementation of Log-Structure File System M. Rosenblum and J. Ousterhout.
26-Oct-15CSE 542: Operating Systems1 File system trace papers The Design and Implementation of a Log- Structured File System. M. Rosenblum, and J.K. Ousterhout.
Guide to MCSE , Second Edition, Enhanced 1 File Storage Basics Basic storage Centers on partitioning physical disk Dynamic storage New method supported.
Serverless Network File Systems Overview by Joseph Thompson.
Deconstructing Storage Arrays Timothy E. Denehy, John Bent, Florentina I. Popovici, Andrea C. Arpaci-Dusseau, Remzi H. Arpaci-Dusseau University of Wisconsin,
Efficient RDF Storage and Retrieval in Jena2 Written by: Kevin Wilkinson, Craig Sayers, Harumi Kuno, Dave Reynolds Presented by: Umer Fareed 파리드.
CS 153 Design of Operating Systems Spring 2015 Lecture 22: File system optimizations.
Advanced UNIX File Systems Berkley Fast File System, Logging File Systems And RAID.
CS 153 Design of Operating Systems Spring 2015 Lecture 21: File Systems.
Semantically-Smart Disk Systems Muthian Sivathanu, Vijayan Prabhakaran, Florentina Popovici, Tim Denehy, Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau University.
Storage Research Meets The Grid Remzi Arpaci-Dusseau.
File Structures. 2 Chapter - Objectives Disk Storage Devices Files of Records Operations on Files Unordered Files Ordered Files Hashed Files Dynamic and.
5 Copyright © 2005, Oracle. All rights reserved. Managing Database Storage Structures.
The concept of RAID in Databases By Junaid Ali Siddiqui.
1 © 2002 hp Introduction to EVA Keith Parris Systems/Software Engineer HP Services Multivendor Systems Engineering Budapest, Hungary 23May 2003 Presentation.
NeST: Network Storage John Bent, Venkateshwaran V Miron Livny, Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau.
GPFS: A Shared-Disk File System for Large Computing Clusters Frank Schmuck & Roger Haskin IBM Almaden Research Center.
Chapter 7 Memory Management Eighth Edition William Stallings Operating Systems: Internals and Design Principles.
Hands-On Microsoft Windows Server 2008 Chapter 7 Configuring and Managing Data Storage.
Transactional Flash V. Prabhakaran, T. L. Rodeheffer, L. Zhou (MSR, Silicon Valley), OSDI 2008 Shimin Chen Big Data Reading Group.
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
W4118 Operating Systems Instructor: Junfeng Yang.
RAID Technology By: Adarsha A,S 1BY08A03. Overview What is RAID Technology? What is RAID Technology? History of RAID History of RAID Techniques/Methods.
Log-Structured Memory for DRAM-Based Storage Stephen Rumble and John Ousterhout Stanford University.
Chapter 2 Memory and process management
Introduction to Operating Systems
ICOM 6005 – Database Management Systems Design
Module IV Memory Organization.
Bridging the Information Gap in Storage Protocol Stacks
UNIT IV RAID.
CSE 451: Operating Systems Autumn 2004 Redundant Arrays of Inexpensive Disks (RAID) Hank Levy 1.
CSE 451: Operating Systems Winter 2004 Module 17 Redundant Arrays of Inexpensive Disks (RAID) Ed Lazowska Allen Center 570.
Supporting Online Analytics with User-Defined Estimation and Early Termination in a MapReduce-Like Framework Yi Wang, Linchuan Chen, Gagan Agrawal The.
Presentation transcript:

Bridging the Information Gap in Storage Protocol Stacks Timothy E. Denehy, Andrea C. Arpaci-Dusseau, and Remzi H. Arpaci-Dusseau University of Wisconsin, Madison

2 of 31 State of Affairs Namespace Files Metadata Layout Liveness Parallelism Redundancy File System Storage System

3 of 31 Information gap may cause problems –Poor performance Partial stripe write operations –Duplicated functionality Logging in file system and storage system –Reduced functionality Storage system lacks knowledge of files Time to re-examine the division of labor Problem

4 of 31 Enhance the storage interface –Expose performance and failure information Use information to provide new functionality –On-line expansion –Dynamic parallelism –Flexible redundancy Our Approach Informed LFS Exposed RAID

5 of 31 Outline E  RAID Overview I·LFS Overview Functionality and Evaluation Conclusion

6 of 31 Goals –Backwards compatibility Block-based interface Linear, concatenated address space –Expose information to the file system above Allows file system to utilize semantic knowledge E  RAID Overview

7 of 31 Region –Contiguous portion of the address space Regions can be added to expand the address space Region composition –RAID: one region for all disks –Exposed: separate regions for each disk –Hybrid E  RAID Regions E  RAID

8 of 31 Exposed on a per-region basis Throughput and queue length Reveals Static disk heterogeneity Dynamic performance and load fluctuations E  RAID Performance Information E  RAID

9 of 31 Exposed on a per-region basis Number of tolerable failures Regions may have different failure characteristics Reveals dynamic failures to file system above E  RAID Failure Information RAID1 E  RAID X

10 of 31 Outline E  RAID Overview I·LFS Overview Functionality and Evaluation Conclusion

11 of 31 Modified NetBSD LFS –All data and metadata is written to a log –Log is a collection of segments –Segment table describes each segment –Cleaner process produces empty segments I·LFS Overview

12 of 31 Goals –Improve performance, functionality, and manageability –Minimize system complexity Exploits E  RAID information to provide –On-line expansion –Dynamic parallelism –Flexible redundancy –Lazy redundancy I·LFS Overview

13 of 31 NetBSD GHz Intel Pentium III Xeon 128 MB RAM Four fast disks –Seagate Cheetah 36XL, 21.6 MB/s Four slow disks –Seagate Barracuda 4XL, 7.5 MB/s I·LFS Experimental Platform

14 of 31 I·LFS Baseline Performance

15 of 31 Goal: expand storage incrementally –Capacity –Performance Ideal: instant disk addition –Minimize downtime –Simplify administration I·LFS supports on-line addition of new disks I·LFS On-line Expansion

16 of 31 E  RAID: an expandable address space Expansion is equivalent to adding empty segments Start with an oversized segment table Activate new portion of segment table I·LFS On-line Expansion Details

17 of 31 I·LFS On-line Expansion Experiment I·LFS takes immediate advantage of each extra disk

18 of 31 Goal: perform well on heterogeneous storage –Static performance differences –Dynamic performance fluctuations Ideal: maximize throughput of the storage system I·LFS writes data proportionate to performance I·LFS Dynamic Parallelism

19 of 31 E  RAID: dynamic performance information Most file system routines are not changed –Aware of only the E  RAID linear address space Segment selection routine –Aware of E  RAID regions and performance –Chooses next segment based on current performance Minimizes changes to the file system I·LFS Dynamic Parallelism Details

20 of 31 I·LFS Static Parallelism Experiment I·LFS provides the full throughput of the system Simple striping runs at the rate of the slowest disk

21 of 31 I·LFS Dynamic Parallelism Experiment I·LFS adjusts to the performance fluctuation

22 of 31 Goal: offer new redundancy options to users Ideal: range of redundancy mechanisms and granularities I·LFS provides mirrored per-file redundancy I·LFS Flexible Redundancy

23 of 31 E  RAID: region failure characteristics Use separate files for redundancy –Even inode N for original files –Odd inode N+1 for redundant files –Original and redundant data in different sets of regions Flexible data placement within the regions Use recursive vnode operations for redundant files –Leverage existing routines to reduce complexity I·LFS Flexible Redundancy Details

24 of 31 I·LFS Flexible Redundancy Experiment I·LFS provides a throughput and reliability tradeoff

25 of 31 Goal: avoid replication performance penalty Ideal: replicate data immediately before failure I·LFS offers redundancy with delayed replication Avoids penalty for redundant, short-lived files I·LFS Lazy Redundancy

26 of 31 E  RAID: region failure characteristics Segments needing replication are flagged Cleaner acts as replicator –Locates flagged segments –Checks data liveness and lifetime –Generates redundant copies of files I·LFS Lazy Redundancy

27 of 31 I·LFS Lazy Redundancy Experiment I·LFS avoids performance penalty for short-lived files

28 of 31 Outline E  RAID Overview I·LFS Overview Functionality and Evaluation Conclusion

29 of 31 Comparison with Traditional Systems On-line expansion –Yes, but capacity only, not performance Dynamic parallelism –Yes, but with duplicated functionality Flexible redundancy –No, the storage system is not aware of file composition Lazy redundancy –No, the storage system is not aware of file deletions

30 of 31 Conclusion Introduced E  RAID and I·LFS Extra information enables new functionality –Difficult or impossible in traditional systems Minimal complexity –19% increase in code size Time to re-examine the division of labor

31 of 31 Questions? Full paper available on the WiND publications page –

32 of 31 Extra Slides

33 of 31 Storage Failure

34 of 31 Crossed-pointer Problem