PARAID: A Gear-Shifting Power-Aware RAID

Slides:



Advertisements
Similar presentations
Conserving Disk Energy in Network Servers ACM 17th annual international conference on Supercomputing Presented by Hsu Hao Chen.
Advertisements

1 Sizing the Streaming Media Cluster Solution for a Given Workload Lucy Cherkasova and Wenting Tang HPLabs.
Energy Efficiency through Burstiness Athanasios E. Papathanasiou and Michael L. Scott University of Rochester, Computer Science Department Rochester, NY.
1 Storage-Aware Caching: Revisiting Caching for Heterogeneous Systems Brian Forney Andrea Arpaci-Dusseau Remzi Arpaci-Dusseau Wisconsin Network Disks University.
Walter Binder University of Lugano, Switzerland Niranjan Suri IHMC, Florida, USA Green Computing: Energy Consumption Optimized Service Hosting.
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
1 Conserving Energy in RAID Systems with Conventional Disks Dong Li, Jun Wang Dept. of Computer Science & Engineering University of Nebraska-Lincoln Peter.
Ceph: A Scalable, High-Performance Distributed File System Priya Bhat, Yonggang Liu, Jing Qin.
RIMAC: Redundancy-based hierarchical I/O cache architecture for energy-efficient, high- performance storage systems Xiaoyu Yao and Jun Wang Computer Architecture.
Shimin Chen Big Data Reading Group Presented and modified by Randall Parabicoli.
Copyright 2009 FUJITSU TECHNOLOGY SOLUTIONS PRIMERGY Servers and Windows Server® 2008 R2 Benefit from an efficient, high performance and flexible platform.
Web Caching Schemes1 A Survey of Web Caching Schemes for the Internet Jia Wang.
Energy Efficient Prefetching – from models to Implementation 6/19/ Adam Manzanares and Xiao Qin Department of Computer Science and Software Engineering.
Energy Efficient Prefetching with Buffer Disks for Cluster File Systems 6/19/ Adam Manzanares and Xiao Qin Department of Computer Science and Software.
Everest: scaling down peak loads through I/O off-loading D. Narayanan, A. Donnelly, E. Thereska, S. Elnikety, A. Rowstron Microsoft Research Cambridge,
1 Exploring Data Reliability Tradeoffs in Replicated Storage Systems NetSysLab The University of British Columbia Abdullah Gharaibeh Matei Ripeanu.
THE DESIGN AND IMPLEMENTATION OF A LOG-STRUCTURED FILE SYSTEM M. Rosenblum and J. K. Ousterhout University of California, Berkeley.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Introduction to Database Systems 1 The Storage Hierarchy and Magnetic Disks Storage Technology: Topic 1.
THE HP AUTORAID HIERARCHICAL STORAGE SYSTEM J. Wilkes, R. Golding, C. Staelin T. Sullivan HP Laboratories, Palo Alto, CA.
Comparing Coordinated Garbage Collection Algorithms for Arrays of Solid-state Drives Junghee Lee, Youngjae Kim, Sarp Oral, Galen M. Shipman, David A. Dillow,
RAID-x: A New Distributed Disk Array for I/O-Centric Cluster Computing Kai Hwang, Hai Jin, and Roy Ho.
Energy Efficiency and Storage Flexibility in the Blue File System Edmund B Nightingale Jason Flinn University of Michigan.
CS 423 – Operating Systems Design Lecture 22 – Power Management Klara Nahrstedt and Raoul Rivas Spring 2013 CS Spring 2013.
1 Exploring Data Reliability Tradeoffs in Replicated Storage Systems NetSysLab The University of British Columbia Abdullah Gharaibeh Advisor: Professor.
Reliability Analysis of An Energy-Aware RAID System Shu Yin Xiao Qin Auburn University.
Cloud Data Center/Storage Power Efficiency Solutions Junyao Zhang 1.
1 Storage Refinement. Outline Disk failures To attack Intermittent failures To attack Media Decay and Write failure –Checksum To attack Disk crash –RAID.
PARAID: The Gear-Shifting Power-Aware RAID Charles Weddle, Mathew Oldham, An-I Andy Wang – Florida State University RuGang Xu, Peter Reiher – University.
PARAID: The Gear-Shifting Power-Aware RAID Charles Weddle, Mathew Oldham, An-I Andy Wang – Florida State University Peter Reiher – University of California,
Exploiting Flash for Energy Efficient Disk Arrays Shimin Chen (Intel Labs) Panos K. Chrysanthis (University of Pittsburgh) Alexandros Labrinidis (University.
1 An SLA-Oriented Capacity Planning Tool for Streaming Media Services Lucy Cherkasova, Wenting Tang, and Sharad Singhal HPLabs,USA.
Introduction and Overview Questions answered in this lecture: What is an operating system? How have operating systems evolved? Why study operating systems?
Tag line, tag line Power Management in Storage Systems Kaladhar Voruganti Technical Director CTO Office, Sunnyvale June 12, 2009.
A Measurement Based Memory Performance Evaluation of High Throughput Servers Garba Isa Yau Department of Computer Engineering King Fahd University of Petroleum.
1 PARAID: A Gear-Shifting Power-Aware RAID Charles Weddle, Mathew Oldham, Jin Qian, An-I Andy Wang – Florida St. University Peter Reiher – University of.
PARAID: A Gear-Shifting Power-Aware RAID Charles Weddle, Mathew Oldham, Jin Qian, An-I Andy Wang – Florida St. University Peter Reiher – University of.
Data Replication and Power Consumption in Data Grids Susan V. Vrbsky, Ming Lei, Karl Smith and Jeff Byrd Department of Computer Science The University.
Towards Dynamic Green-Sizing for Database Servers Mustafa Korkmaz, Alexey Karyakin, Martin Karsten, Kenneth Salem University of Waterloo.
Example: Rumor Performance Evaluation Andy Wang CIS 5930 Computer Systems Performance Analysis.
Ceph: A Scalable, High-Performance Distributed File System
Conquest-2: Improving Energy Efficiency and Performance Through a Disk/RAM Hybrid File System An-I Andy Wang Florida State University (NSF CCR ,
Embedded System Lab. 정범종 A_DRM: Architecture-aware Distributed Resource Management of Virtualized Clusters H. Wang et al. VEE, 2015.
Improving Disk Throughput in Data-Intensive Servers Enrique V. Carrera and Ricardo Bianchini Department of Computer Science Rutgers University.
1 PARAID: A Gear-Shifting Power-Aware RAID Charles Weddle, Mathew Oldham, Jin Qian, An-I Andy Wang – Florida St. University Peter Reiher – University of.
File management and Performance. File Systems Architecture device drivers physical I/O (PIOCS) logical I/O (LIOCS) access methods File organization and.
Best Available Technologies: External Storage Overview of Opportunities and Impacts November 18, 2015.
Ensieea Rizwani An energy-efficient management mechanism for large-scale server clusters By: Zhenghua Xue, Dong, Ma, Fan, Mei 1.
Department of Computer Sciences, University of Wisconsin Madison DADA – Dynamic Allocation of Disk Area Jayaram Bobba Vivek Shrivastava.
GPFS: A Shared-Disk File System for Large Computing Clusters Frank Schmuck & Roger Haskin IBM Almaden Research Center.
Maximizing Performance – Why is the disk subsystem crucial to console performance and what’s the best disk configuration. Extending Performance – How.
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
By Harshal Ghule Guided by Mrs. Anita Mahajan G.H.Raisoni Institute Of Engineering And Technology.
1 PARAID: A Gear-Shifting Power-Aware RAID Charles Weddle, Mathew Oldham, Jin Qian, An-I Andy Wang – Florida St. University Peter Reiher – University of.
Motivation Energy costs are rising –An increasing concern for servers –No longer limited to laptops Energy consumption of disk drives –24% of the power.
Input and Output Optimization in Linux for Appropriate Resource Allocation and Management James Avery King.
Energy Aware Network Operations
PARAID: A Gear-Shifting Power-Aware RAID
HP AutoRAID (Lecture 5, cs262a)
Green cloud computing 2 Cs 595 Lecture 15.
Optimizing the Migration of Virtual Computers
PA an Coordinated Memory Caching for Parallel Jobs
The Composite-File File System: Decoupling the One-to-one Mapping of Files and Metadata for Better Performance Shuanglong Zhang, Helen Catanese, Andy An-I.
Energy Efficiency and Storage Flexibility in the Blue File System
THE HP AUTORAID HIERARCHICAL STORAGE SYSTEM
Overview Continuation from Monday (File system implementation)
Qingbo Zhu, Asim Shankar and Yuanyuan Zhou
Energy Efficiency and Storage Flexibility in the Blue File System
Specialized Cloud Architectures
Dong Hyun Kang, Changwoo Min, Young Ik Eom
Presentation transcript:

PARAID: A Gear-Shifting Power-Aware RAID Charles Weddle, Mathew Oldham, Jin Qian, An-I Andy Wang – Florida St. University Peter Reiher – University of California, Los Angeles Geoff Kuenning – Harvey Mudd College

Motivation Energy costs are rising Energy consumption of disk drives An increasing concern for servers No longer limited to laptops Energy consumption of disk drives 24% of the power usage in web servers 27% of electricity cost for data centers More energy  more heat  more cooling  lower computational density  more space  more costs Is it possible to reduce energy consumption without degrading performance while maintaining reliability? PARAID: A Gear-Shifting Power-Aware RAID

Challenges Energy Performance Reliability Not enough opportunities to spin down RAIDs Performance Essential for peak loads Reliability Server-class drives are not designed for frequent power switching PARAID: A Gear-Shifting Power-Aware RAID

Existing Work Most trade performance for energy savings directly e.g. vary speed of disks Most are simulated results PARAID: A Gear-Shifting Power-Aware RAID

Observations RAID is configured for peak performance RAID keeps all drives spinning for light loads Unused storage capacity Over-provision of storage capacity Unused storage can be traded for energy savings Fluctuating load Cyclic fluctuation of loads Infrequent on-off power transitions can be effective PARAID: A Gear-Shifting Power-Aware RAID

Performance vs. Energy Optimizations Performance benefits Realized under heavy loads Energy benefits Realized instantaneously

Power-Aware RAID Skewed striping for energy savings Preserving peak performance Maintaining reliability Evaluation Conclusion Questions PARAID: A Gear-Shifting Power-Aware RAID

Skewed Striping for Energy Saving Use over-provisioned spare storage Organized into hierarchical overlapping subsets 1 2 3 4 5 RAID PARAID: A Gear-Shifting Power-Aware RAID

Skewed Striping for Energy Saving Each set analogous to gears in automobiles 1 2 3 4 5 RAID Gears 1 2 3 PARAID: A Gear-Shifting Power-Aware RAID

Skewed Striping for Energy Saving Soft states can be reclaimed for space Persist across reboots 1 2 3 4 5 Soft States RAID Gears 1 2 3 PARAID: A Gear-Shifting Power-Aware RAID

Skewed Striping for Energy Saving Operate in gear 1 Disks 4 and 5 are powered off 1 2 3 4 5 Soft States RAID Gears 1 2 3 PARAID: A Gear-Shifting Power-Aware RAID

Skewed Striping for Energy Saving Approximate the workload Gear shift into most appropriate gear Minimize the opportunity lost to save power Conventional RAID PARAID Energy ( Powered On Disks ) workload Workload ( Disk Parallelism ) PARAID: A Gear-Shifting Power-Aware RAID

Skewed Striping for Energy Saving Adapt to cyclic fluctuating workload Gear shift when gear utilization threshold is met time load utilization threshold gear shift PARAID: A Gear-Shifting Power-Aware RAID

Preserving Peak Performance Operate in the highest gear When the system demands peak performance Uses the same disk layout Maximize parallelism within each gear Load is balanced Uniform striping pattern Delay block replication until gear shifts Capture block writes PARAID: A Gear-Shifting Power-Aware RAID

Maintaining Reliability Reuse existing RAID levels (RAID-5) Also used in various gears Drives have a limited number of power cycles Ration number of power cycles PARAID: A Gear-Shifting Power-Aware RAID

Maintaining Reliability Busy disk stay powered on, idle disks stay powered off Outside disks are role exchanged with middle disks busy disks power cycled idle role exchange Disk 1 Gear 1 Gear 2 Gear 3 Disk 2 Disk 3 Disk 4 Disk 5 Disk 6 PARAID: A Gear-Shifting Power-Aware RAID

Logical Component Design File system RAID PARAID block mapping Disk device driver User space Linux kernel Soft RAID Reliability manager Load monitor Gear manager Admin tool PARAID: A Gear-Shifting Power-Aware RAID

Data Layout Resembles the data flow of RAID 1+0 Parity for 5 disks does not work for 4 disks For example, replicated block 12 on disk 3 Disk 1 Disk 2 Disk 3 Disk 4 Disk 5 Gear 1 RAID-5 (1-4) 8 12 ((1-4),8,12) 16 20 (16,20,_) _ Gear 2 1 2 3 4 5 6 7 (5-8) 9 10 (9-12) 11 13 (13-16) 14 15 (17-20) 17 18 19 PARAID: A Gear-Shifting Power-Aware RAID

Data Layout Cascading parity updates For example, updating block 8 on disk 5 Disk 1 Disk 2 Disk 3 Disk 4 Disk 5 Gear 1 RAID-5 (1-4) 8 12 ((1-4),8,12) 16 20 (16,20,_) _ Gear 2 1 2 3 4 5 6 7 (5-8) 9 10 (9-12) 11 13 (13-16) 14 15 (17-20) 17 18 19 PARAID: A Gear-Shifting Power-Aware RAID

Update Propagation Up-shift propagation (e.g. shifting from 3 to 5 disks) Full synchronization On-demand synchronization Need to respect block dependency Downshift propagation PARAID: A Gear-Shifting Power-Aware RAID

Asymmetric Gear-Shifting Policies Up-shift (aggressive) Moving utilization average + moving standard deviation > utilization threshold Downshift (conservative) Modified utilization moving average + moving standard deviation < utilization threshold Moving average modified to account for fewer drives and extra parity updates PARAID: A Gear-Shifting Power-Aware RAID

Implementation Prototyped in Linux 2.6.5 Open source, software RAID Implemented block I/O handler, monitor, disk manager Implemented user admin tool to configure device Updated Raid Tools to recognize PARAID level PARAID: A Gear-Shifting Power-Aware RAID

Evaluation Challenges Prototyping PARAID Commercial machines Conceptual barriers Benchmarks designed to measure peak performance Trace replay Time consuming PARAID: A Gear-Shifting Power-Aware RAID

Evaluation Measurement framework multimeter USB cable client server power supply 12v & 5v power lines measurement probes SCSI cable crossover cable Xeon 2.8 Ghz, 512 MB RAM 36.7 GB 15k RPM SCSI P4 2.8 Ghz, 1 GB RAM 160 GB 7200 RPM SATA RAID BOOT PARAID: A Gear-Shifting Power-Aware RAID

Evaluation Three different workloads using two different RAID settings Web trace - RAID level 0 (2-disk gear 1, 5-disk gear 2) Mostly read activity Cello99 - RAID level 5 (3-disk gear 1, 5-disk gear 2) I/O-intensive workload with writes PostMark - RAID level 5 Measure peak performance and gear shifting overhead Speed up trace playback To match hardware Explore range of speed up factors and power savings PARAID: A Gear-Shifting Power-Aware RAID

Web Trace UCLA CS Dept Web Servers (8/11/2006 – 8/14/2006) File system: ~32 GB (~500k files) Trace replay: ~95k requests with ~4 GB data (~260 MB unique) PARAID: A Gear-Shifting Power-Aware RAID

Web Trace Power Savings 64x – 60 requests/sec Energy Savings 64x - 34% 128x - 28% 256x - 10% 128x – 120 requests/sec 256x – 240 requests/sec PARAID: A Gear-Shifting Power-Aware RAID

Web Trace Latency Overhead 256x 256x - within 2.7% 64x - 240% 80ms vs. 33ms 128x 64x PARAID: A Gear-Shifting Power-Aware RAID

Web Trace Bandwidth Overhead 256x 256x - within 1.3% in high gear 128x PARAID: A Gear-Shifting Power-Aware RAID

Cello99 Trace Cello99 Workload HP Storage Research Labs 50 hours beginning on 9/12/1999 1.5 million requests (12 GB) to 440MB of unique blocks I/O-intensive with 42% writes PARAID: A Gear-Shifting Power-Aware RAID

Cello99 Power Savings Energy Savings 32x - 13% 64x - 8.2% 128x - 3.5% 32x – 270 requests/sec Energy Savings 32x - 13% 64x - 8.2% 128x - 3.5% 64x – 550 requests/sec 128x – 1000 requests/sec PARAID: A Gear-Shifting Power-Aware RAID

Cello99 Completion Time Overhead 32x - 1.8ms, 26% slower due to time spent in low gear 64x 32x PARAID: A Gear-Shifting Power-Aware RAID

Cello99 Bandwidth Overhead < 1% degra- dation during peak hours 128x Overhead < 1% degra- dation during peak hours 64x 32x PARAID: A Gear-Shifting Power-Aware RAID

PostMark Benchmark Popular synthetic benchmark Generates ISP-style workloads Stresses peak read/write performance of storage device PARAID: A Gear-Shifting Power-Aware RAID

Postmark Performance PARAID: A Gear-Shifting Power-Aware RAID

Postmark Power Savings PARAID: A Gear-Shifting Power-Aware RAID

Related Work Pergamum EERAID RIMAC Hibernator MAID PDC BlueFS Hibernator is an energy efficient RAID that shows up to 65% energy savings through the use of multi-speed disks MAID is an energy efficient RAID that migrates infrequently accessed data to rarely used disks. PDC is an energy efficient RAID that puts the most popular data on the first disk, second most popular data on the second disk, This is at the expense of parallelism. EERAID and RIMAC assume the use of a nonvolatile cache at the disk controller level and the knowledge of cache content to conserver energy in RAIDs. BlueFS is a distributed file system that uses a flexible cache hierarchy to decide when and where to access data based on the energy characteristics of each device. PARAID: A Gear-Shifting Power-Aware RAID

Ongoing Work Try more workloads Optimize PARAID gear configuration Explore asynchronous update propagation Speed up recovery Live testing PARAID: A Gear-Shifting Power-Aware RAID

Lessons Learned Third version of design, early design too complicated Data alignment problems Difficult to measure system under normal load Hard to predict workload transformations due to complex system optimizations Challenging to match trace environments PARAID: A Gear-Shifting Power-Aware RAID

Conclusion PARAID reuses standard RAID-levels without special hardware while decreasing their energy use by 34%. Optimized version can save even more energy Empirical evaluation important PARAID: A Gear-Shifting Power-Aware RAID

PARAID Recovery 2.7 times slower than conventional raid For example, 2 gear PARAID device First, the soft state must recover Second, data must be propagated Third, conventional raid must recover Recovery not as bad for read intensive workloads

Number of gear switches Number of gear switches PARAID Gear-Shifting Web Trace Gear-Shifting Stats 256x 128x 64x Number of gear switches 15.2 8.0 2.0 % time spent in low gear 52% 88% 98% % extra I/Os for update propagations 0.63% 0.37% 0.21% Cello99 Gear-Shifting Stats 128x 64x 32x Number of gear switches 6.0 5.6 5.4 % time spent in low gear 47% 74% 88% % extra I/Os for update propagations 8.0% 15% 21%