1 Storage-Aware Caching: Revisiting Caching for Heterogeneous Systems Brian Forney Andrea Arpaci-Dusseau Remzi Arpaci-Dusseau Wisconsin Network Disks University.

Slides:



Advertisements
Similar presentations
Dynamic Replica Placement for Scalable Content Delivery Yan Chen, Randy H. Katz, John D. Kubiatowicz {yanchen, randy, EECS Department.
Advertisements

Dissemination-based Data Delivery Using Broadcast Disks.
RAID Redundant Arrays of Independent Disks Courtesy of Satya, Fall 99.
Conserving Disk Energy in Network Servers ACM 17th annual international conference on Supercomputing Presented by Hsu Hao Chen.
HyLog: A High Performance Approach to Managing Disk Layout Wenguang Wang Yanping Zhao Rick Bunt Department of Computer Science University of Saskatchewan.
4.4 Page replacement algorithms
ULC: An Unified Placement and Replacement Protocol in Multi-level Storage Systems Song Jiang and Xiaodong Zhang College of William and Mary.
Full-System Timing-First Simulation Carl J. Mauer Mark D. Hill and David A. Wood Computer Sciences Department University of Wisconsin—Madison.
Application-Aware Memory Channel Partitioning † Sai Prashanth Muralidhara § Lavanya Subramanian † † Onur Mutlu † Mahmut Kandemir § ‡ Thomas Moscibroda.
1 A Hybrid Adaptive Feedback Based Prefetcher Santhosh Verma, David Koppelman and Lu Peng Louisiana State University.
Managing Wire Delay in Large CMP Caches Bradford M. Beckmann David A. Wood Multifacet Project University of Wisconsin-Madison MICRO /8/04.
1 Improving Direct-Mapped Cache Performance by the Addition of a Small Fully-Associative Cache and Prefetch Buffers By Sreemukha Kandlakunta Phani Shashank.
A File is Not a File: Understanding the I/O Behavior of Apple Desktop Applications Tyler Harter, Chris Dragga, Michael Vaughn, Andrea C. Arpaci-Dusseau,
1 Storage-Aware Caching: Revisiting Caching for Heterogeneous Storage Systems (Fast’02) Brian Forney Andrea Arpaci-Dusseau Remzi Arpaci-Dusseau Wisconsin.
Operating Systems ECE344 Ashvin Goel ECE University of Toronto Disks and RAID.
FlashVM: Virtual Memory Management on Flash Mohit Saxena and Michael M. Swift Introduction Flash storage is the largest change to memory and storage systems.
1 Web Server Performance in a WAN Environment Vincent W. Freeh Computer Science North Carolina State Vsevolod V. Panteleenko Computer Science & Engineering.
SCAN: A Dynamic, Scalable, and Efficient Content Distribution Network Yan Chen, Randy H. Katz, John D. Kubiatowicz {yanchen, randy,
Web Caching Schemes1 A Survey of Web Caching Schemes for the Internet Jia Wang.
Lecture 17 I/O Optimization. Disk Organization Tracks: concentric rings around disk surface Sectors: arc of track, minimum unit of transfer Cylinder:
Improving Proxy Cache Performance: Analysis of Three Replacement Policies Dilley, J.; Arlitt, M. A journal paper of IEEE Internet Computing, Volume: 3.
ISCSI Performance in Integrated LAN/SAN Environment Li Yin U.C. Berkeley.
Energy Efficient Prefetching – from models to Implementation 6/19/ Adam Manzanares and Xiao Qin Department of Computer Science and Software Engineering.
Lecture 3: A Case for RAID (Part 1) Prof. Shahram Ghandeharizadeh Computer Science Department University of Southern California.
Energy Efficient Prefetching with Buffer Disks for Cluster File Systems 6/19/ Adam Manzanares and Xiao Qin Department of Computer Science and Software.
Locality-Aware Request Distribution in Cluster-based Network Servers 1. Introduction and Motivation --- Why have this idea? 2. Strategies --- How to implement?
Differentiated Multimedia Web Services Using Quality Aware Transcoding S. Chandra, C.Schlatter Ellis and A.Vahdat InfoCom 2000, IEEE Journal on Selected.
Storage System: RAID Questions answered in this lecture: What is RAID? How does one trade-off between: performance, capacity, and reliability? What is.
Frangipani: A Scalable Distributed File System C. A. Thekkath, T. Mann, and E. K. Lee Systems Research Center Digital Equipment Corporation.
“Five minute rule ten years later and other computer storage rules of thumb” Authors: Jim Gray, Goetz Graefe Reviewed by: Nagapramod Mandagere Biplob Debnath.
Object-based Storage Long Liu Outline Why do we need object based storage? What is object based storage? How to take advantage of it? What's.
Redundant Array of Independent Disks
Toolbox for Dimensioning Windows Storage Systems Jalil Boukhobza, Claude Timsit 12/09/2006 Versailles Saint Quentin University.
SEDA: An Architecture for Well-Conditioned, Scalable Internet Services
PARAID: The Gear-Shifting Power-Aware RAID Charles Weddle, Mathew Oldham, An-I Andy Wang – Florida State University Peter Reiher – University of California,
An I/O Simulator for Windows Systems Jalil Boukhobza, Claude Timsit 27/10/2004 Versailles Saint Quentin University laboratory.
Swapping to Remote Memory over InfiniBand: An Approach using a High Performance Network Block Device Shuang LiangRanjit NoronhaDhabaleswar K. Panda IEEE.
Row Buffer Locality Aware Caching Policies for Hybrid Memories HanBin Yoon Justin Meza Rachata Ausavarungnirun Rachael Harding Onur Mutlu.
Exploiting Gray-Box Knowledge of Buffer Cache Management Nathan C. Burnett, John Bent, Andrea C. Arpaci-Dusseau, Remzi H. Arpaci-Dusseau University of.
Managing Distributed, Shared L2 Caches through OS-Level Page Allocation Jason Bosko March 5 th, 2008 Based on “Managing Distributed, Shared L2 Caches through.
1 Virtual Machine Memory Access Tracing With Hypervisor Exclusive Cache USENIX ‘07 Pin Lu & Kai Shen Department of Computer Science University of Rochester.
1 Admission Control and Request Scheduling in E-Commerce Web Sites Sameh Elnikety, EPFL Erich Nahum, IBM Watson John Tracey, IBM Watson Willy Zwaenepoel,
Latency Reduction Techniques for Remote Memory Access in ANEMONE Mark Lewandowski Department of Computer Science Florida State University.
Improving Disk Throughput in Data-Intensive Servers Enrique V. Carrera and Ricardo Bianchini Department of Computer Science Rutgers University.
Measuring the Capacity of a Web Server USENIX Sympo. on Internet Tech. and Sys. ‘ Koo-Min Ahn.
Introduction: Memory Management 2 Ideally programmers want memory that is large fast non volatile Memory hierarchy small amount of fast, expensive memory.
CSE 451: Operating Systems Spring 2012 Module 16 BSD UNIX Fast File System Ed Lazowska Allen Center 570.
1 Hidra: History Based Dynamic Resource Allocation For Server Clusters Jayanth Gummaraju 1 and Yoshio Turner 2 1 Stanford University, CA, USA 2 Hewlett-Packard.
MiddleMan: A Video Caching Proxy Server NOSSDAV 2000 Brian Smith Department of Computer Science Cornell University Ithaca, NY Soam Acharya Inktomi Corporation.
Transforming Policies into Mechanisms with Infokernel Andrea C. Arpaci-Dusseau, Remzi H. Arpaci-Dusseau, Nathan C. Burnett, Timothy E. Denehy, Thomas J.
On the Placement of Web Server Replicas Yu Cai. Paper On the Placement of Web Server Replicas Lili Qiu, Venkata N. Padmanabhan, Geoffrey M. Voelker Infocom.
1 Lecture 16: Data Storage Wednesday, November 6, 2006.
Kalman Filter and Data Streaming Presented By :- Ankur Jain Department of Computer Science 7/21/03.
Informed Prefetching and Caching R. Hugo Patterson, Garth A. Gibson, Eka Ginting, Daniel Stodolsky, Jim Zelenka.
Performance directed energy management using BOS technique
RAID Redundant Arrays of Independent Disks
CSI 400/500 Operating Systems Spring 2009
Storage Virtualization
Auburn University COMP7500 Advanced Operating Systems I/O-Aware Load Balancing Techniques (2) Dr. Xiao Qin Auburn University.
Scalable High Performance Main Memory System Using PCM Technology
Part V Memory System Design
Row Buffer Locality Aware Caching Policies for Hybrid Memories
Predictive Performance
Cluster Resource Management: A Scalable Approach
Cooperative Caching, Simplified
Web switch support for differentiated services
Admission Control and Request Scheduling in E-Commerce Web Sites
Qingbo Zhu, Asim Shankar and Yuanyuan Zhou
Lecture 15 Reading: Bacon 7.6, 7.7
Virtual Memory: Working Sets
Presentation transcript:

1 Storage-Aware Caching: Revisiting Caching for Heterogeneous Systems Brian Forney Andrea Arpaci-Dusseau Remzi Arpaci-Dusseau Wisconsin Network Disks University of Wisconsin at Madison

2 Circa 1970 OS policy Intense research period in policies Wide variety developed; many used today –Examples: Clock, LRU Simple storage environment –Focus: workload –Assumption: consistent retrieval cost buffer cache App

3 Today Rich storage environment –Devices attached in many ways –More devices –Increased device sophistication Mismatch: Need to reevaluate LAN WAN policy buffer cache App

4 Problem illustration Uniform workload Two disks LRU policy Slow disk is bottleneck Problem: Policy is oblivious –Does not filter well fastslow

5 General solution Integrate workload and device performance –Balance work across devices –Work: cumulative delay Cannot throw out: –Existing non-cost aware policy research –Existing caching software

6 Our solution: Overview Generic partitioning framework –Old idea –One-to-one mapping: device  partition –Each partition has cost-oblivious policy –Adjust partition sizes Advantages –Aggregates performance information –Easily and quickly adapts to workload and device performance changes –Integrates well with existing software Key: How to pick the partition sizes

7 Outline Motivation Solution overview Taxonomy Dynamic partitioning algorithm Evaluation Summary

8 Partitioning algorithms Static –Pro: Simple –Con: Wasteful Dynamic –Adapts to workload Hotspots Access pattern changes –Handles device performance faults Used dynamic

9 Our algorithm: Overview 1.Observe: Determine per-device cumulative delay 2.Act: Repartition cache 3.Save & reset: Clear last W requests Observe Act Save & reset

10 Algorithm: Observe Want accurate system balance view Record per-device cumulative delay for last W completed disk requests –At client –Includes network time

11 Algorithm: Act Categorize each partition –Page consumers Cumulative delay above threshold  possible bottleneck –Page suppliers Cumulative delay below mean  lose pages without decreasing performance –Neither Always have page suppliers if there are page consumers Page consumerPage supplierNeither Page consumerPage supplierNeither Before After

12 Page consumers How many pages? Depends on state: –Warming Cumulative delay increasing Aggressively add pages: reduce queuing –Warm Cumulative delay constant Conservatively add pages –Cooling Cumulative delay decreasing Do nothing; naturally decreases

13 Dynamic partitioning Eager –Immediately change partition sizes –Pro: Matches observation –Con: Some pages temporarily unused Lazy –Change partition sizes on demand –Pro: Easier to implement –Con: May cause over correction

14 Outline Motivation Solution overview Taxonomy Dynamic partitioning algorithm Evaluation Summary

15 Evaluation methodology Simulator Workloads: synthetic and web Caching, RAID-0 client LogGP network With endpoint contention Disks 16 IBM 9LZX First-order model: queuing, seek time & rotational time

16 Evaluation methodology Introduced performance heterogeneity –Disk aging Used current technology trends –Seek and rotation: 10% decrease/year –Bandwidth: 40% increase/year Scenarios –Single disk degradation: Single disk, multiple ages –Incremental upgrades: Multiple disks, two ages –Fault injection Understand dynamic device performance change and device sharing effects Talk only shows single disk degradation

17 Evaluated policies Cost-oblivious: LRU, Clock Storage-aware: Eager LRU, Lazy LRU, Lazy Clock (Clock-Lottery) Comparison: LANDLORD –Cost-aware, non-partitioned LRU –Same as web caching algorithm –Integration problems with modern OSes

18 Synthetic Workload: Read requests, exponentially distributed around 34 KB, uniform load across disks A single slow disk greatly impacts performance. Eager LRU, Lazy LRU, Lazy Clock, and LANDLORD robust as slow disk performance degrades

19 Web Workload: 1 day image server trace at UC Berkeley, reads & writes Eager LRU and LANDLORD are the most robust.

20 Summary Problem: Mismatch between storage environment and cache policy –Current buffer cache policies lack device information Policies need to include storage environment information Our solution: Generic partitioning framework –Aggregates performance information –Adapts quickly –Allows for use of existing policies

21 Questions? More information at

22

23 Future work Implementation in Linux Cost-benefit algorithm Study of integration with prefetching and layout

24 Problems with LANDLORD Does not mesh well with unified buffer caches (assumes LRU) LRU-based: not always desirable –Example: Databases Suffers from a “memory effect” –Can be much slower to adapt

25 Disk aging of an IBM 9LZX Age (years)Bandwidth (MB/s)Avg. seek time (ms) Avg. rotational delay (ms)

26 Web without writes Workload: Web workload where writes are replaced with reads Eager LRU and LANDLORD are the most robust.

27

28 Problem illustration OS Applications FastSlow Disks

29 Problem illustration OS Applications FastSlow Disks

30 Lack of information buffer cache OS Applications Disks DFSes, NAS, & new devices drivers

31 Solution: overview Partition the cache –One per device –Cost-oblivious policies (e.g. LRU) in partitions –Aggregate device perf. Dynamically reallocate pages buffer cache policy

32 Forms of dynamic partitioning Eager –Change sizes

33 Tomorrow New devices New paradigms Increasingly rich storage environment Mismatch: Reevaluation needed LAN WAN buffer cache policy ? ?

34 WiND project Wisconsin Network Disks Building manageable distributed storage Focus: Local-area networked storage Issues similar in wide-area