BUCS— A Bottom Up Caching Structure for Storage Servers Ming Zhang and Dr. Ken Qing Yang HPCL, Dept. of ECE URI Storage Volume Data storage plays an essential.

Slides:



Advertisements
Similar presentations
IS 4506 Tuning and Monitoring Internet Information Server.
Advertisements

Engenio 7900 HPC Storage System. 2 LSI Confidential LSI In HPC LSI (Engenio Storage Group) has a rich, successful history of deploying storage solutions.
Preparing For Server Installation Instructor: Enoch E. Damson.
Protocols and software for exploiting Myrinet clusters Congduc Pham and the main contributors P. Geoffray, L. Prylli, B. Tourancheau, R. Westrelin.
The TickerTAIP Parallel RAID Architecture P. Cao, S. B. Lim S. Venkatraman, J. Wilkes HP Labs.
R.A.I.D. Copyright © 2005 by James Hug Redundant Array of Independent (or Inexpensive) Disks.
Chapter 5: Server Hardware and Availability. Hardware Reliability and LAN The more reliable a component, the more expensive it is. Server hardware is.
Reporter:PCLee With a significant increase in the design complexity of cores and associated communication among them, post-silicon validation.
ANALYZING STORAGE SYSTEM WORKLOADS Paul G. Sikalinda, Pieter S. Kritzinger {psikalin, DNA Research Group Computer Science Department.
NETWORK LOAD BALANCING NLB.  Network Load Balancing (NLB) is a Clustering Technology.  Windows Based. (windows server).  To scale performance, Network.
Memory Hierarchy. Smaller and faster, (per byte) storage devices Larger, slower, and cheaper (per byte) storage devices.
An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool.
Web Caching Schemes1 A Survey of Web Caching Schemes for the Internet Jia Wang.
Introduction to Systems Architecture Kieran Mathieson.
Chapter 1 Introduction 1.1A Brief Overview - Parallel Databases and Grid Databases 1.2Parallel Query Processing: Motivations 1.3Parallel Query Processing:
1 Software Testing and Quality Assurance Lecture 40 – Software Quality Assurance.
Chapter 3 Chapter 3: Server Hardware. Chapter 3 Learning Objectives n Describe the base system requirements for Windows NT 4.0 Server n Explain how to.
©Company confidential 1 Performance Testing for TM & D – An Overview.
1 Presenter: Chien-Chih Chen Proceedings of the 2002 workshop on Memory system performance.
An Efficient Programmable 10 Gigabit Ethernet Network Interface Card Paul Willmann, Hyong-youb Kim, Scott Rixner, and Vijay S. Pai.
© 2001 by Prentice Hall8-1 Local Area Networks, 3rd Edition David A. Stamper Part 3: Software Chapter 8 Client/Server Architecture.
Hystor : Making the Best Use of Solid State Drivers in High Performance Storage Systems Presenter : Dong Chang.
RAID: High-Performance, Reliable Secondary Storage Mei Qing & Chaoxia Liao Nov. 20, 2003.
Module 9 Review Questions 1. The ability for a system to continue when a hardware failure occurs is A. Failure tolerance B. Hardware tolerance C. Fault.
Report : Zhen Ming Wu 2008 IEEE 9th Grid Computing Conference.
Chapter 3 Memory Management: Virtual Memory
Continuous resource monitoring for self-predicting DBMS Dushyanth Narayanan 1 Eno Thereska 2 Anastassia Ailamaki 2 1 Microsoft Research-Cambridge, 2 Carnegie.
1 Fault Tolerance in the Nonstop Cyclone System By Scott Chan Robert Jardine Presented by Phuc Nguyen.
Design and Performance Evaluation of Networked Storage Architectures Xubin He July 25,2002 Dept. of Electrical and Computer Engineering.
18 June 2001 Optimizing Distributed System Performance via Adaptive Middleware Load Balancing Ossama Othman Douglas C. Schmidt
A Unified, Low-overhead Framework to Support Continuous Profiling and Optimization Xubin (Ben) He Storage Technology & Architecture Research(STAR)
Database Edition for Sybase Sales Presentation. Market Drivers DBAs are facing immense time pressure in an environment with ever-increasing data Continuous.
Emalayan Vairavanathan
Copyright © 2010, Scryer Analytics, LLC. All rights reserved. Optimizing SAS System Performance − A Platform Perspective Patrick McDonald Scryer Analytics,
RAPID-Cache – A Reliable and Inexpensive Write Cache for Disk I/O Systems Yiming Hu Qing Yang Tycho Nightingale.
CSE 451: Operating Systems Section 10 Project 3 wrap-up, final exam review.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
LECC2003 AmsterdamMatthias Müller A RobIn Prototype for a PCI-Bus based Atlas Readout-System B. Gorini, M. Joos, J. Petersen (CERN, Geneva) A. Kugel, R.
Hardware Trends. Contents Memory Hard Disks Processors Network Accessories Future.
1 Martin Schulz, Lawrence Livermore National Laboratory Brian White, Sally A. McKee, Cornell University Hsien-Hsin Lee, Georgia Institute of Technology.
Our work on virtualization Chen Haogang, Wang Xiaolin {hchen, Institute of Network and Information Systems School of Electrical Engineering.
Securing and Monitoring 10GbE WAN Links Steven Carter Center for Computational Sciences Oak Ridge National Laboratory.
A Measurement Based Memory Performance Evaluation of High Throughput Servers Garba Isa Yau Department of Computer Engineering King Fahd University of Petroleum.
Srihari Makineni & Ravi Iyer Communications Technology Lab
I/O Computer Organization II 1 Introduction I/O devices can be characterized by – Behavior: input, output, storage – Partner: human or machine – Data rate:
Introduction to DFS. Distributed File Systems A file system whose clients, servers and storage devices are dispersed among the machines of a distributed.
Increasing Web Server Throughput with Network Interface Data Caching October 9, 2002 Hyong-youb Kim, Vijay S. Pai, and Scott Rixner Rice Computer Architecture.
Business Data Communications, Fourth Edition Chapter 11: Network Management.
Fast Crash Recovery in RAMCloud. Motivation The role of DRAM has been increasing – Facebook used 150TB of DRAM For 200TB of disk storage However, there.
CS 4396 Computer Networks Lab Router Architectures.
Improving Disk Throughput in Data-Intensive Servers Enrique V. Carrera and Ricardo Bianchini Department of Computer Science Rutgers University.
Introduction: Memory Management 2 Ideally programmers want memory that is large fast non volatile Memory hierarchy small amount of fast, expensive memory.
Rick Claus Sr. Technical Evangelist,
Full and Para Virtualization
© 2006 EMC Corporation. All rights reserved. The Host Environment Module 2.1.
STORAGE ARCHITECTURE/ MASTER): Disk Storage: What Are Your Options? Randy Kerns Senior Partner The Evaluator Group.
Exploiting Instruction Streams To Prevent Intrusion Milena Milenkovic.
Chapter 8 System Management Semester 2. Objectives  Evaluating an operating system  Cooperation among components  The role of memory, processor,
Jeffrey Ellak CS 147. Topics What is memory hierarchy? What are the different types of memory? What is in charge of accessing memory?
Exploiting Task-level Concurrency in a Programmable Network Interface June 11, 2003 Hyong-youb Kim, Vijay S. Pai, and Scott Rixner Rice Computer Architecture.
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
COMP7500 Advanced Operating Systems I/O-Aware Load Balancing Techniques Dr. Xiao Qin Auburn University
1© Copyright 2015 EMC Corporation. All rights reserved. NUMA(YEY) BY JACOB KUGLER.
A Framework For Trusted Instruction Execution Via Basic Block Signature Verification Milena Milenković, Aleksandar Milenković, and Emil Jovanov Electrical.
A PRESENTATION ON VIRTUAL MEMORY (PAGING) Submitted to Submitted by Prof. Dr. Ashwani kumar Ritesh verma Dept. Of Physics Mtech (Instrumentation) Roll.
Chapter 1: Introduction
Overview Introduction VPS Understanding VPS Architecture
Storage Networking Protocols
(Architectural Support for) Semantically-Smart Disk Systems
Database System Architectures
Presentation transcript:

BUCS— A Bottom Up Caching Structure for Storage Servers Ming Zhang and Dr. Ken Qing Yang HPCL, Dept. of ECE URI Storage Volume Data storage plays an essential role in today’s fast-growing data-intensive network services. Online data storage doubles every 9 months How much data are there? 1. Read (Text) 100 KB/hr, 25 GB/lifetime per person 2. Hear 10KB/s) 40 MB/hr, 10 TB/lifetime per person 3. See MB/s) 2 GB/hr, 500TB/lifetime per person Storage Cost in an IT Dept. Storage Speed A Server-to-Storage Bottleneck Current Storage Servers: Motivations Data bus is becoming a bottleneck - 1 Gigabit NIC support 2 Gb/s (duplex) - 10Gb/s NIC is on the way - A 10Gb/s TOE can achieve 7.9Gb/s - 6 SATA RAID0 can achieve >300MB/s - 1 PCI bus: 66 * 8 = 533 MB/s - PCI-X (1GB/s = 8Gb/s) -PCI Express, InfiniBand Motivations Embedded systems have become more powerful than ever BUCS Functional marriage between HBA and NIC Caching at controller level Data are placed at lower level caches Replacing using LRU among L1, L2, Disk Only meta data are passed to bus and RAM Most reads and writes from network are done in lower level caches with minimum bus transactions BUCS Controller Prototype Read Performance (Single Client) Write Performance (Single Client) Performance (Four Clients) TPC-C Trace Results Request Response Time Randomly chosen 10K requests. Conclusions A New Cache Hierarchy Structure Eliminate bus bottleneck Reduce Response time Increase system throughput by 3 times Compliance with Existing Standards Ready to be used HELP--- Hardware Environment for Low-overhead Profiling Ming Zhang and Ken Qing Yang HPCL, Dept. of ECE, URI Why Profiling? System profiling has been an important mechanism to observe system activities Profiling-based optimization has become a key technique in computer designs Continuous and online optimization is needed because of dynamic nature of computer systems Traditional approaches suffer from high overhead to already overloaded systems HELP— Hardware Environment for Low-overhead Profiling Offload computing overheads from host processors to an embedded processor Continuous feedback loop model: 1. Low overhead profiling of system events 2. Parallel processing of raw data and setting up new policies 3. Applying the new policies to host HELP Architecture Low cost, low power embedded processor Expandable with secondary PCI slot Interface with host via standard PCI slot Adaptive Caching Policy IOMeter results of buffer cache with random write workloads HELP can help by adaptively setting cache policies Potential Applications Performance: - Low overhead profiling - Adaptive pre-fetching and caching policies - Online code optimization and recompilation Availability: - Monitor system events and report failures or faults Security: - Monitor abnormal system accesses, high risk events, intrusion detection …… Conclusion HELP is a low cost, low power tool for system profiling and optimization Plug-and-Play device Can be applied to any computer system with PCI slots “Offload” feature makes it superior to other existing tools. Computer Architecture Research at HPCL ( Storage cost as proportion of total IT spending as compared to server cost (Src Brocade) BUCS – Bottom Up Caching Structure Host RAM CPU HBA Disk Ctrlr NIC Cache Sys RAM CPU HBA Disk Crlr NIC Bus is in a critical path. System Bus network System bus network Host HELP Measured Performance Run PostMark and popular Linux profiling tool, LTT The following table shows the measured time and overhead HELP reduces overhead of profiling to negligible Read Write