John Ousterhout Stanford University RAMCloud Overview and Update SEDCL Forum January, 2015.

Slides:



Advertisements
Similar presentations
MinCopysets: Derandomizing Replication in Cloud Storage
Advertisements

Fast Crash Recovery in RAMCloud
Two phase commit. Failures in a distributed system Consistency requires agreement among multiple servers –Is transaction X committed? –Have all servers.
Memory and Object Management in a Distributed RAM-based Storage System Thesis Proposal Steve Rumble April 23 rd, 2012.
PFabric: Minimal Near-Optimal Datacenter Transport Mohammad Alizadeh Shuang Yang, Milad Sharif, Sachin Katti, Nick McKeown, Balaji Prabhakar, Scott Shenker.
John Ousterhout Stanford University RAMCloud Overview and Update SEDCL Retreat June, 2014.
RAMCloud: Scalable High-Performance Storage Entirely in DRAM John Ousterhout Stanford University (with Nandu Jayakumar, Diego Ongaro, Mendel Rosenblum,
Pond The OceanStore Prototype. Pond -- Dennis Geels -- January 2003 Talk Outline System overview Implementation status Results from FAST paper Conclusion.
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
RAMCloud Scalable High-Performance Storage Entirely in DRAM John Ousterhout, Parag Agrawal, David Erickson, Christos Kozyrakis, Jacob Leverich, David Mazières,
RAMCloud 1.0 John Ousterhout Stanford University (with Arjun Gopalan, Ashish Gupta, Ankita Kejriwal, Collin Lee, Behnam Montazeri, Diego Ongaro, Seo Jin.
Log-Structured Memory for DRAM-Based Storage Stephen Rumble, Ankita Kejriwal, and John Ousterhout Stanford University.
Two phase commit. What we’ve learnt so far Sequential consistency –All nodes agree on a total order of ops on a single object Crash recovery –An operation.
Sinfonia: A New Paradigm for Building Scalable Distributed Systems Marcos K. Aguilera, Arif Merchant, Mehul Shah, Alistair Veitch, Christonos Karamanolis.
SEDCL: Stanford Experimental Data Center Laboratory.
CS 140 Lecture Notes: Technology and Operating SystemsSlide 1 Technology Changes Mid-1980’s2009Change CPU speed15 MHz2 GHz133x Memory size8 MB4 GB500x.
OceanStore: An Architecture for Global-Scale Persistent Storage Professor John Kubiatowicz, University of California at Berkeley
Introducing the Platform Lab John Ousterhout Stanford University.
CS 142 Lecture Notes: Large-Scale Web ApplicationsSlide 1 RAMCloud Overview ● Storage for datacenters ● commodity servers ● GB DRAM/server.
MetaSync File Synchronization Across Multiple Untrusted Storage Services Seungyeop Han Haichen Shen, Taesoo Kim*, Arvind Krishnamurthy,
Distributed storage for structured data
BigTable CSE 490h, Autumn What is BigTable? z “A BigTable is a sparse, distributed, persistent multidimensional sorted map. The map is indexed by.
Metrics for RAMCloud Recovery John Ousterhout Stanford University.
The RAMCloud Storage System
Memory and Object Management in RAMCloud Steve Rumble December 3 rd, 2013.
RAMCloud Design Review Recovery Ryan Stutsman April 1,
A Compilation of RAMCloud Slides (through Oct. 2012) John Ousterhout Stanford University.
Assumptions Hypothesis Hopes RAMCloud Mendel Rosenblum Stanford University.
What We Have Learned From RAMCloud John Ousterhout Stanford University (with Asaf Cidon, Ankita Kejriwal, Diego Ongaro, Mendel Rosenblum, Stephen Rumble,
1 The Google File System Reporter: You-Wei Zhang.
SEDCL/Platform Lab Retreat John Ousterhout Stanford University.
RAMCloud Overview John Ousterhout Stanford University.
RAMCloud: a Low-Latency Datacenter Storage System John Ousterhout Stanford University
RAMCloud: Concept and Challenges John Ousterhout Stanford University.
RAMCloud: A Low-Latency Datacenter Storage System Ankita Kejriwal Stanford University (Joint work with Diego Ongaro, Ryan Stutsman, Steve Rumble, Mendel.
Some key-value stores using log-structure Zhichao Liang LevelDB Riak.
Cool ideas from RAMCloud Diego Ongaro Stanford University Joint work with Asaf Cidon, Ankita Kejriwal, John Ousterhout, Mendel Rosenblum, Stephen Rumble,
VLDB2012 Hoang Tam Vo #1, Sheng Wang #2, Divyakant Agrawal †3, Gang Chen §4, Beng Chin Ooi #5 #National University of Singapore, †University of California,
RAMCloud: System Performance Measurements (Jun ‘11) Nandu Jayakumar
1 Moshe Shadmon ScaleDB Scaling MySQL in the Cloud.
Low-Latency Datacenters John Ousterhout Platform Lab Retreat May 29, 2015.
Durability and Crash Recovery for Distributed In-Memory Storage Ryan Stutsman, Asaf Cidon, Ankita Kejriwal, Ali Mashtizadeh, Aravind Narayanan, Diego Ongaro,
RAMCloud: Low-latency DRAM-based storage Jonathan Ellithorpe, Arjun Gopalan, Ashish Gupta, Ankita Kejriwal, Collin Lee, Behnam Montazeri, Diego Ongaro,
RAMCloud: Scalable High-Performance Storage Entirely in DRAM John Ousterhout Stanford University (with Christos Kozyrakis, David Mazières, Aravind Narayanan,
An Introduction to Consensus with Raft
Log-structured Memory for DRAM-based Storage Stephen Rumble, John Ousterhout Center for Future Architectures Research Storage3.2: Architectures.
CS 140 Lecture Notes: Technology and Operating Systems Slide 1 Technology Changes Mid-1980’s2012Change CPU speed15 MHz2.5 GHz167x Memory size8 MB4 GB500x.
Fast Crash Recovery in RAMCloud. Motivation The role of DRAM has been increasing – Facebook used 150TB of DRAM For 200TB of disk storage However, there.
Experiences With RAMCloud Applications John Ousterhout, Jonathan Ellithorpe, Bob Brown.
INTRODUCTION TO DBS Database: a collection of data describing the activities of one or more related organizations DBMS: software designed to assist in.
RAMCloud Overview and Status John Ousterhout Stanford University.
SEDCL, June 7-8, Table Enumeration in RAMCloud Elliott Slaughter.
Problem-solving on large-scale clusters: theory and applications Lecture 4: GFS & Course Wrap-up.
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition File System Implementation.
Plethora: Infrastructure and System Design. Introduction Peer-to-Peer (P2P) networks: –Self-organizing distributed systems –Nodes receive and provide.
Implementing Linearizability at Large Scale and Low Latency Collin Lee, Seo Jin Park, Ankita Kejriwal, † Satoshi Matsushita, John Ousterhout Platform Lab.
CSci8211: Distributed Systems: RAMCloud 1 Distributed Shared Memory/Storage Case Study: RAMCloud Developed by Stanford Platform Lab  Key Idea: Scalable.
John Ousterhout Stanford University RAMCloud Overview and Update SEDCL Retreat June, 2013.
Technology Drill Down: Windows Azure Platform Eric Nelson | ISV Application Architect | Microsoft UK |
The Raft Consensus Algorithm Diego Ongaro and John Ousterhout Stanford University.
© 2009 IBM Corporation Statements of IBM future plans and directions are provided for information purposes only. Plans and direction are subject to change.
RAMCloud and the Low-Latency Datacenter John Ousterhout Stanford Platform Laboratory.
The Stanford Platform Laboratory John Ousterhout and Guru Parulkar Stanford University
Log-Structured Memory for DRAM-Based Storage Stephen Rumble and John Ousterhout Stanford University.
CPS 512 midterm exam #1, 10/7/2016 Your name please: ___________________ NetID:___________ /60 /40 /10.
Flat Datacenter Storage
RAMCloud Architecture
Building a Database on S3
RAMCloud Architecture
Lecture 15 Reading: Bacon 7.6, 7.7
Presentation transcript:

John Ousterhout Stanford University RAMCloud Overview and Update SEDCL Forum January, 2015

● Quick overview of RAMCloud ● Progress since June retreat ● Current projects:  Secondary indexes  Multi-object transactions  New transport architecture January 30, 2015RAMCloud Overview and UpdateSlide 2 Outline

General-purpose storage system for large-scale applications: ● All data is stored in DRAM at all times ● As durable and available as disk ● Simple key-value data model ● Large scale: servers, 100+ TB ● Low latency: 5-10 µs remote access time Project goal: enable a new class of data-intensive applications January 30, 2015RAMCloud Overview and UpdateSlide 3 What is RAMCloud?

January 30, 2015RAMCloud Overview and UpdateSlide 4 RAMCloud Architecture Master Backup Master Backup Master Backup Master Backup … Appl. Library Appl. Library Appl. Library Appl. Library … Datacenter Network Coordinator 1000 – 10,000 Storage Servers 1000 – 100,000 Application Servers Commodity Servers GB per server High-speed networking: ● 5 µs round-trip ● Full bisection bwidth

Data Model: Key-Value Store ● Basic operations:  read(tableId, key) => blob, version  write(tableId, key, blob) => version  delete(tableId, key) ● Other operations:  cwrite(tableId, key, blob, version) => version  Enumerate objects in table  Efficient multi-read, multi-write  Atomic increment ● Not in RAMCloud 1.0:  Atomic updates of multiple objects  Secondary indexes January 30, 2015RAMCloud Overview and UpdateSlide 5 Tables (Only overwrite if version matches) Key (≤ 64KB) Version (64b) Blob (≤ 1MB) Object

● Raft project (new consensus algorithm) finished:  Best Paper Award at USENIX ATC  Diego Ongaro graduated in September  Usage continues to grow ● Incremental performance improvements:  Small random reads: 5.0µs → 4.7µs  Small durable writes: 15µs → 13.5µs ● Ongoing experiments with potential applications (more details in upcoming talk) January 30, 2015RAMCloud Overview and UpdateSlide 6 Updates

● Participants: Ankita Kejriwal, Stephen Yang ● Many interesting issues:  Representation (objects, indexes)  Scalability  Consistency  Crash recovery ● Status in June:  Skeletal implementation running  Many restrictions, missing features (e.g. fixed-size keys) ● Progress:  Implemented indexlet split/migrate for reconfiguration  Reworked B-tree implementation to eliminate restrictions  Working towards SOSP paper submission January 30, 2015RAMCloud Overview and UpdateSlide 7 Secondary Indexes

● Participants: Collin Lee, Seojin Park, Ankita Kejriwal ● Based on general-purpose linearizability support  Discussed at June retreat  Completed this fall ● Commit protocol designed:  Client-driven (similar to Sinfonia)  Capitalizes on linearizability infrastructure  Detailed talk coming next ● Implementation underway ● Targeting SOSP paper (March) January 30, 2015RAMCloud Overview and UpdateSlide 8 Multi-Object Transactions

● Participants: Behnam Montazeri, Henry Qin, Mohammad Alizadeh ● New network protocol for datacenter RPC:  Replace TCP/IP: better latency, scalability  Receiver-driven congestion/flow control  Design & simulation just getting started ● New threading architecture:  Goals: ● Minimize thread crossings ● Reduce latency (polling-based) ● Improve throughput  Design work just starting January 30, 2015RAMCloud Overview and UpdateSlide 9 Clean-Slate Transport Redesign

● Lots of work in progress ● Should have more results by June retreat January 30, 2015RAMCloud Overview and UpdateSlide 10 Conclusion