Concurrency Control & Caching Consistency Issues and Survey Dingshan He November 18, 2002.

Slides:



Advertisements
Similar presentations
What is OceanStore? - 10^10 users with files each - Goals: Durability, Availability, Enc. & Auth, High performance - Worldwide infrastructure to.
Advertisements

Consistency and Replication Chapter 7 Part II Replica Management & Consistency Protocols.
Sanjay Ghemawat, Howard Gobioff and Shun-Tak Leung
G O O G L E F I L E S Y S T E M 陳 仕融 黃 振凱 林 佑恩 Z 1.
Computer Science Lecture 20, page 1 CS677: Distributed OS Today: Coda, xFS Case Study: Coda File System Brief overview of other recent file systems –xFS.
Study of Hurricane and Tornado Operating Systems By Shubhanan Bakre.
G Robert Grimm New York University Disconnected Operation in the Coda File System.
Disconnected Operation in the Coda File System James J. Kistler and M. Satyanarayanan Carnegie Mellon University Presented by Deepak Mehtani.
CS 582 / CMPE 481 Distributed Systems
Coda file system: Disconnected operation By Wallis Chau May 7, 2003.
Computer Science Lecture 21, page 1 CS677: Distributed OS Today: Coda, xFS Case Study: Coda File System Brief overview of other recent file systems –xFS.
Data Sharing in OSD Environment Dingshan He September 30, 2002.
Distributed Systems Fall 2009 Replication Fall 20095DV0203 Outline Group communication Fault-tolerant services –Passive and active replication Highly.
Jeff Chheng Jun Du.  Distributed file system  Designed for scalability, security, and high availability  Descendant of version 2 of Andrew File System.
Distributed File System: Data Storage for Networks Large and Small Pei Cao Cisco Systems, Inc.
Mobility Presented by: Mohamed Elhawary. Mobility Distributed file systems increase availability Remote failures may cause serious troubles Server replication.
Client-Server Computing in Mobile Environments
Northwestern University 2007 Winter – EECS 443 Advanced Operating Systems The Google File System S. Ghemawat, H. Gobioff and S-T. Leung, The Google File.
Distributed Databases
FARSITE: Federated, Available, and Reliable Storage for an Incompletely Trusted Environment.
Distributed Systems Tutorial 11 – Yahoo! PNUTS written by Alex Libov Based on OSCON 2011 presentation winter semester,
CSC 456 Operating Systems Seminar Presentation (11/13/2012) Leon Weingard, Liang Xin The Google File System.
1 6.4 Distribution Protocols Different ways of propagating/distributing updates to replicas, independent of the consistency model. First design issue.
Distributed Systems Principles and Paradigms Chapter 10 Distributed File Systems 01 Introduction 02 Communication 03 Processes 04 Naming 05 Synchronization.
Mobility in Distributed Computing With Special Emphasis on Data Mobility.
Pond: the OceanStore Prototype Sean Rhea, Patric Eaton, Dennis Gells, Hakim Weatherspoon, Ben Zhao, and John Kubiatowicz University of California, Berkeley.
1 Moshe Shadmon ScaleDB Scaling MySQL in the Cloud.
Replication ( ) by Ramya Balakumar
Distributed File Systems Case Studies: Sprite Coda.
Distributed File Systems Overview  A file system is an abstract data type – an abstraction of a storage device.  A distributed file system is available.
Chapter 20 Distributed File Systems Copyright © 2008.
Distributed File System By Manshu Zhang. Outline Basic Concepts Current project Hadoop Distributed File System Future work Reference.
CEPH: A SCALABLE, HIGH-PERFORMANCE DISTRIBUTED FILE SYSTEM S. A. Weil, S. A. Brandt, E. L. Miller D. D. E. Long, C. Maltzahn U. C. Santa Cruz OSDI 2006.
MapReduce and GFS. Introduction r To understand Google’s file system let us look at the sort of processing that needs to be done r We will look at MapReduce.
Presenters: Rezan Amiri Sahar Delroshan
IM NTU Distributed Information Systems 2004 Replication Management -- 1 Replication Management Yih-Kuen Tsay Dept. of Information Management National Taiwan.
Robustness in the Salus scalable block store Yang Wang, Manos Kapritsos, Zuocheng Ren, Prince Mahajan, Jeevitha Kirubanandam, Lorenzo Alvisi, and Mike.
 Distributed file systems having transaction facility need to support distributed transaction service.  A distributed transaction service is an extension.
INTRODUCTION TO DBS Database: a collection of data describing the activities of one or more related organizations DBMS: software designed to assist in.
GFS. Google r Servers are a mix of commodity machines and machines specifically designed for Google m Not necessarily the fastest m Purchases are based.
CS425 / CSE424 / ECE428 — Distributed Systems — Fall 2011 Some material derived from slides by Prashant Shenoy (Umass) & courses.washington.edu/css434/students/Coda.ppt.
Distributed File Systems
Caching Consistency and Concurrency Control Contact: Dingshan He
Fault Tolerance and Replication
Write Conflicts in Optimistic Replication Problem: replicas may accept conflicting writes. How to detect/resolve the conflicts? client B client A replica.
POND: THE OCEANSTORE PROTOTYPE S. Rea, P. Eaton, D. Geels, H. Weatherspoon, J. Kubiatowicz U. C. Berkeley.
Chapter 7: Consistency & Replication IV - REPLICATION MANAGEMENT By Jyothsna Natarajan Instructor: Prof. Yanqing Zhang Course: Advanced Operating Systems.
Introduction to Distributed Databases Yiwei Wu. Introduction A distributed database is a database in which portions of the database are stored on multiple.
11.6 Distributed File Systems Consistency and Replication Xiaolong Wu Instructor: Dr Yanqing Zhang Advanced Operating System.
Highly Available Services and Transactions with Replicated Data Jason Lenthe.
GPFS: A Shared-Disk File System for Large Computing Clusters Frank Schmuck & Roger Haskin IBM Almaden Research Center.
Computer Science Lecture 19, page 1 CS677: Distributed OS Last class: Distributed File Systems Issues in distributed file systems Sun’s Network File System.
THE EVOLUTION OF CODA M. Satyanarayanan Carnegie-Mellon University.
Dsitributed File Systems
Distributed File System. Outline Basic Concepts Current project Hadoop Distributed File System Future work Reference.
Mobility Victoria Krafft CS /25/05. General Idea People and their machines move around Machines want to share data Networks and machines fail Network.
Nomadic File Systems Uri Moszkowicz 05/02/02.
Outline Announcements Fault Tolerance.
7.1. CONSISTENCY AND REPLICATION INTRODUCTION
Today: Coda, xFS Case Study: Coda File System
File-System Interface
Distributed Systems CS
Replica Placement Model: We consider objects (and don’t worry whether they contain just data or code, or both) Distinguish different processes: A process.
DISTRIBUTED SYSTEMS Principles and Paradigms Second Edition ANDREW S
THE GOOGLE FILE SYSTEM.
Lecture 4: File-System Interface
System-Level Support CIS 640.
Distributed Databases
Presentation transcript:

Concurrency Control & Caching Consistency Issues and Survey Dingshan He November 18, 2002.

Outline Infrastructure assumptions Concurrency control & caching consistency issues in the infrastructure Survey concurrency control & caching consistency solutions in existing systems –Storage Tank –Oceanstore –Coda Discussion

Infrastructure Assumptions Entities –Clients –OSD’s –Regional Managers Mobility –Clients could have high mobility –OSD’s have mediate mobility –Regional Managers are relatively static

Infrastructure Assumptions (cont.) Connectivity –Disconnection is possible at any place –Clients could have weak connectivity (low bandwidth/long latency) Any of the three kinds of entities can be dynamically created and inserted into the infrastructure

Client Behavior Caching information for performance as well as in expectation of disconnection High mobility –Transfer of regional managers –Changing of concurrency control & caching consistency information Weak connectivity –Reduce message volume

OSD Behavior Mobility –Transfer of regional managers –Handing over of concurrency control & caching consistency management task –Redirecting requests to new regional managers

Regional Manager Behavior Support transferring of clients and OSD’s Efficiently performs handing over Disconnection: –Regional managers get partitioned –Maintain strong consistency within connected partitions –Maintain enough information for reintegration

Exploiting Object Features No single solution could satisfy all situations Object should have its own requirement Our design should identify these requirements, abstract them into several levels, and applies corresponding mechanism accordingly

Survey of Several Existing Systems IBM Storage Tank Oceanstore Coda

Storage Tank with OSD

IBM Storage Tank Protocol A locking and data consistency model Allows the IBM Storage Tank distributed storage system to look and behave like a local file system Objective: provides strong data consistency between clients and servers in distributed environment

Storage Tank Features Concurrency control –Semi-preemptive session locks –Byte-range locks (mandatory and advisory) –Cache coherency data locks Sequential consistency Direct I/O for caching applications (database) Publish consistency for web updates Aggressive caching –Write-back caching of data and metadata –Session state via semi-preemptive locks

Storage Tank Features (cont.) Data consistency across client failures –Leases for failure detection and coordinated recovery Implicit leases Opportunistic renewal

Storage Tank Client Cache Data Metadata Locks

Comments on Storage Tank designed to provide performance that is comparable to that of file systems built on bus-attached, high-performance storage. Works in data center model Restricted to enterprise-wide data sharing, physically

Oceanstore’s Update model An update is a list of predicate-action pairs If some predicate = true, the update commits Each update is applied atomically Can perform many useful predicates and actions against encrypted data –Search over encrypted data –Delete and append using a position-dependent block cipher

Oceanstore Consistency Solution User a two tier architecture –Primary tier: uses distributed consistency Replicas use Byzantine agreement protocol Replicas sign decisions using proactive signatures –Secondary tier: acts as a distributed read/write cache Kept up-to-date via “push” or “pull” Supports connected and disconnected modes of operation

Oceanstore Update Serialization Clients optimistically timestamp updates with commit times Secondary replicas order updates by timestamps tentatively Primary tier picks total order guided by timestamps using Byzantine agreement protocol

Comments on Oceanstore Similar infrastructure It does not separate metadata and data

Coda Volume Management Volume Storage Group (VSG) : set of servers with replicas of a volume Degree of replication and identity of replication site are specified when a volume is created Above info. is stored in volume replication database presenting at every server Venus keeps track of Available VSG (AVSG) for every volume from which it has cached data

Coda Read/Write Strategy Client obtains data from one member of its AVSG called the preferred server Other servers are contacted to verify that the preferred server does have latest copy When a file is closed after modification it is transferred in parallel to all members of the AVSG

Coda’s Disconnected Operation Aim: to provide a file system with resilience to network failures Venus performs as a pseudo-server Updates have to be revalidated with respect to integrity and protection by real servers Venus records sufficient information to replay update activity in a per-volume log called replay log

Coda’s Reintegration The replay log is shipped in parallel to the AVSG, and executed independently at each member Each replica of an object is tagged with a storeid Storeid of objects mentioned in replay log vs. storeid of server’s replica of the object 1) Lock referenced objects, 2) validate and execute each operation, 3) data transfer, and 4) commits transaction and release locks

Comments on Coda Designed for specific application, campus environment particularly Optimistic replica control –Conflicting updates –Security of cached replica