Mobility Presented by: Mohamed Elhawary. Mobility Distributed file systems increase availability Remote failures may cause serious troubles Server replication.

Slides:



Advertisements
Similar presentations
Types of Distributed Database Systems
Advertisements

Eventual Consistency Jinyang. Sequential consistency Sequential consistency properties: –Latest read must see latest write Handles caching –All writes.
Computer Science Lecture 20, page 1 CS677: Distributed OS Today: Coda, xFS Case Study: Coda File System Brief overview of other recent file systems –xFS.
File System for Mobile Computing Quick overview of AFS identify the issues for file system design in wireless and mobile environments Design Options for.
Using DSVM to Implement a Distributed File System Ramon Lawrence Dept. of Computer Science
Overview of Mobile Computing (3): File System. File System for Mobile Computing Issues for file system design in wireless and mobile environments Design.
G Robert Grimm New York University Disconnected Operation in the Coda File System.
Disconnected Operation in the Coda File System James J. Kistler and M. Satyanarayanan Carnegie Mellon University Presented by Deepak Mehtani.
Disconnected Operation in the Coda File System James J. Kistler and M. Satyanarayanan Carnegie Mellon University Presented by Cong.
CS 582 / CMPE 481 Distributed Systems
Coda file system: Disconnected operation By Wallis Chau May 7, 2003.
“Managing Update Conflicts in Bayou, a Weakly Connected Replicated Storage System ” Distributed Systems Κωνσταντακοπούλου Τζένη.
Flexible Update Propagation for Weakly Consistent Replication Karin Petersen, Mike K. Spreitzer, Douglas B. Terry, Marvin M. Theimer and Alan J. Demers.
Computer Science Lecture 21, page 1 CS677: Distributed OS Today: Coda, xFS Case Study: Coda File System Brief overview of other recent file systems –xFS.
Department of Electrical Engineering
CSS490 Replication & Fault Tolerance
HA LVS Coda - Devjani Sinha. High Availability-HA n critical commercial applications move on the Internet n An elegant highly available system may have.
Disconnected Operation In The Coda File System James J Kistler & M Satyanarayanan Carnegie Mellon University Presented By Prashanth L Anmol N M Yulong.
G Robert Grimm New York University Bayou: A Weakly Connected Replicated Storage System.
Concurrency Control & Caching Consistency Issues and Survey Dingshan He November 18, 2002.
Jeff Chheng Jun Du.  Distributed file system  Designed for scalability, security, and high availability  Descendant of version 2 of Andrew File System.
University of Pennsylvania 11/21/00CSE 3801 Distributed File Systems CSE 380 Lecture Note 14 Insup Lee.
Distributed File System: Data Storage for Networks Large and Small Pei Cao Cisco Systems, Inc.
Managing Update Conflicts in Bayou, a Weakly Connected Replicated Storage System D. B. Terry, M. M. Theimer, K. Petersen, A. J. Demers, M. J. Spreitzer.
Client-Server Computing in Mobile Environments
Distributed Databases
Distributed File Systems Concepts & Overview. Goals and Criteria Goal: present to a user a coherent, efficient, and manageable system for long-term data.
Distributed Deadlocks and Transaction Recovery.
Adaptation for Mobile Data Access (DM1 & PL1) Yerang Hur Mar. 24, 1998.
Update Propagation with Variable Connectivity Prasun Dewan Department of Computer Science University of North Carolina
1 The Google File System Reporter: You-Wei Zhang.
Distributed Systems Principles and Paradigms Chapter 10 Distributed File Systems 01 Introduction 02 Communication 03 Processes 04 Naming 05 Synchronization.
Mobility in Distributed Computing With Special Emphasis on Data Mobility.
1 Consistency of Replicated Data in Weakly Connected Systems CS444N, Spring 2002 Instructor: Mary Baker.
Distributed File Systems Overview  A file system is an abstract data type – an abstraction of a storage device.  A distributed file system is available.
VICTORIA UNIVERSITY OF WELLINGTON Te Whare Wananga o te Upoko o te Ika a Maui SWEN 432 Advanced Database Design and Implementation Data Versioning Lecturer.
Introduction to DFS. Distributed File Systems A file system whose clients, servers and storage devices are dispersed among the machines of a distributed.
DISCONNECTED OPERATION IN THE CODA FILE SYSTEM J. J. Kistler M. Sataynarayanan Carnegie-Mellon University.
CODA: A HIGHLY AVAILABLE FILE SYSTEM FOR A DISTRIBUTED WORKSTATION ENVIRONMENT M. Satyanarayanan, J. J. Kistler, P. Kumar, M. E. Okasaki, E. H. Siegel,
Mobile File System Byung Chul Tak. AFS  Andrew File System Distributed computing environment developed at CMU provides transparent access to remote shared.
CS425 / CSE424 / ECE428 — Distributed Systems — Fall 2011 Some material derived from slides by Prashant Shenoy (Umass) & courses.washington.edu/css434/students/Coda.ppt.
Distributed File Systems
Information/File Access and Sharing Coda: A Case Study J. Kistler, M. Satyanarayanan. Disconnected operation in the Coda File System. ACM Transaction on.
ENERGY-EFFICIENCY AND STORAGE FLEXIBILITY IN THE BLUE FILE SYSTEM E. B. Nightingale and J. Flinn University of Michigan.
Fault Tolerance and Replication
D u k e S y s t e m s Asynchronous Replicated State Machines (Causal Multicast and All That) Jeff Chase Duke University.
Write Conflicts in Optimistic Replication Problem: replicas may accept conflicting writes. How to detect/resolve the conflicts? client B client A replica.
An Architecture for Mobile Databases By Vishal Desai.
Eventual Consistency Jinyang. Review: Sequential consistency Sequential consistency properties: –All read/write ops follow some total ordering –Read must.
Highly Available Services and Transactions with Replicated Data Jason Lenthe.
Computer Science Lecture 19, page 1 CS677: Distributed OS Last class: Distributed File Systems Issues in distributed file systems Sun’s Network File System.
THE EVOLUTION OF CODA M. Satyanarayanan Carnegie-Mellon University.
Distributed File System. Outline Basic Concepts Current project Hadoop Distributed File System Future work Reference.
Feb 22, 2001CSCI {4,6}900: Ubiquitous Computing1 Announcements Send today with people in your project group. People seem to be dropping off and I.
Computer Science Lecture 19, page 1 CS677: Distributed OS Last Class: Fault tolerance Reliable communication –One-one communication –One-many communication.
Mobility Victoria Krafft CS /25/05. General Idea People and their machines move around Machines want to share data Networks and machines fail Network.
Mobile File Systems.
Nomadic File Systems Uri Moszkowicz 05/02/02.
Disconnected Operation in the Coda File System
Presented By Abhishek Khaitan Adhip Joshi
Today: Coda, xFS Case Study: Coda File System
Distributed File Systems
Distributed File Systems
Outline Announcements Lab2 Distributed File Systems 1/17/2019 COP5611.
Introduction of Week 13 Return assignment 11-1 and 3-1-5
Replica Placement Model: We consider objects (and don’t worry whether they contain just data or code, or both) Distinguish different processes: A process.
Outline Review of Quiz #1 Distributed File Systems 4/20/2019 COP5611.
Last Class: Web Caching
System-Level Support CIS 640.
University of Wisconsin-Madison Presented by: Nick Kirchem
Presentation transcript:

Mobility Presented by: Mohamed Elhawary

Mobility Distributed file systems increase availability Remote failures may cause serious troubles Server replication (higher quality) Client replication, inferior, needs periodic revalidation Cache coherence to combine performance, availability and quality

Mobility (Cont’d) Client server as Coda Pairwise as Bayou Two level hierarchy as Oracle

Disconnected Operation in Coda Enable clients to continue accessing data during failures Cashing data on the clients with right back upon reconnection Designed for a large num. of clients and a smaller num. of servers Server replication and call back break Preserve transparency

Scalability Whole file cashing, cash miss on open only Simple but involve communication overhead Placing functionality on the clients unless violating security or integrity Avoids system wide rapid change

Optimistic replica control Disconnection is involuntary Extended disconnections Leases place time bound on locks but disconnected client loses control after expiration Low degree of write sharing in Unix Applied to server replication for transparency

Client structure Venus is a user level process Works as a cache manager

Venus States Hoarding state: normal operation relying on server replication Reintegration state: resynchronize its state Venus can be in different states with respect to different volumes

Hoarding Hoard useful data in anticipation of disconnection Combines implicit and explicit source of information File reference behavior can’t be predicted Disconnections are unpredictable Priority of cached objects is a function of user given priority (unsafe) and recent usage (hard to manage) Hierarchical cache management

Hoard walking Periodically restores cache equilibrium Unexpected disconnection before a scheduled walk is a problem Revaluate name bindings Revaluate priorities of all entries Solve the callback break problem Policies for directories and files

Emulation Venus functions as a pseudo server Generates temporary file identifiers Behaves as faithful as possible, but no guarantee Cache entries of modified assume infinite priority, except when cache is full Use a replay log Use nonvolatile storage Meta data changes through atomic transactions

Reintegration Venus propagates changes made in emulation Update suspended in the volume being integrated until completion, it may take time! Replace temporary ids with permanent ones, but may be avoided Ship replay log Execute replay as a single transaction, is that fair?

Replay algorithm Log is parsed Log is validated using store id Store create a shadow file and data transfer is deferred Data transfer Commits and release locks In case of failures, user can replay selectively Using dependence graphs may enhance reintegration

Evaluation

Evaluation (cont’d)

Flexible update propagation in Bayou Weekly consistent replication can accommodate propagation policies Don’t assume client server as in Coda Exchange update operations and not the changes, suite low B/W networks Relies on the theory of epidemics Storage vs. bandwidth

Basic Anti-entropy Storage consists of an ordered log of updates and a database Two replicas bring each other up-to-date by agreeing on the set of writes in their logs Servers assigns monotonically increasing accept stamp to new writes Write propagates with stamps and server id Partial accept order to maintain the prefix property Needs to be implemented over TCP

Basic Anti-entropy (Cont’d)

Effective write-log Each replica discard stable writes as a pruned prefix from write log They may have not fully propagated and may require full database transfer Storage bandwidth tradeoff Assigns monotonically increasing CSN to committed writes and infinity to tentative writes Write A precedes write B if…

Anti-entropy with com. writes

Write log truncation To test if a server is missing writes, S.O version vector Penalty involve database transfer and roll back

Write log truncation (Cont’d)

Session guarantees Causal order as a refinement of accept order Each server maintains a logical clock that advances with new writes Total order through

Light weight creation and retirement Versions vectors are updated to include or exclude servers Creation and retirement are handled as a write that propagates via anti-entropy is S i ’s server id T k,i + 1 initialize accept stamp counter Linearly creation increase id’s and version vectors dramatically

Disadvantages When num of replicas is larger than update rate, cost of exchanging version vectors is high If the update rate is larger than the commit rate, the write log will grow large

Policies When to reconcile Which replicas to reconcile How to truncate the write log Selecting a server to create a new replica Depend on network characteristics and threshold values

Evaluation Experiments were taken for an application, how far is that really represent the general dist. File system Only committed writes are propagated Sparc running SunOS (SS) 486 laptops running linux (486) 3000 and 100 byte messages between 3 replicas 520 public key overheads and 1316 bytes update schema and padding

Execution vs. writes propagated

Execution for 100 writes

Network independent

Effect of server creations When servers created from initial one, 20K version vector for 1000 servers When linearly created, 4 MB version vectors for 1000 servers Time to test if a write is among those represented by version vector may grow cubically if linearly created

Conclusion Disconnected operation through Cedar and PCMAIL is less transparent than Coda Golding’s time stamped anti entropy protocol is similar to bayou, with heavier weight mechanism to create replicas and less aggressive to claim storage