Where in the world is my data? Sudarshan Kadambi Yahoo! Research VLDB 2011 Joint work with Jianjun Chen, Brian Cooper, Adam Silberstein, David Lomax, Erwin.

Slides:



Advertisements
Similar presentations
Dynamic Replica Placement for Scalable Content Delivery Yan Chen, Randy H. Katz, John D. Kubiatowicz {yanchen, randy, EECS Department.
Advertisements

Copyright © 2007 Ramez Elmasri and Shamkant B. Navathe Slide
Supporting Cooperative Caching in Disruption Tolerant Networks
Alex Cheung and Hans-Arno Jacobsen August, 14 th 2009 MIDDLEWARE SYSTEMS RESEARCH GROUP.
PNUTS: Yahoo!’s Hosted Data Serving Platform Brian F. Cooper, Raghu Ramakrishnan, Utkarsh Srivastava, Adam Silberstein, Philip Bohannon, HansArno Jacobsen,
Consistency and Replication Chapter 7 Part II Replica Management & Consistency Protocols.
PNUTS: Yahoo’s Hosted Data Serving Platform Jonathan Danaparamita jdanap at umich dot edu University of Michigan EECS 584, Fall Some slides/illustrations.
Search and Replication in Unstructured Peer-to-Peer Networks Pei Cao, Christine Lv., Edith Cohen, Kai Li and Scott Shenker ICS 2002.
PNUTS: Yahoo!’s Hosted Data Serving Platform Yahoo! Research present by Liyan & Fang.
Benchmarking Cloud Serving Systems with YCSB Brian F. Cooper, Adam Silberstein, Erwin Tam, Raghu Ramakrishnan, Russell Sears Yahoo! Research Presenter.
BY VAIBHAV NACHANKAR ARVIND DWARAKANATH Evaluation of Hbase Read/Write (A study of Hbase and it’s benchmarks)
1 Content Delivery Networks iBAND2 May 24, 1999 Dave Farber CTO Sandpiper Networks, Inc.
Cooperative Caching of Dynamic Content on a Distributed Web Server Vegard Holmedahl, Ben Smith, Tao Yang Speaker: SeungLak Choi, DB Lab., CS Dept.
CS 347Notes 021 CS 347: Parallel and Distributed Data Management Notes02: Distributed DB Design Hector Garcia-Molina.
Database Replication techniques: a Three Parameter Classification Authors : Database Replication techniques: a Three Parameter Classification Authors :
Storage Management and Caching in PAST, a large-scale, persistent peer- to-peer storage utility Authors: Antony Rowstorn (Microsoft Research) Peter Druschel.
Beneficial Caching in Mobile Ad Hoc Networks Bin Tang, Samir Das, Himanshu Gupta Computer Science Department Stony Brook University.
Peer-to-Peer Based Multimedia Distribution Service Zhe Xiang, Qian Zhang, Wenwu Zhu, Zhensheng Zhang IEEE Transactions on Multimedia, Vol. 6, No. 2, April.
Distributed Systems Fall 2009 Replication Fall 20095DV0203 Outline Group communication Fault-tolerant services –Passive and active replication Highly.
1 Drafting Behind Akamai (Travelocity-Based Detouring) AoJan Su, David R. Choffnes, Aleksandar Kuzmanovic, and Fabian E. Bustamante Department of Electrical.
Performance Evaluation of Peer-to-Peer Video Streaming Systems Wilson, W.F. Poon The Chinese University of Hong Kong.
Introspective Replica Management Yan Chen, Hakim Weatherspoon, and Dennis Geels Our project developed and evaluated a replica management algorithm suitable.
Tanenbaum & Van Steen, Distributed Systems: Principles and Paradigms, 2e, (c) 2007 Prentice-Hall, Inc. All rights reserved Client-Centric.
1 Awareness Services for Digital Libraries Arturo Crespo Hector Garcia-Molina Stanford University.
Concurrency Control & Caching Consistency Issues and Survey Dingshan He November 18, 2002.
On Fairness, Optimizing Replica Selection in Data Grids Husni Hamad E. AL-Mistarihi and Chan Huah Yong IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS,
Consistency Rationing in the Cloud: Pay only when it matters Tim Kraska, Martin Hentschel, Gustavo Alonso, Donald Kossmann Systems Modified.
Chapter 9 Overview  Reasons to monitor SQL Server  Performance Monitoring and Tuning  Tools for Monitoring SQL Server  Common Monitoring and Tuning.
Inexpensive Scalable Information Access Many Internet applications need to access data for millions of concurrent users Relational DBMS technology cannot.
PNUTS: YAHOO!’S HOSTED DATA SERVING PLATFORM FENGLI ZHANG.
Distributed Data Stores – Facebook Presented by Ben Gooding University of Arkansas – April 21, 2015.
Achieving Load Balance and Effective Caching in Clustered Web Servers Richard B. Bunt Derek L. Eager Gregory M. Oster Carey L. Williamson Department of.
Distributed Systems Tutorial 11 – Yahoo! PNUTS written by Alex Libov Based on OSCON 2011 presentation winter semester,
1 6.4 Distribution Protocols Different ways of propagating/distributing updates to replicas, independent of the consistency model. First design issue.
Zhen Feng, Mingwei Xu, Yu Wang and Qing Li Tsinghua University, Beijing, China, Globalcom2013 – NGN Symposium Katto Lab Hiroto Kisara AN ARCHITECTURE FOR.
Database Replication Policies for Dynamic Content Applications Gokul Soundararajan, Cristiana Amza, Ashvin Goel University of Toronto EuroSys 2006: Leuven,
Consistency And Replication
1 BitHoc: BitTorrent for wireless ad hoc networks Jointly with: Chadi Barakat Jayeoung Choi Anwar Al Hamra Thierry Turletti EPI PLANETE 28/02/2008 MAESTRO/PLANETE.
Web Cache Replacement Policies: Properties, Limitations and Implications Fabrício Benevenuto, Fernando Duarte, Virgílio Almeida, Jussara Almeida Computer.
Ahmad Al-Shishtawy 1,2,Tareq Jamal Khan 1, and Vladimir Vlassov KTH Royal Institute of Technology, Stockholm, Sweden {ahmadas, tareqjk,
Using the Small-World Model to Improve Freenet Performance Hui Zhang Ashish Goel Ramesh Govindan USC.
Alireza Angabini Advanced DB class Dr. M.Rahgozar Fall 88.
Hotspot Detection in a Service Oriented Architecture Pranay Anchuri,
Consistent and Efficient Database Replication based on Group Communication Bettina Kemme School of Computer Science McGill University, Montreal.
CEPH: A SCALABLE, HIGH-PERFORMANCE DISTRIBUTED FILE SYSTEM S. A. Weil, S. A. Brandt, E. L. Miller D. D. E. Long, C. Maltzahn U. C. Santa Cruz OSDI 2006.
ECO-DNS: Expected Consistency Optimization for DNS Chen Stephanos Matsumoto Adrian Perrig © 2013 Stephanos Matsumoto1.
Architecture for Caching Responses with Multiple Dynamic Dependencies in Multi-Tier Data- Centers over InfiniBand S. Narravula, P. Balaji, K. Vaidyanathan,
Garo Bournoutian and Alex Orailoglu Proceedings of the 45th ACM/IEEE Design Automation Conference (DAC’08) June /10/28.
PNUTS PNUTS: Yahoo!’s Hosted Data Serving Platform Brian F. Cooper, Raghu Ramakrishnan, Utkarsh Srivastava, Adam Silberstein, Philip Bohannon, HansArno.
Preventive Replication in Database Cluster Esther Pacitti, Cedric Coulon, Patrick Valduriez, M. Tamer Özsu* LINA / INRIA – Atlas Group University of Nantes.
Databases Illuminated
PMIT-6101 Advanced Database Systems By- Jesmin Akhter Assistant Professor, IIT, Jahangirnagar University.
Database replication policies for dynamic content applications Gokul Soundararajan, Cristiana Amza, Ashvin Goel University of Toronto Presented by Ahmed.
Mobile Data Access1 Replication, Caching, Prefetching and Hoarding for Mobile Computing.
Eduardo Gutarra Velez. Outline Distributed Filesystems Motivation Google Filesystem Architecture The Metadata Consistency Model File Mutation.
PROP: A Scalable and Reliable P2P Assisted Proxy Streaming System Computer Science Department College of William and Mary Lei Guo, Songqing Chen, and Xiaodong.
Authors Brian F. Cooper, Raghu Ramakrishnan, Utkarsh Srivastava, Adam Silberstein, Philip Bohannon, Hans-Arno Jacobsen, Nick Puz, Daniel Weaver, Ramana.
Distributed File Systems
Ing. Erick López Ch. M.R.I. Replicación Oracle. What is Replication  Replication is the process of copying and maintaining schema objects in multiple.
R*: An overview of the Architecture By R. Williams et al. Presented by D. Kontos Instructor : Dr. Megalooikonomou.
DATABASE REPLICATION DISTRIBUTED DATABASE. O VERVIEW Replication : process of copying and maintaining database object, in multiple database that make.
1 Benchmarking Cloud Serving Systems with YCSB Brian F. Cooper, Adam Silberstein, Erwin Tam, Raghu Ramakrishnan and Russell Sears Yahoo! Research.
Topics in Distributed Databases Database System Implementation CSE 507 Some slides adapted from Navathe et. Al and Silberchatz et. Al.
Practical Database Design and Tuning
Gregory Kesden, CSE-291 (Storage Systems) Fall 2017
Gregory Kesden, CSE-291 (Cloud Computing) Fall 2016
Distributed Systems CS
Benchmarking Cloud Serving Systems with YCSB
THE GOOGLE FILE SYSTEM.
Presentation transcript:

Where in the world is my data? Sudarshan Kadambi Yahoo! Research VLDB 2011 Joint work with Jianjun Chen, Brian Cooper, Adam Silberstein, David Lomax, Erwin Tam, Raghu Ramakrishnan and Hector Garcia-Molina

Problem Description  Consider a distributed database, with replicas kept in-sync via. an asynchronous replication mechanism.  Consider a social networking application that uses this distributed database.  Consider a user who is based in Europe.  If user’s record is never accessed in Asia, we shouldn't need to pay the network/disk bandwidth to update the record in the Asian replica.

Criteria used to replicate a given record  Dynamic Factors  How often is the record read vs. updated?  Latency of forwarded reads.  Static Factors  Legal Constraints  Critical data items such as billing records might have additional replication requirements. In this presentation, we’ll look at selective replication at a record level that respects policy constraints and minimize replication costs and is tuned to support latency guarantees.

Architecture  PNUTS.  Asynchronous Replication.  Timeline Consistency.  Replicate everywhere.  With selective replication, some replicas have a full copy of record, others only have stubs.  Each stub has the primary key and additional metadata such as list of replicas that have a full copy of the record.  Read for a record at a replica that contains a stub will result in a forwarded read.

Optimization Problem Given the following constraints:  Policy constraints that define the allowable and mandatory locations for full replicas of each record, and the minimum number of full replicas for each record, and  A latency SLA which specifies that a specified fraction of read requests must be served by a local, full replica Choose a replication strategy to minimize the sum of replication bandwidth and forwarding bandwidth for a given work- load. Note: Total Bandwidth = Update Bandwidth + Forwarding Bandwidth

Policy Constraints Based on legal dictates, availability needs and other application requirements. [ CONSTRAINT I ] IF TABLE_NAME = "Users” THEN SET 'MIN_COPIES' = 2 SET 'INCL_LIST' = ’USWest' CONSTRAINT_PRI = 0

Policy Constraints (contd.) [ CONSTRAINT II ] IF TABLE_NAME = "Users" AND FIELD_STR('home_location') = 'france' THEN SET 'MIN_COPIES' = 3 AND SET 'EXCL_LIST' = ’Asia' CONSTRAINT_PRI = 1

Constraint Enforcement  Master makes an initial placement decision when the record is inserted.  R and stub (R) are published to the messaging layer in a single transaction.  If record contents change, full copies can migrate (promotions/demotions).  Constraints are validated when they're supplied.  Our system don't allow constraints to be changed after data is inserted.

Dynamic Placement

Retention Interval  Too short: locations will be quick to surrender full replicas.  Too long: single read can cause a full replica to be retained for a long time.

Latency constraints  Dynamic placement places full copies where reads exceeds writes and stubs elsewhere.  Might be necessary to make extra full copies so that latency SLA is met.  One way to accomplish is by increasing the number of copies.  Another is to increase the retention interval I.

Experimental Setup  Social networking application.  Users have a home location from where their reads and writes originate.  Varied the remote probability, the read/write ratio, the size of reads/writes and user mobility.  For constraint schemes, we use min copies of 2. Each record must have a full copy at the user's home location.

Configuration  Clusters in data centers in US, India, Singapore.  100,000 1 /KB records.  5M read/write operations for each data point.  For dynamic schemes, generated a trace of 6M operations and used the first 1M for warmup.

Varying read/write proportion Insight: Dynamic scheme performs well with increasing number of writes, as it can keep as few as one copy. Due to the adaptation overhead, Dynamic with Constraints performs worse than Static Constraints.

Varying read/write proportion Insight: Latency of the dynamic scheme increases as write proportion increases, as the likelihood increases that an update reaches an expired full replica and causes the demotion of that replica to a stub. Hence there is fewer full replicas, increasing overall latency.

Impact of Locality Insight: As remote probability increases, even though the proportion of writes remains the same at 10%, the effect of those writes get amplified as a higher proportion of records at a replica are obtained adaptively.

Impact of Locality Insight: Static Constraints pays the penalty of having to repeatedly do forwarded reads for friend’s records, without being able to store those records locally.

Real Data Trace  10 days of logs,170,000 unique users, 32 million operations.  The trace is read-heavy; About 40% operations are remote.  Dynamic with Constraints get similar average read latencies (about 4ms) as Full.  Total bandwidth for Full is 8 Mb and 6.8 Mb for Dynamic with Constraints.

Comparison with other techniques  Caching.  Replication.  Our Technique: Caching + Replication  Minimum bookkeeping.  Local + Global decision making.  Also applicable to other web databases such as BigTable and Cassandra.

Related Work  Adaptive Dynamic Replication (Wolfson et al.)  Data Replication in Mariposa (economic model)  Minimal cost replication for availability (Yu and Vahdat)  Cache placement based on analyzing distributed query plan (Kossmann et al.)  Replication strategies in peer to peer networks (Cohen and Shenker)

Conclusion  Proposed mechanism for selectively replicating data at a record granularity while respecting policy constraints.  Currently being rolled out to production at Yahoo.  Examined a dynamic placement scheme with small bookkeeping overhead.  Experimental results show significant improvement in bandwidth usage.  Tunable in order to meet latency constraints.

Thank You

Eventual Consistency  Application can apply concurrent updates to different replicas of the same record in parallel.  PNUTS publishes changes asynchronously to other replicas and resolves conflicts using the local timestamp of each write.  With selective replication: updates are not published to stubs, this may cause a replica to not eventually receive all changes to a record.  To address this, require a full replica to republish its write after detecting overlapping promotions for other replicas of the same record.