Viceroy: A scalable and dynamic emulation of the Butterfly Presented in CS294-4 by Sailesh Krishnamurthy Sep 22, 2003.

Slides:



Advertisements
Similar presentations
SkipNet: A Scalable Overlay Network with Practical Locality Properties Nick Harvey, Mike Jones, Stefan Saroiu, Marvin Theimer, Alec Wolman Microsoft Research.
Advertisements

Dynamic Replica Placement for Scalable Content Delivery Yan Chen, Randy H. Katz, John D. Kubiatowicz {yanchen, randy, EECS Department.
Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan MIT and Berkeley presented by Daniel Figueiredo Chord: A Scalable Peer-to-peer.
Pastry Peter Druschel, Rice University Antony Rowstron, Microsoft Research UK Some slides are borrowed from the original presentation by the authors.
Scalable Content-Addressable Network Lintao Liu
1 Greedy Forwarding in Dynamic Scale-Free Networks Embedded in Hyperbolic Metric Spaces Dmitri Krioukov CAIDA/UCSD Joint work with F. Papadopoulos, M.
1Department of Electrical Engineering and Computer Science, University of Michigan, USA. 2Department of Computer Science, National University of Singapore,
CHORD – peer to peer lookup protocol Shankar Karthik Vaithianathan & Aravind Sivaraman University of Central Florida.
Viceroy: Scalable Emulation of Butterfly Networks For Distributed Hash Tables By: Dahlia Malkhi, Moni Naor & David Ratajzcak Nov. 11, 2003 Presented by.
Technische Universität Chemnitz Kurt Tutschku Vertretung - Professur Rechner- netze und verteilte Systeme Chord - A Distributed Hash Table Yimei Liao.
The Chord P2P Network Some slides have been borowed from the original presentation by the authors.
Chord: A scalable peer-to- peer lookup service for Internet applications Ion Stoica, Robert Morris, David Karger, M. Frans Kaashock, Hari Balakrishnan.
Distributed Hash Tables CPE 401 / 601 Computer Network Systems Modified from Ashwin Bharambe and Robert Morris.
Online Social Networks and Media Navigation in a small world.
Symphony: Distributed Hashing in a Small World Gurmeet Singh Manku Mayank Bawa Prabhakar Raghavan Presented by Satpreet Singh.
Lecture 7 CS 728 Searchable Networks. Errata: Differences between Copying and Preferential Attachment In generative model: let p k be fraction of nodes.
Common approach 1. Define space: assign random ID (160-bit) to each node and key 2. Define a metric topology in this space,  that is, the space of keys.
Small-World Graphs for High Performance Networking Reem Alshahrani Kent State University.
P2P Network Structured Networks (III) Distributed Hash Tables Pedro García López Universitat Rovira I Virgili
Topics in Reliable Distributed Systems Lecture 2, Fall Dr. Idit Keidar.
SkipNet: A Scalable Overlay Network with Practical Locality Properties Nick Harvey, Mike Jones, Stefan Saroiu, Marvin Theimer, Alec Wolman Presented by.
Chapter 6: Transform and Conquer Trees, Red-Black Trees The Design and Analysis of Algorithms.
SkipNet Christian Schmidt-Madsen, Peter Tiedemann,
Fault-tolerant Routing in Peer-to-Peer Systems James Aspnes Zoë Diamadi Gauri Shah Yale University PODC 2002.
SkipNet: A Scalable Overlay Network with Practical Locality Properties Nick Harvey, Mike Jones, Stefan Saroiu, Marvin Theimer, Alec Wolman Microsoft Research.
Distributed Lookup Systems
Chord: A Scalable Peer-to-peer Lookup Service for Internet Applications Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek and Hari alakrishnan.
SkipNet: A Scaleable Overlay Network With Practical Locality Properties Presented by Rachel Rubin CS294-4: Peer-to-Peer Systems By Nicholas Harvey, Michael.
Topics in Reliable Distributed Systems Fall Dr. Idit Keidar.
P2P Course, Structured systems 1 Skip Net (9/11/05)
P2P Course, Structured systems 1 Introduction (26/10/05)
File Sharing : Hash/Lookup Yossi Shasho (HW in last slide) Based on Chord: A Scalable Peer-to-peer Lookup Service for Internet ApplicationsChord: A Scalable.
Distributed Hash Table Systems Hui Zhang University of Southern California.
Chord & CFS Presenter: Gang ZhouNov. 11th, University of Virginia.
Chord: A Scalable Peer-to-peer Lookup Protocol for Internet Applications Xiaozhou Li COS 461: Computer Networks (precept 04/06/12) Princeton University.
Searching via Your Neighbor’s Neighbor: The Power of Lookahead in P2P Networks Moni Naor Udi Wieder The Weizmann Institute of Science Gurmeet Manku Stanford.
AVL Trees Amanuel Lemma CS252 Algoithms Dec. 14, 2000.
Computer Algorithms Submitted by: Rishi Jethwa Suvarna Angal.
Content Addressable Network CAN. The CAN is essentially a distributed Internet-scale hash table that maps file names to their location in the network.
Online Social Networks and Media
1 Detecting and Reducing Partition Nodes in Limited-routing-hop Overlay Networks Zhenhua Li and Guihai Chen State Key Laboratory for Novel Software Technology.
1 Trees 4: AVL Trees Section 4.4. Motivation When building a binary search tree, what type of trees would we like? Example: 3, 5, 8, 20, 18, 13, 22 2.
Dynamic Networks for Peer-to-Peer Systems Pierre Fraigniaud CNRS LRI, Univ. Paris Sud Joint work with Philippe Gauron.
1 Secure Peer-to-Peer File Sharing Frans Kaashoek, David Karger, Robert Morris, Ion Stoica, Hari Balakrishnan MIT Laboratory.
Chord Advanced issues. Analysis Theorem. Search takes O (log N) time (Note that in general, 2 m may be much larger than N) Proof. After log N forwarding.
Lecture 12 Distributed Hash Tables CPE 401/601 Computer Network Systems slides are modified from Jennifer Rexford.
LightFlood: An Efficient Flooding Scheme for File Search in Unstructured P2P Systems Song Jiang, Lei Guo, and Xiaodong Zhang College of William and Mary.
Idit Keidar, Principles of Reliable Distributed Systems, Technion EE, Spring Principles of Reliable Distributed Systems Lecture 2: Distributed Hash.
© 2004 Goodrich, Tamassia Trees
Chord Advanced issues. Analysis Search takes O(log(N)) time –Proof 1 (intuition): At each step, distance between query and peer hosting the object reduces.
Dynamic Networks for Peer-to-Peer Systems Pierre Fraigniaud CNRS Lab. de Recherche en Informatique (LRI) Univ. Paris-Sud, Orsay Joint work with Philippe.
CSE 486/586, Spring 2014 CSE 486/586 Distributed Systems Distributed Hash Tables Steve Ko Computer Sciences and Engineering University at Buffalo.
Algorithms and Techniques in Structured Scalable Peer-to-Peer Networks
Two Peer-to-Peer Networking Approaches Ken Calvert Net Seminar, 23 October 2001 Note: Many slides “borrowed” from S. Ratnasamy’s Qualifying Exam talk.
HYPERCUBE ALGORITHMS-1
ICDCS 05Adaptive Counting Networks Srikanta Tirthapura Elec. And Computer Engg. Iowa State University.
Analysis of the Evolution of Peer-to-Peer Systems Proseminar “Peer – to – Peer Information Systems” WS 04/05 Prof. Gerhard Weikum Speaker : Emil Zankov.
CS694 - DHT1 Distributed Hash Table Systems Hui Zhang University of Southern California.
Incrementally Improving Lookup Latency in Distributed Hash Table Systems Hui Zhang 1, Ashish Goel 2, Ramesh Govindan 1 1 University of Southern California.
CSE 486/586 Distributed Systems Distributed Hash Tables
Chord: A Scalable Peer-to-Peer Lookup Service for Internet Applications * CS587x Lecture Department of Computer Science Iowa State University *I. Stoica,
Peer-to-Peer Networks 07 Degree Optimal Networks
CSE 486/586 Distributed Systems Distributed Hash Tables
The Chord P2P Network Some slides have been borrowed from the original presentation by the authors.
Know thy Neighbor’s Neighbor Better Routing for Skip Graphs and Small Worlds Moni Naor Udi Wieder.
Peer-to-Peer and Social Networks
Interconnection Network Design Lecture 14
Chord Advanced issues.
By: Dahlia Malkhi, Moni Naor & David Ratajzcak Nov. 11, 2003
CSE 486/586 Distributed Systems Distributed Hash Tables
Presentation transcript:

Viceroy: A scalable and dynamic emulation of the Butterfly Presented in CS294-4 by Sailesh Krishnamurthy Sep 22, 2003

Viceroy Goals: An overlay routing network with Logarithmic path lengths Constant join/leave cost Balanced congestion (~ log(n)/n) Keys, servers mapped to unit ring [0,1) In Chord: each node has all log(n) links In Viceroy: Each node has one log(n) link A link to 1/2 k distance points to a node with a link to 1/2 (k+1) distance

Viceroy Topology Each node has a level between 1..log(n) A `level-k’ node has “right child”: A long-range link to distance 1/2 k (approx.), to a level-(k+1)’ node “left child”: A local link to `level-(k+1)’ node Level ring links (pred,succ of the same level) “up”: A local link to ‘level-(k-1)’ node Ring links (pred,succ on the ring)

Viceroy Topology 01 Level 1 Level 2 Level 3 x y

Distributed level selection Select Level (s) Let n0 = 1/d(s,succ(s)) Select a level among [1 … log(n0)] uniformly at random Sanity When n servers present, then w.h.p. every server estimates: log(n/2logn)<=log(n0)<=3logn Any level l <log(n/2logn) is “sane”

Lookup target in Viceroy Three phases: Proceed to root: up-links to a `level-1’ node for level k = 1..log(n) If distance < 1/2 k use down-left (short link) If distance >= 1/2 k use down-right (long link) If reqd down link doesn’t exist (or if you overshoot target) break to next phase Traverse the ring (pred/succ links, whichever is closer)

Analysis Constant out-degree Expected constant in-degree log(n ) w.h.p. O(log(n )) lookup steps w.h.p log(n ) to `level-1’ node log(n ) for binary search log(n ) for final local search Congestion: Expected log(n )/n lookup load O(log 2 (n )/n) w.h.p.

Improving Lookup Simple lookup - third phase may be too long May be log 2 n links to traverse :-( Fancy lookup Use a combination of global and level rings to get a dilation of log(n) w.h.p. Greedy approach - use the level links if you are still too far away from the target.

What happens on node leave? Problem: we have constant out-degree, but even with a sparse network we could have log(n) in-degree on average Solution: “buckets” - extra background process Idea: improve identity/level-selection so that we have constant number of nodes in each stretch of (log(n)/n) nodes Maintain ~n buckets of ~log(n) contiguous non-overlapping nodes.

Bucket properties Size: always θ(log(n)) Merge with neighbouring bucket when size falls below log(n) Split bucket when size grows above clog(n) Diversity in bucket Each level in [1..log(n)] is represented by (1,c) nodes. Claim: this limits the indegree to 2c - how ?

Some comparisons SkipNets log(n) log(n) ?? SkipList

Some questions How important are const degree networks ? Dilation and congestion same as chord Depends on bucket mechanism What about fault tolerance ?