Tapestry : An Infrastructure for Fault-tolerant Wide-area Location and Routing Presenter: Chunyuan Liao March 6, 2002 Ben Y.Zhao, John Kubiatowicz, and.

Slides:



Advertisements
Similar presentations
Brocade: Landmark Routing on Peer to Peer Networks Ben Y. Zhao Yitao Duan, Ling Huang, Anthony Joseph, John Kubiatowicz IPTPS, March 2002.
Advertisements

Dynamic Replica Placement for Scalable Content Delivery Yan Chen, Randy H. Katz, John D. Kubiatowicz {yanchen, randy, EECS Department.
Tapestry: Scalable and Fault-tolerant Routing and Location Stanford Networking Seminar October 2001 Ben Y. Zhao
Tapestry: Decentralized Routing and Location SPAM Summer 2001 Ben Y. Zhao CS Division, U. C. Berkeley.
Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan MIT and Berkeley presented by Daniel Figueiredo Chord: A Scalable Peer-to-peer.
Pastry Peter Druschel, Rice University Antony Rowstron, Microsoft Research UK Some slides are borrowed from the original presentation by the authors.
Ranveer Chandra , Kenneth P. Birman Department of Computer Science
1 Accessing nearby copies of replicated objects Greg Plaxton, Rajmohan Rajaraman, Andrea Richa SPAA 1997.
Common approach 1. Define space: assign random ID (160-bit) to each node and key 2. Define a metric topology in this space,  that is, the space of keys.
The Oceanstore Regenerative Wide-area Location Mechanism Ben Zhao John Kubiatowicz Anthony Joseph Endeavor Retreat, June 2000.
Small-world Overlay P2P Network
SCAN: A Dynamic, Scalable, and Efficient Content Distribution Network Yan Chen, Randy H. Katz, John D. Kubiatowicz {yanchen, randy,
15-441: Computer Networking Lecture 26: Networking Future.
Peer to Peer File Sharing Huseyin Ozgur TAN. What is Peer-to-Peer?  Every node is designed to(but may not by user choice) provide some service that helps.
Internet Indirection Infrastructure Ion Stoica UC Berkeley.
A Scalable Content-Addressable Network Authors: S. Ratnasamy, P. Francis, M. Handley, R. Karp, S. Shenker University of California, Berkeley Presenter:
Scalable Adaptive Data Dissemination Under Heterogeneous Environment Yan Chen, John Kubiatowicz and Ben Zhao UC Berkeley.
Each mesh represents a single hop on the route to a given root. Sibling nodes maintain pointers to each other. Each referrer has pointers to the desired.
Tapestry: Wide-area Location and Routing Ben Y. Zhao John Kubiatowicz Anthony D. Joseph U. C. Berkeley.
Weaving a Tapestry Distributed Algorithms for Secure Node Integration, Routing and Fault Handling Ben Y. Zhao (John Kubiatowicz, Anthony Joseph) Fault-tolerant.
Anonymous Gossip: Improving Multicast Reliability in Mobile Ad-Hoc Networks Ranveer Chandra (joint work with Venugopalan Ramasubramanian and Ken Birman)
OceanStore: An Architecture for Global-Scale Persistent Storage Professor John Kubiatowicz, University of California at Berkeley
1 CS 194: Distributed Systems Distributed Hash Tables Scott Shenker and Ion Stoica Computer Science Division Department of Electrical Engineering and Computer.
Decentralized Location Services CS273 Guest Lecture April 24, 2001 Ben Y. Zhao.
Or, Providing High Availability and Adaptability in a Decentralized System Tapestry: Fault-resilient Wide-area Location and Routing Issues Facing Wide-area.
Wide-area cooperative storage with CFS
Or, Providing Scalable, Decentralized Location and Routing Network Services Tapestry: Fault-tolerant Wide-area Application Infrastructure Motivation and.
P2P Course, Structured systems 1 Skip Net (9/11/05)
1/17/01 Changing the Tapestry— Inserting and Deleting Nodes Kris Hildrum, UC Berkeley Joint work with John Kubiatowicz, Satish.
OceanStore An Architecture for Global-Scale Persistent Storage Motivation Feature Application Specific Components - Secure Naming - Update - Access Control-
Tapestry: A Resilient Global-scale Overlay for Service Deployment Ben Y. Zhao, Ling Huang, Jeremy Stribling, Sean C. Rhea, Anthony D. Joseph, and John.
“Umbrella”: A novel fixed-size DHT protocol A.D. Sotiriou.
Tapestry An off-the-wall routing protocol? Presented by Peter, Erik, and Morten.
Internet Indirection Infrastructure (i3) Ion Stoica, Daniel Adkins, Shelley Zhuang, Scott Shenker, Sonesh Surana UC Berkeley SIGCOMM 2002.
Tapestry GTK Devaroy (07CS1012) Kintali Bala Kishan (07CS1024) G Rahul (07CS3009)
1 Plaxton Routing. 2 Introduction Plaxton routing is a scalable mechanism for accessing nearby copies of objects. Plaxton mesh is a data structure that.
Arnold N. Pears, CoRE Group Uppsala University 3 rd Swedish Networking Workshop Marholmen, September Why Tapestry is not Pastry Presenter.
Failure Resilience in the Peer-to-Peer-System OceanStore Speaker: Corinna Richter.
Chord & CFS Presenter: Gang ZhouNov. 11th, University of Virginia.
Pond: the OceanStore Prototype Sean Rhea, Patric Eaton, Dennis Gells, Hakim Weatherspoon, Ben Zhao, and John Kubiatowicz University of California, Berkeley.
Overcast: Reliable Multicasting with an Overlay Network CS294 Paul Burstein 9/15/2003.
Brocade Landmark Routing on P2P Networks Gisik Kwon April 9, 2002.
Information-Centric Networks07a-1 Week 7 / Paper 1 Internet Indirection Infrastructure –Ion Stoica, Daniel Adkins, Shelley Zhuang, Scott Shenker, Sonesh.
Chord: A Scalable Peer-to-peer Lookup Protocol for Internet Applications Xiaozhou Li COS 461: Computer Networks (precept 04/06/12) Princeton University.
Vincent Matossian September 21st 2001 ECE 579 An Overview of Decentralized Discovery mechanisms.
Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan MIT and Berkeley presented by Daniel Figueiredo Chord: A Scalable Peer-to-peer.
Dynamic Source Routing in ad hoc wireless networks Alexander Stojanovic IST Lisabon 1.
MapReduce and GFS. Introduction r To understand Google’s file system let us look at the sort of processing that needs to be done r We will look at MapReduce.
An IP Address Based Caching Scheme for Peer-to-Peer Networks Ronaldo Alves Ferreira Joint work with Ananth Grama and Suresh Jagannathan Department of Computer.
1 More on Plaxton routing There are n nodes, and log B n digits in the id, where B = 2 b The neighbor table of each node consists of - primary neighbors.
Chord: A Scalable Peer-to-peer Lookup Service for Internet Applications Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan Presented.
OceanStore: An Architecture for Global- Scale Persistent Storage.
Scalable Content- Addressable Networks Prepared by Kuhan Paramsothy March 5, 2007.
Paper Survey of DHT Distributed Hash Table. Usages Directory service  Very little amount of information, such as URI, metadata, … Storage  Data, such.
Tapestry: A Resilient Global-scale Overlay for Service Deployment 1 Ben Y. Zhao, Ling Huang, Jeremy Stribling, Sean C. Rhea, Anthony D. Joseph, and John.
PROP: A Scalable and Reliable P2P Assisted Proxy Streaming System Computer Science Department College of William and Mary Lei Guo, Songqing Chen, and Xiaodong.
Peer to Peer Network Design Discovery and Routing algorithms
CS 6401 Overlay Networks Outline Overlay networks overview Routing overlays Resilient Overlay Networks Content Distribution Networks.
Tapestry : An Infrastructure for Fault-tolerant Wide-area Location and Routing Presenter : Lee Youn Do Oct 5, 2005 Ben Y.Zhao, John Kubiatowicz, and Anthony.
Improving Fault Tolerance in AODV Matthew J. Miller Jungmin So.
1 Plaxton Routing. 2 History Greg Plaxton, Rajmohan Rajaraman, Andrea Richa. Accessing nearby copies of replicated objects, SPAA 1997 Used in several.
Incrementally Improving Lookup Latency in Distributed Hash Table Systems Hui Zhang 1, Ashish Goel 2, Ramesh Govindan 1 1 University of Southern California.
CS791Aravind Elango Maintenance-Free Global Data Storage Sean Rhea, Chris Wells, Patrick Eaten, Dennis Geels, Ben Zhao, Hakim Weatherspoon and John Kubiatowicz.
OceanStore: An Architecture for Global-Scale Persistent Storage
Accessing nearby copies of replicated objects
Outline Midterm results summary Distributed file systems – continued
John D. Kubiatowicz UC Berkeley
Tapestry: Scalable and Fault-tolerant Routing and Location
MIT LCS Proceedings of the 2001 ACM SIGCOMM Conference
DSDV Destination-Sequenced Distance-Vector Routing Protocol
Presentation transcript:

Tapestry : An Infrastructure for Fault-tolerant Wide-area Location and Routing Presenter: Chunyuan Liao March 6, 2002 Ben Y.Zhao, John Kubiatowicz, and Anthony D,Josephetc. Computer Science Division University of California, Berkeley

Outline  Challenges  System overview  Operations, concerned issues & solutions Route Route Locate Locate Publish Publish  Evaluation & Conclusion  Implementation  Summary & Comments Insert Insert Delete Delete Move Move

Project background  Driving force : Ubiquitous Computing  OceanStore – A data utility infrastructure  Goals: –Based on the current untrusted Infrastructure –Achieve Nomadic Data  Anytime, Anywhere  Highly scalable, reliable and fault-tolerant and fault-tolerant  Basic issues: –Data Location –Routing

Challenges  How to achieve naming, location and routing with a complex & chaotic computing environment  Dynamic nature –Mobile and replicated Data & Services –Complex interaction between components, even in motion  Traditional approaches –fail to address the extreme dynamic nature

Tapestry : An infrastructure for Fault-tolerant wide-area Location and Routing  An overlay Location & Routing infrastructure built on the IP  Features –Highly scalable : Decentralized, Point-2-PointSelf-Organizing –Highly fault-tolerant: Redundancy, Adaptation –Good locality Content-based routing&location –Highly durable

Basic Model of Tapestry  Originated in Plaxton Scheme  Basic components: –Nodes ServersRoutersClients –Objects Data or Services –Link Point-2-Point link

Operations in Trapestry  Naming  Routing  Object Location  Publishing Objects  Inserting/Deleting Objects  Mobile Objects

Tapestry - Naming  Node ID/Object ID –A fixed length bit string (4 bits in each level ) 84F8, F8, 9098 –Global –Randomly generated –Location-Independent –Even distributed –Not unique ( shared by replicas )

Routing : Rules  Suffix matching ( similar to Plaxton ) –Incrementally routing digital by digital –Maximum hops : log b (N) 6789 B4F Msg to 4598 B437

Routing : Neighbor maps A table with b*log b (N) entries The i-th level neighbor share (i-1) suffix chunks Entry( i, j ) Pointer to the neighbor “ j” + (i-1) suffix Secondary Neighbors Back Pointers Create bi-direction link 0642

Routing : Fault-tolerant  Detect Server/Link failure –TCP time out( Ping ) –Periodic “heart beat” msg along back pointers  Resist fault –Secondary neighbor  Recover –Probing message –Second Chance

Locating : basic procedure  4 phrases locating –Map the Object ID to a “virtual” Node ID –Route the request to that node –Arrive the surrogate or“root for the object –Direct to the server Client : B4F8 Server : B F734 B Surrogate Routing

Locating : Surrogate Routing(1)  Given any client at different place, how to find the same “root”? –Plaxton 1.Find the nodes with the maximum matching suffix (Stop at the empty entry in neighbor map) 2.Order them with the global knowledge 3.Choose the No.1 –Tapestry 1.Go further than Plaxton( choose an alternate entry ) 2.Stop at a neighbor map where there is only one non-empty entry pointed to node R 3.R is the root

Locating : Surrogate Routing(2) F3145 E B7645 B B B1145 Assumption: 1.Every node is reachable Ensure the same “patterns” 2.Even distributed ID Ensure less and less nodes in mapping table Conclusion: 1. Root can always be found 2. E. of Sur. Route is 2

Publishing  Similar to locating 1.Server send msg and pretends to locate the object 2.Find the surrogate node as the “root” for the Obj. 3.Save the related info there, such as 3.Save the related info there, such as Server :B4F F734 B Surrogate Routing

Locating/Publishing : Fault-Tolerant & Locality  Multiple “root” (better than Plaxton) –Map the Obj. ID to several “root” –Publish/Locate can be executed simultaneously  Cache 2-tuple  Cache 2-tuple –Clients can get the on the way to the root –Intermediate notes can receive multiple for the same Obj., the nearest one is chosen

Insert a new node: basic procedure 1.Get an Node ID 2.Begin with a “Gateway node” G 3.Pretends to route to itself 4.Establish nearly optimal neighbor map during the “pseudo routing” by coping & Choosing nearest ones. 5.Go back and notify neighbors Gateway node : B4F F734 B Surrogate Routing New node : 1234

Delete a note Most simple operation  Explicitly notify the neighbors with back pointers  Use Soft sate Don’t send “heart beat” messages and republish messages any more

Maintain System Consistency  Components in a Tapestry node –Neighbor map –Back pointers –Object-Location pointers –Hotspot Monitor –Object store  Main correct status –Soft sate –Proactive explicit update

Soft state  Advantage –Easy to implement –Suited to slowly changing systems  Disadvantage –Tradeoff between bandwidth overhead and level of consistency –Not suited to the fast changing systems –Example : Bytes for the republishing for a server can be 1400MB (!) in a single interval.

Proactive explicit update ( PEU )  Proactive explicit updates –Epoch number  sequence # of the rounds –Expanded 3-tuple    Soft state : backup resort

PEU : Node Mobility Root C D E F A * * B Move Object 123 from A to B Republishing (123,B) Deleting (123,A) with “LostHopID”

PEU : Recover location pointers Root E F C A Server B Exiting Notification D Reconstruction (O,S,B) Deleting Old Data

Introspective Optimization : A dapting to the changing environment  Load balance 1.Periodically Ping by refresher thread 2.Update neighbor pointers  Hotspot 1.Find the source of the heavy traffic, “Hotspot” 2.Pub the desired data near the hotspot

Evaluation  Gain –Good Locality –Low Location latency –High Stability –High Fault-tolerence  Cost –Bandwidth overhead linear to the replicas

Implementation  Packet level simulators are finished in C  Used to support other applications –such as OceanStore –Bayeus, application-level multicast protocol  Future Working –Security issues –Mobile-IP like functionality

Summary  Urgent need for new Location/Routing Scheme  Features of Tapestry –Location-independent naming –Integration of location and routing –Content-based routing –Support for the dynamic environment inserting/deleting/moving Node/Object

Comments and Questions  Paradox or discrepancy? The underlying IP has bad scalability, how can Tapestry achieve high scalability? Just for demo!  What’s the relation between the IP and Tapestry? Tapestry doesn’t intend to replace IP, it just tries to establish a higher level locating & routing infrastructure to support the content- based operation.  How can we achieve the same goal without IP?