Peer-to-Peer 15-441.

Slides:



Advertisements
Similar presentations
P2P data retrieval DHT (Distributed Hash Tables) Partially based on Hellerstein’s presentation at VLDB2004.
Advertisements

Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan MIT and Berkeley presented by Daniel Figueiredo Chord: A Scalable Peer-to-peer.
Peer to Peer and Distributed Hash Tables
Clayton Sullivan PEER-TO-PEER NETWORKS. INTRODUCTION What is a Peer-To-Peer Network A Peer Application Overlay Network Network Architecture and System.
L-19 P2P. Scaling Problem Millions of clients  server and network meltdown 2.
Application Layer Overlays IS250 Spring 2010 John Chuang.
P2p, Spring 05 1 Topics in Database Systems: Data Management in Peer-to-Peer Systems March 29, 2005.
Peer-to-Peer Jeff Pang Spring Spring 2004, Jeff Pang2 Intro Quickly grown in popularity –Dozens or hundreds of file sharing applications.
Topics in Reliable Distributed Systems Lecture 2, Fall Dr. Idit Keidar.
Peer-to-Peer Credits: Slides adapted from J. Pang, B. Richardson, I. Stoica, M. Cuenca, B. Cohen, S. Ramesh, D. Qui, R. Shrikant, Xiaoyan Li and R. Martin.
P2P: Advanced Topics Filesystems over DHTs and P2P research Vyas Sekar.
Chord-over-Chord Overlay Sudhindra Rao Ph.D Qualifier Exam Department of ECECS.
Freenet A Distributed Anonymous Information Storage and Retrieval System I Clarke O Sandberg I Clarke O Sandberg B WileyT W Hong.
Topics in Reliable Distributed Systems Fall Dr. Idit Keidar.
1 CS 194: Distributed Systems Distributed Hash Tables Scott Shenker and Ion Stoica Computer Science Division Department of Electrical Engineering and Computer.
1 Seminar: Information Management in the Web Gnutella, Freenet and more: an overview of file sharing architectures Thomas Zahn.
Peer To Peer Distributed Systems Pete Keleher. Why Distributed Systems? l Aggregate resources! –memory –disk –CPU cycles l Proximity to physical stuff.
Wide-area cooperative storage with CFS
Peer-to-Peer Networks Slides largely adopted from Ion Stoica’s lecture at UCB.
Peer-to-Peer Networking Credit slides from J. Pang, B. Richardson, I. Stoica.
File Sharing : Hash/Lookup Yossi Shasho (HW in last slide) Based on Chord: A Scalable Peer-to-peer Lookup Service for Internet ApplicationsChord: A Scalable.
Winter 2008 P2P1 Peer-to-Peer Networks: Unstructured and Structured What is a peer-to-peer network? Unstructured Peer-to-Peer Networks –Napster –Gnutella.
1 Freenet  Addition goals to file location: -Provide publisher anonymity, security -Resistant to attacks – a third party shouldn’t be able to deny the.
Peer-to-Peer Scaling Problem Millions of clients  server and network meltdown.
1CS 6401 Peer-to-Peer Networks Outline Overview Gnutella Structured Overlays BitTorrent.
Peer-to-Peer Outline p2p file sharing techniques –Downloading: Whole-file vs. chunks –Searching Centralized index (Napster, etc.) Flooding (Gnutella,
CS 640: Introduction to Computer Networks Yu-Chi Lai Lecture 18 - Peer-to-Peer.
Distributed Systems Lecture 21 – CDN & Peer-to-Peer.
P2P File Sharing Systems
INTRODUCTION TO PEER TO PEER NETWORKS Z.M. Joseph CSE 6392 – DB Exploration Spring 2006 CSE, UT Arlington.
Freenet. Anonymity  Napster, Gnutella, Kazaa do not provide anonymity  Users know who they are downloading from  Others know who sent a query  Freenet.
1 Napster & Gnutella An Overview. 2 About Napster Distributed application allowing users to search and exchange MP3 files. Written by Shawn Fanning in.
Introduction Widespread unstructured P2P network

Content Overlays (Nick Feamster). 2 Content Overlays Distributed content storage and retrieval Two primary approaches: –Structured overlay –Unstructured.
Introduction of P2P systems
Peer to Peer Research survey TingYang Chang. Intro. Of P2P Computers of the system was known as peers which sharing data files with each other. Build.
Chord: A Scalable Peer-to-peer Lookup Protocol for Internet Applications Xiaozhou Li COS 461: Computer Networks (precept 04/06/12) Princeton University.
1 Distributed Hash Tables (DHTs) Lars Jørgen Lillehovde Jo Grimstad Bang Distributed Hash Tables (DHTs)
Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan MIT and Berkeley presented by Daniel Figueiredo Chord: A Scalable Peer-to-peer.
Peer-to-Peer File Sharing Jennifer Rexford COS 461: Computer Networks Lectures: MW 10-10:50am in Architecture N101
SIGCOMM 2001 Lecture slides by Dr. Yingwu Zhu Chord: A Scalable Peer-to-peer Lookup Service for Internet Applications.
1 Peer-to-Peer Technologies Seminar by: Kunal Goswami (05IT6006) School of Information Technology Guided by: Prof. C.R.Mandal, School of Information Technology.
PEER TO PEER (P2P) NETWORK By: Linda Rockson 11/28/06.
15-744: Computer Networking L-22: P2P. Lecture 22: Peer-to-Peer Networks Typically each member stores/provides access to content Has quickly.
CS 640: Introduction to Computer Networks Aditya Akella Lecture 24 - Peer-to-Peer.
1 Secure Peer-to-Peer File Sharing Frans Kaashoek, David Karger, Robert Morris, Ion Stoica, Hari Balakrishnan MIT Laboratory.
Computer Networking P2P. Why P2P? Scaling: system scales with number of clients, by definition Eliminate centralization: Eliminate single point.
EE324 DISTRIBUTED SYSTEMS L17 P2P. Scaling Problem 2  Millions of clients  server and network meltdown.
Hui Zhang, Fall Computer Networking CDN and P2P.
Peer-to-Peer File Sharing
Algorithms and Techniques in Structured Scalable Peer-to-Peer Networks
LOOKING UP DATA IN P2P SYSTEMS Hari Balakrishnan M. Frans Kaashoek David Karger Robert Morris Ion Stoica MIT LCS.
CS Spring 2014 CS 414 – Multimedia Systems Design Lecture 37 – Introduction to P2P (Part 1) Klara Nahrstedt.
Two Peer-to-Peer Networking Approaches Ken Calvert Net Seminar, 23 October 2001 Note: Many slides “borrowed” from S. Ratnasamy’s Qualifying Exam talk.
INTERNET TECHNOLOGIES Week 10 Peer to Peer Paradigm 1.
P2P Search COP6731 Advanced Database Systems. P2P Computing  Powerful personal computer Share computing resources P2P Computing  Advantages: Shared.
P2P Search COP P2P Search Techniques Centralized P2P systems  e.g. Napster, Decentralized & unstructured P2P systems  e.g. Gnutella.
15-829A/18-849B/95-811A/19-729A Internet-Scale Sensor Systems: Design and Policy Review.
Distributed Computing Systems Peer-to-Peer. Definition of Peer-to-Peer (P2P) Significant autonomy from central servers Exploits resources at edges of.
Chord: A Scalable Peer-to-Peer Lookup Service for Internet Applications * CS587x Lecture Department of Computer Science Iowa State University *I. Stoica,
CS Spring 2010 CS 414 – Multimedia Systems Design Lecture 24 – Introduction to Peer-to-Peer (P2P) Systems Klara Nahrstedt (presented by Long Vu)
Distributed Hash Tables. DHTs ● Like it sounds – a distributed hash table ● Put(Key, Value) ● Get(Key) -> Value.
(some slides from D. Anderson, K. Harras and others)
CS 268: Lecture 22 (Peer-to-Peer Networks)
Distributed Systems CDN & Peer-to-Peer.
مظفر بگ محمدی دانشگاه ایلام
EE 122: Peer-to-Peer (P2P) Networks
DHT Routing Geometries and Chord
#02 Peer to Peer Networking
Presentation transcript:

Peer-to-Peer 15-441

Scaling Problem Millions of clients  server and network meltdown

P2P System Leverage the resources of client machines (peers) App end-point vs. Infrastructure vs. waypoints Transition: infrastructure -> remove them Leverage the resources of client machines (peers) Computation, storage, bandwidth

Why p2p? Harness lots of spare capacity 1 Big Fast Server: 1Gbit/s, $10k/month++ 2,000 cable modems: 1Gbit/s, $ ?? 1M end-hosts: Uh, wow. Build self-managing systems / Deal with huge scale Same techniques attractive for both companies / servers / p2p E.g., Akamai’s 14,000 nodes Google’s 100,000+ nodes

Outline p2p file sharing techniques Downloading: Whole-file vs. chunks Searching Centralized index (Napster, etc.) Flooding (Gnutella, etc.) Smarter flooding (KaZaA, …) Routing (Freenet, etc.) Uses of p2p - what works well, what doesn’t? servers vs. arbitrary nodes Hard state (backups!) vs soft-state (caches) Challenges Fairness, freeloading, security, …

P2p file-sharing Quickly grown in popularity Dozens or hundreds of file sharing applications 35 million American adults use P2P networks -- 29% of all Internet users in US! Audio/Video transfer now dominates traffic on the Internet

What’s out there? Central Flood Super-node flood Route Whole File Napster Gnutella Freenet Chunk Based BitTorrent KaZaA (bytes, not chunks) DHTs eDonkey2000

Searching N2 N1 N3 ? N4 N6 N5 Key=“title” Value=MP3 data… Internet Client Publisher 1000s of nodes. Set of nodes may change… Lookup(“title”) N4 N6 N5

Searching 2 Needles vs. Haystacks Search expressiveness Searching for top 40, or an obscure punk track from 1981 that nobody’s heard of? Search expressiveness Whole word? Regular expressions? File names? Attributes? Whole-text search? (e.g., p2p gnutella or p2p google?)

Framework Common Primitives: Join: how to I begin participating? Publish: how do I advertise my file? Search: how to I find a file? Fetch: how to I retrieve a file?

Next Topic... Centralized Database Query Flooding Napster Query Flooding Gnutella Intelligent Query Flooding KaZaA Swarming BitTorrent Unstructured Overlay Routing Freenet Structured Overlay Routing Distributed Hash Tables

Napster: History 1999: Sean Fanning launches Napster Peaked at 1.5 million simultaneous users Jul 2001: Napster shuts down

Napster: Overiew Centralized Database: Join: on startup, client contacts central server Publish: reports list of files to central server Search: query the server => return someone that stores the requested file Fetch: get the file directly from peer

Napster: Publish insert(X, 123.2.21.23) ... Publish I have X, Y, and Z! 123.2.21.23

Napster: Search 123.2.0.18 Fetch search(A) --> 123.2.0.18 Query Reply Where is file A?

Napster: Discussion Pros: Simple Search scope is O(1) Controllable (pro or con?) Cons: Server maintains O(N) State Server does all processing Single point of failure

Next Topic... Centralized Database Query Flooding Napster Query Flooding Gnutella Intelligent Query Flooding KaZaA Swarming BitTorrent Unstructured Overlay Routing Freenet Structured Overlay Routing Distributed Hash Tables

Gnutella: History In 2000, J. Frankel and T. Pepper from Nullsoft released Gnutella Soon many other clients: Bearshare, Morpheus, LimeWire, etc. In 2001, many protocol enhancements including “ultrapeers”

Gnutella: Overview Query Flooding: Join: on startup, client contacts a few other nodes; these become its “neighbors” Publish: no need Search: ask neighbors, who ask their neighbors, and so on... when/if found, reply to sender. TTL limits propagation Fetch: get the file directly from peer

Gnutella: Search I have file A. Reply Query Where is file A?

Gnutella: Discussion Pros: Cons: Fully de-centralized Search cost distributed Processing @ each node permits powerful search semantics Cons: Search scope is O(N) Search time is O(???) Nodes leave often, network unstable TTL-limited search works well for haystacks. For scalability, does NOT search every node. May have to re-issue query later

KaZaA: History In 2001, KaZaA created by Dutch company Kazaa BV Single network called FastTrack used by other clients as well: Morpheus, giFT, etc. Eventually protocol changed so other clients could no longer talk to it Most popular file sharing network today with >10 million users (number varies)

KaZaA: Overview “Smart” Query Flooding: Join: on startup, client contacts a “supernode” ... may at some point become one itself Publish: send list of files to supernode Search: send query to supernode, supernodes flood query amongst themselves. Fetch: get the file directly from peer(s); can fetch simultaneously from multiple peers

KaZaA: Network Design “Super Nodes”

KaZaA: File Insert insert(X, 123.2.21.23) ... Publish I have X!

KaZaA: File Search search(A) --> 123.2.0.18 123.2.22.50 Replies Query Where is file A?

KaZaA: Fetching More than one node may have requested file... How to tell? Must be able to distinguish identical files Not necessarily same filename Same filename not necessarily same file... Use Hash of file KaZaA uses UUHash: fast, but not secure Alternatives: MD5, SHA-1 How to fetch? Get bytes [0..1000] from A, [1001...2000] from B Alternative: Erasure Codes

KaZaA: Discussion Pros: Cons: Tries to take into account node heterogeneity: Bandwidth Host Computational Resources Host Availability (?) Rumored to take into account network locality Cons: Mechanisms easy to circumvent Still no real guarantees on search scope or search time Similar behavior to gnutella, but better.

Stability and Superpeers Why superpeers? Query consolidation Many connected nodes may have only a few files Propagating a query to a sub-node would take more b/w than answering it yourself Caching effect Requires network stability Superpeer selection is time-based How long you’ve been on is a good predictor of how long you’ll be around.

BitTorrent: History In 2002, B. Cohen debuted BitTorrent Key Motivation: Popularity exhibits temporal locality (Flash Crowds) E.g., Slashdot effect, CNN on 9/11, new movie/game release Focused on Efficient Fetching, not Searching: Distribute the same file to all peers Single publisher, multiple downloaders Has some “real” publishers: Blizzard Entertainment using it to distribute the beta of their new game

BitTorrent: Overview Swarming: Big differences from Napster: Join: contact centralized “tracker” server, get a list of peers. Publish: Run a tracker server. Search: Out-of-band. E.g., use Google to find a tracker for the file you want. Fetch: Download chunks of the file from your peers. Upload chunks you have to them. Big differences from Napster: Chunk based downloading (sound familiar? :) “few large files” focus Anti-freeloading mechanisms

BitTorrent: Publish/Join Tracker

BitTorrent: Fetch

BitTorrent: Sharing Strategy Employ “Tit-for-tat” sharing strategy A is downloading from some other people A will let the fastest N of those download from him Be optimistic: occasionally let freeloaders download Otherwise no one would ever start! Also allows you to discover better peers to download from when they reciprocate Let N peop Goal: Pareto Efficiency Game Theory: “No change can make anyone better off without making others worse off” Does it work? (don’t know!)

BitTorrent: Summary Pros: Cons: Works reasonably well in practice Gives peers incentive to share resources; avoids freeloaders Cons: Pareto Efficiency relative weak condition Central tracker server needed to bootstrap swarm (Tracker is a design choice, not a requirement, as you know from your projects. Could easily combine with other approaches.)

Next Topic... Centralized Database Query Flooding Napster Query Flooding Gnutella Intelligent Query Flooding KaZaA Swarming BitTorrent Unstructured Overlay Routing Freenet Structured Overlay Routing Distributed Hash Tables (DHT)

Distributed Hash Tables Academic answer to p2p Goals Guatanteed lookup success Provable bounds on search time Provable scalability Makes some things harder Fuzzy queries / full-text search / etc. Read-write, not read-only Hot Topic in networking since introduction in ~2000/2001

DHT: Overview Abstraction: a distributed “hash-table” (DHT) data structure: put(id, item); item = get(id); Implementation: nodes in system form a distributed data structure Can be Ring, Tree, Hypercube, Skip List, Butterfly Network, ...

DHT: Overview (2) Structured Overlay Routing: Join: On startup, contact a “bootstrap” node and integrate yourself into the distributed data structure; get a node id Publish: Route publication for file id toward a close node id along the data structure Search: Route a query for file id toward a close node id. Data structure guarantees that query will meet the publication. Fetch: Two options: Publication contains actual file => fetch from where query stops Publication says “I have file X” => query tells you 128.2.1.3 has X, use IP routing to get X from 128.2.1.3

DHT: Example - Chord Associate to each node and file a unique id in an uni-dimensional space (a Ring) E.g., pick from the range [0...2m] Usually the hash of the file or IP address Properties: Routing table size is O(log N) , where N is the total number of nodes Guarantees that a file is found in O(log N) hops from MIT in 2001

DHT: Consistent Hashing Key 5 K5 Node 105 N105 K20 Circular ID space N32 N90 K80 A key is stored at its successor: node with next higher ID

DHT: Chord Basic Lookup N120 N10 “Where is key 80?” N105 N32 “N90 has K80” Just need to make progress, and not overshoot. Will talk about initialization later. And robustness. Now, how about speed? K80 N90 N60

DHT: Chord “Finger Table” 1/4 1/2 1/8 1/16 1/32 1/64 1/128 N80 Small tables, but multi-hop lookup. Table entries: IP address and Chord ID. Navigate in ID space, route queries closer to successor. Log(n) tables, log(n) hops. Route to a document between ¼ and ½ … Entry i in the finger table of node n is the first node that succeeds or equals n + 2i In other words, the ith finger points 1/2n-i way around the ring

DHT: Chord Join Assume an identifier space [0..8] Node n1 joins 1 7 6 Succ. Table i id+2i succ 0 2 1 1 3 1 2 5 1 1 7 6 2 5 3 4

DHT: Chord Join Node n2 joins 1 7 6 2 5 3 4 Succ. Table i id+2i succ i id+2i succ 0 2 2 1 3 1 2 5 1 1 7 6 2 Succ. Table i id+2i succ 0 3 1 1 4 1 2 6 1 5 3 4

DHT: Chord Join Nodes n0, n6 join 1 7 6 2 5 3 4 Succ. Table i id+2i succ 0 1 1 1 2 2 2 4 0 Nodes n0, n6 join Succ. Table i id+2i succ 0 2 2 1 3 6 2 5 6 1 7 Succ. Table i id+2i succ 0 7 0 1 0 0 2 2 2 6 2 Succ. Table i id+2i succ 0 3 6 1 4 6 2 6 6 5 3 4

DHT: Chord Join Nodes: n1, n2, n0, n6 Items: f7, f2 1 7 6 2 5 3 4 Succ. Table Items i id+2i succ 0 1 1 1 2 2 2 4 0 7 Nodes: n1, n2, n0, n6 Items: f7, f2 Succ. Table Items 1 7 i id+2i succ 0 2 2 1 3 6 2 5 6 1 Succ. Table 6 2 i id+2i succ 0 7 0 1 0 0 2 2 2 Succ. Table i id+2i succ 0 3 6 1 4 6 2 6 6 5 3 4

DHT: Chord Routing Upon receiving a query for item id, a node: Succ. Table Items i id+2i succ 0 1 1 1 2 2 2 4 0 7 Upon receiving a query for item id, a node: Checks whether stores the item locally If not, forwards the query to the largest node in its successor table that does not exceed id Succ. Table Items 1 7 i id+2i succ 0 2 2 1 3 6 2 5 6 1 query(7) Succ. Table 6 2 i id+2i succ 0 7 0 1 0 0 2 2 2 Succ. Table i id+2i succ 0 3 6 1 4 6 2 6 6 5 3 4

DHT: Chord Summary Routing table size? Routing time? Log N fingers Each hop expects to 1/2 the distance to the desired id => expect O(log N) hops.

DHT: Discussion Pros: Cons: Guaranteed Lookup O(log N) per node state and search scope Cons: No one uses them? (only one file sharing app) Supporting non-exact match search is hard

When are p2p / DHTs useful? Caching and “soft-state” data Works well! BitTorrent, KaZaA, etc., all use peers as caches for hot data Finding read-only data Limited flooding finds hay DHTs find needles BUT

A Peer-to-peer Google? Complex intersection queries (“the” + “who”) Billions of hits for each term alone Sophisticated ranking Must compare many results before returning a subset to user Very, very hard for a DHT / p2p system Need high inter-node bandwidth (This is exactly what Google does - massive clusters)

Writable, persistent p2p Do you trust your data to 100,000 monkeys? Node availability hurts Ex: Store 5 copies of data on different nodes When someone goes away, you must replicate the data they held Hard drives are *huge*, but cable modem upload bandwidth is tiny - perhaps 10 Gbytes/day Takes many days to upload contents of 200GB hard drive. Very expensive leave/replication situation!

P2P: Summary Many different styles; remember pros and cons of each centralized, flooding, swarming, unstructured and structured routing Lessons learned: Single points of failure are very bad Flooding messages to everyone is bad Underlying network topology is important Not all nodes are equal Need incentives to discourage freeloading Privacy and security are important Structure can provide theoretical bounds and guarantees

Extra Slides

KaZaA: Usage Patterns KaZaA is more than one workload! Many files < 10MB (e.g., Audio Files) Many files > 100MB (e.g., Movies) from Gummadi et al., SOSP 2003

KaZaA: Usage Patterns (2) KaZaA is not Zipf! FileSharing: “Request-once” Web: “Request-repeatedly” from Gummadi et al., SOSP 2003

KaZaA: Usage Patterns (3) What we saw: A few big files consume most of the bandwidth Many files are fetched once per client but still very popular Solution? Caching! from Gummadi et al., SOSP 2003

Freenet: History In 1999, I. Clarke started the Freenet project Basic Idea: Employ Internet-like routing on the overlay network to publish and locate files Addition goals: Provide anonymity and security Make censorship difficult

Freenet: Overview Routed Queries: Join: on startup, client contacts a few other nodes it knows about; gets a unique node id Publish: route file contents toward the file id. File is stored at node with id closest to file id Search: route query for file id toward the closest node id Fetch: when query reaches a node containing file id, it returns the file to the sender

Freenet: Routing Tables id – file identifier (e.g., hash of file) next_hop – another node that stores the file id file – file identified by id being stored on the local node Forwarding of query for file id If file id stored locally, then stop Forward data back to upstream requestor If not, search for the “closest” id in the table, and forward the message to the corresponding next_hop If data is not found, failure is reported back Requestor then tries next closest match in routing table id next_hop file … …

Freenet: Routing query(10) n1 n2 1 4 n1 f4 12 n2 f12 5 n3 9 n3 f9 4’ 4

Freenet: Routing Properties “Close” file ids tend to be stored on the same node Why? Publications of similar file ids route toward the same place Network tend to be a “small world” Small number of nodes have large number of neighbors (i.e., ~ “six-degrees of separation”) Consequence: Most queries only traverse a small number of hops to find the file

Freenet: Anonymity & Security Randomly modify source of packet as it traverses the network Can use “mix-nets” or onion-routing Security & Censorship resistance No constraints on how to choose ids for files => easy to have to files collide, creating “denial of service” (censorship) Solution: have a id type that requires a private key signature that is verified when updating the file Cache file on the reverse path of queries/publications => attempt to “replace” file with bogus data will just cause the file to be replicated more!

Freenet: Discussion Pros: Cons: Intelligent routing makes queries relatively short Search scope small (only nodes along search path involved); no flooding Anonymity properties may give you “plausible deniability” Cons: Still no provable guarantees! Anonymity features make it hard to measure, debug