Self-service, with applications to distributed classifier construction Michael K. Reiter and Asad Samar April 27, 2006 Properties & Related Work Self-Service.

Slides:



Advertisements
Similar presentations
Dynamic Replica Placement for Scalable Content Delivery Yan Chen, Randy H. Katz, John D. Kubiatowicz {yanchen, randy, EECS Department.
Advertisements

CAN 1.Distributed Hash Tables a)DHT recap b)Uses c)Example – CAN.
1 Intrusion Monitoring of Malicious Routing Behavior Poornima Balasubramanyam Karl Levitt Computer Security Laboratory Department of Computer Science UCDavis.
Replication. Topics r Why Replication? r System Model r Consistency Models r One approach to consistency management and dealing with failures.
Clayton Sullivan PEER-TO-PEER NETWORKS. INTRODUCTION What is a Peer-To-Peer Network A Peer Application Overlay Network Network Architecture and System.
X stream Project proposal. Project goals: Students Students: Academic Supervisor Academic Supervisor: Advisors: Developing and Implementing a large scale.
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
1 Accessing nearby copies of replicated objects Greg Plaxton, Rajmohan Rajaraman, Andrea Richa SPAA 1997.
Small-world Overlay P2P Network
Distributed Systems Fall 2010 Replication Fall 20105DV0203 Outline Group communication Fault-tolerant services –Passive and active replication Highly.
Business Continuity and DR, A Practical Implementation Mich Talebzadeh, Consultant, Deutsche Bank
Probabilistic Aggregation in Distributed Networks Ling Huang, Ben Zhao, Anthony Joseph and John Kubiatowicz {hling, ravenben, adj,
P2P: Advanced Topics Filesystems over DHTs and P2P research Vyas Sekar.
Probabilistic Data Aggregation Ling Huang, Ben Zhao, Anthony Joseph Sahara Retreat January, 2004.
presented by Hasan SÖZER1 Scalable P2P Search Daniel A. Menascé George Mason University.
SCALLOP A Scalable and Load-Balanced Peer- to-Peer Lookup Protocol for High- Performance Distributed System Jerry Chou, Tai-Yi Huang & Kuang-Li Huang Embedded.
Efficient Parallelization for AMR MHD Multiphysics Calculations Implementation in AstroBEAR.
©Silberschatz, Korth and Sudarshan18.1Database System Concepts Centralized Systems Run on a single computer system and do not interact with other computer.
What Can Databases Do for Peer-to-Peer Steven Gribble, Alon Halevy, Zachary Ives, Maya Rodrig, Dan Suciu Presented by: Ryan Huebsch CS294-4 P2P Systems.
Distributed File System: Data Storage for Networks Large and Small Pei Cao Cisco Systems, Inc.
Amazon’s Dynamo System The material is taken from “Dynamo: Amazon’s Highly Available Key-value Store,” by G. DeCandia, D. Hastorun, M. Jampani, G. Kakulapati,
GrIDS -- A Graph Based Intrusion Detection System For Large Networks Paper by S. Staniford-Chen et. al.
Thesis Proposal Data Consistency in DHTs. Background Peer-to-peer systems have become increasingly popular Lots of P2P applications around us –File sharing,
1 The Design of a Robust Peer-to-Peer System Gisik Kwon Dept. of Computer Science and Engineering Arizona State University Reference: SIGOPS European Workshop.
Project Presentation Students: Yan Michalevsky Asaf Cidon Supervisors: Alexander Shraer Assoc. Prof. Idit Keidar.
1 Resilient and Coherence Preserving Dissemination of Dynamic Data Using Cooperating Peers Shetal Shah, IIT Bombay Kirthi Ramamritham, IIT Bombay Prashant.
High Throughput Computing on P2P Networks Carlos Pérez Miguel
Peer-to-Peer Distributed Shared Memory? Gabriel Antoniu, Luc Bougé, Mathieu Jan IRISA / INRIA & ENS Cachan/Bretagne France Dagstuhl seminar, October 2003.
Advanced Computer Networks Topic 2: Characterization of Distributed Systems.
The Replica Location Service The Globus Project™ And The DataGrid Project Copyright (c) 2002 University of Chicago and The University of Southern California.
Carnegie Mellon Increasing Intrusion Tolerance Via Scalable Redundancy Mike Reiter Natassa Ailamaki Greg Ganger Priya Narasimhan Chuck Cranor.
Copyright © George Coulouris, Jean Dollimore, Tim Kindberg This material is made available for private study and for direct.
DHT-based unicast for mobile ad hoc networks Thomas Zahn, Jochen Schiller Institute of Computer Science Freie Universitat Berlin 報告 : 羅世豪.
1 Distributed Hash Table CS780-3 Lecture Notes In courtesy of Heng Yin.
Research of P2P Architecture based on Cloud Computing Speaker : 吳靖緯 MA0G0101.
GLOBAL EDGE SOFTWERE LTD1 R EMOTE F ILE S HARING - Ardhanareesh Aradhyamath.
Plethora: Infrastructure and System Design. Introduction Peer-to-Peer (P2P) networks: –Self-organizing distributed systems –Nodes receive and provide.
PRIN WOMEN PROJECT Research Unit: University of Naples Federico II G. Ferraiuolo
Peer to Peer Network Design Discovery and Routing algorithms
1 Lecture 19: Scalable Protocols & Synch Topics: coherence protocols for distributed shared-memory multiprocessors and synchronization (Sections )
Chapter 7: Consistency & Replication IV - REPLICATION MANAGEMENT By Jyothsna Natarajan Instructor: Prof. Yanqing Zhang Course: Advanced Operating Systems.
Peer-to-Peer Systems: An Overview Hongyu Li. Outline  Introduction  Characteristics of P2P  Algorithms  P2P Applications  Conclusion.
CS Spring 2014 CS 414 – Multimedia Systems Design Lecture 37 – Introduction to P2P (Part 1) Klara Nahrstedt.
INTERNET TECHNOLOGIES Week 10 Peer to Peer Paradigm 1.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 1.
CMSC 691B Multi-Agent System A Scalable Architecture for Peer to Peer Agent by Naveen Srinivasan.
1 Plaxton Routing. 2 History Greg Plaxton, Rajmohan Rajaraman, Andrea Richa. Accessing nearby copies of replicated objects, SPAA 1997 Used in several.
Seminar On Rain Technology
CS Spring 2010 CS 414 – Multimedia Systems Design Lecture 24 – Introduction to Peer-to-Peer (P2P) Systems Klara Nahrstedt (presented by Long Vu)
Asstt. Professor Adeel Akram. Other Novel Routing Approaches Link reversal Aimed for highly dynamic networks Goal: to identify some path, as opposed.
The Biologically Inspired Distributed File System: An Emergent Thinker Instantiation Presented by Dr. Ying Lu.
rain technology (redundant array of independent nodes)
REPLICATION & LOAD BALANCING
ETHANE: TAKING CONTROL OF THE ENTERPRISE
Distributed Computing
CHAPTER 3 Architectures for Distributed Systems
Plethora: Infrastructure and System Design
Accessing nearby copies of replicated objects
Replication Middleware for Cloud Based Storage Service
DHT Routing Geometries and Chord
A Case for Mutual Notification
Distributed computing deals with hardware
Outline Announcements Lab2 Distributed File Systems 1/17/2019 COP5611.
Service-Oriented Computing: Semantics, Processes, Agents
Deterministic and Semantically Organized Network Topology
Distributed Database Management Systems
Outline Review of Quiz #1 Distributed File Systems 4/20/2019 COP5611.
ContinuStreaming: Achieving High Playback Continuity of Gossip-based Peer-to-Peer Streaming IPDPS 2008 LI Zhenhua Dept. Computer, Nanjing University.
Lecture 18: Coherence and Synchronization
Sisi Duan Assistant Professor Information Systems
Presentation transcript:

Self-service, with applications to distributed classifier construction Michael K. Reiter and Asad Samar April 27, 2006 Properties & Related Work Self-Service achieves three properties Scalability Single clusters are only incrementally scalable Consistency Most P2P systems do not achieve strong consistency properties Multiple clusters (one per object) lack consistency for multi-object operations Failure recovery DHTs supporting atomic operations do not recover from failures Self-Service recovers from failures but may lose recent operations Consistency Scalability Fault recovery P2P systems [read-only, eventual consistency] Single clusters (Dynamic CDNs) Self-Service Distributed directories, DHT with atomic operations Multiple clusters (e.g., Oceanstore) Self-Service is a different point in this design space Service ImplementationsOur approach Centralized implementation Centralized services Provides strong consistency, but Scales poorly Cluster implementation Cluster-based services Can provide strong consistency, but Requires adding resources to the cluster as number of clients increases Self-service Scales to a large client population (particularly compute scalability) Does not require adding resources to the server as clients increase Provides strong consistency guarantees Can recover from client failures … but best suited for services where Clients are trustworthy (ongoing work) and churn is low Service state can be decomposed into small objects such that client operations typically involve only a few The Self-Service approach Clients perform their own operations No resource increase required at server as clients increase Algorithms guarantee strong consistency properties Can recover from client failures, though at some cost Clients are arranged in a tree rooted at server Service state divided into small objects Each client retrieves the objects required for its operation And applies the operation locally ababababab `a’ gets the object from root and applies its operation, modifying the object. `b’ gets object from `a’ and applies its operation, modifying the object again. Algorithms Nodes arrange themselves in a logical distributed queue And take turns to apply their operations on the object a b c d e abcdabcdebcde a b c d e a b c d e b has requested from a, c from b and d from c. e now requests from d. a completes operation and transfers to b. If a node disconnects, any objects in its sub-tree are reconstituted from latest state seen at any connected node Connected nodes continue to make progress even if a node in the distributed queue gets disconnected p' p c a b d Object history p' p c a b d Object history Object is in the disconnected sub-tree Object reconstituted at the parent of disconnected node Motivation Building models for network traffic classification and intrusion detection From traffic seen at multiple vantage points Globally share information to build accurate models Centralized scenario Send raw data to server Server updates model Server makes new model available Scalability and privacy concerns Our “self-service” approach All participating networks placed in a distributed tree Participating network pulls model Participating network updates model Participating network makes new model available Scales well and raw data not shared with other networks Experiments Implemented algorithms and traffic modeling application Micro-benchmarks on PlanetLab Model building experiments (CPU intensive) on local cluster Compared with a centralized implementation of same service