Distributed Monitoring and Management Presented by: Ahmed Khurshid Abdullah Al-Nayeem CS 525 Spring 2009 Advanced Distributed Systems.

Slides:



Advertisements
Similar presentations
Dynamic Replica Placement for Scalable Content Delivery Yan Chen, Randy H. Katz, John D. Kubiatowicz {yanchen, randy, EECS Department.
Advertisements

Efficient Event-based Resource Discovery Wei Yan*, Songlin Hu*, Vinod Muthusamy +, Hans-Arno Jacobsen +, Li Zha* * Chinese Academy of Sciences, Beijing.
Software-defined networking: Change is hard Ratul Mahajan with Chi-Yao Hong, Rohan Gandhi, Xin Jin, Harry Liu, Vijay Gill, Srikanth Kandula, Mohan Nanduri,
Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan MIT and Berkeley presented by Daniel Figueiredo Chord: A Scalable Peer-to-peer.
Dynamo: Amazon's Highly Available Key-value Store Distributed Storage Systems CS presented by: Hussam Abu-Libdeh.
Epidemic Techniques Algorithms and Implementations.
Ranveer Chandra , Kenneth P. Birman Department of Computer Science
Gossip Scheduling for Periodic Streams in Ad-hoc WSNs Ercan Ucan, Nathanael Thompson, Indranil Gupta Department of Computer Science University of Illinois.
LightFlood: An Optimal Flooding Scheme for File Search in Unstructured P2P Systems Song Jiang, Lei Guo, and Xiaodong Zhang College of William and Mary.
1 Failure Recovery for Priority Progress Multicast Jung-Rung Han Supervisor: Charles Krasic.
SCAN: A Dynamic, Scalable, and Efficient Content Distribution Network Yan Chen, Randy H. Katz, John D. Kubiatowicz {yanchen, randy,
Scribe: A Large-Scale and Decentralized Application-Level Multicast Infrastructure Miguel Castro, Peter Druschel, Anne-Marie Kermarrec, and Antony L. T.
Web Caching Schemes1 A Survey of Web Caching Schemes for the Internet Jia Wang.
Dept. of Computer Science & Engineering, CUHK1 Trust- and Clustering-Based Authentication Services in Mobile Ad Hoc Networks Edith Ngai and Michael R.
CMPE 150- Introduction to Computer Networks 1 CMPE 150 Fall 2005 Lecture 22 Introduction to Computer Networks.
Dissemination protocols for large sensor networks Fan Ye, Haiyun Luo, Songwu Lu and Lixia Zhang Department of Computer Science UCLA Chien Kang Wu.
Application Layer Multicast
presented by Hasan SÖZER1 Scalable P2P Search Daniel A. Menascé George Mason University.
SCALLOP A Scalable and Load-Balanced Peer- to-Peer Lookup Protocol for High- Performance Distributed System Jerry Chou, Tai-Yi Huang & Kuang-Li Huang Embedded.
Chord-over-Chord Overlay Sudhindra Rao Ph.D Qualifier Exam Department of ECECS.
CS218 – Final Project A “Small-Scale” Application- Level Multicast Tree Protocol Jason Lee, Lih Chen & Prabash Nanayakkara Tutor: Li Lao.
Adaptive Web Caching Lixia Zhang, Sally Floyd, and Van Jacob-son. In the 2nd Web Caching Workshop, Boulder, Colorado, April 25, System Laboratory,
Epidemic Techniques Chiu Wah So (Kelvin). Database Replication Why do we replicate database? – Low latency – High availability To achieve strong (sequential)
1 Spring Semester 2007, Dept. of Computer Science, Technion Internet Networking recitation #5 Mobile Ad-Hoc Networks TBRPF.
Wide-area cooperative storage with CFS
CS 580S Sensor Networks and Systems Professor Kyoung Don Kang Lecture 7 February 13, 2006.
11 SERVER CLUSTERING Chapter 6. Chapter 6: SERVER CLUSTERING2 OVERVIEW  List the types of server clusters.  Determine which type of cluster to use for.
1 CS 525 Advanced Distributed Systems Spring 2011 Indranil Gupta (Indy) Distributed Monitoring April 5, 2011 All Slides © IG Acknowledgments: Some slides.
Roger ZimmermannCOMPSAC 2004, September 30 Spatial Data Query Support in Peer-to-Peer Systems Roger Zimmermann, Wei-Shinn Ku, and Haojun Wang Computer.
Communication (II) Chapter 4
Collaborative Content Delivery Werner Vogels Robbert van Renesse, Ken Birman Dept. of Computer Science, Cornell University A peer-to-peer solution for.
Chord & CFS Presenter: Gang ZhouNov. 11th, University of Virginia.
5.1 Tanenbaum & Van Steen, Distributed Systems: Principles and Paradigms, 2e, (c) 2007 Prentice-Hall, Inc. All rights reserved DISTRIBUTED.
Overcast: Reliable Multicasting with an Overlay Network CS294 Paul Burstein 9/15/2003.
Tsunami: Maintaining High Bandwidth Under Dynamic Network Conditions Dejan Kostić, Ryan Braud, Charles Killian, Eric Vandekieft, James W. Anderson, Alex.
Adaptive Web Caching CS411 Dynamic Web-Based Systems Flying Pig Fei Teng/Long Zhao/Pallavi Shinde Computer Science Department.
Load-Balancing Routing in Multichannel Hybrid Wireless Networks With Single Network Interface So, J.; Vaidya, N. H.; Vehicular Technology, IEEE Transactions.
The Replica Location Service The Globus Project™ And The DataGrid Project Copyright (c) 2002 University of Chicago and The University of Southern California.
November 17, 2015Department of Computer Sciences, UT Austin1 SDIMS: A Scalable Distributed Information Management System Praveen Yalagandula Mike Dahlin.
Scalable Content- Addressable Networks Prepared by Kuhan Paramsothy March 5, 2007.
Paper Survey of DHT Distributed Hash Table. Usages Directory service  Very little amount of information, such as URI, metadata, … Storage  Data, such.
Scalable Self-Repairing Publish/Subscribe Robbert van Renesse Ken Birman Werner Vogels Cornell University.
Intradomain Traffic Engineering By Behzad Akbari These slides are based in part upon slides of J. Rexford (Princeton university)
1 Secure Peer-to-Peer File Sharing Frans Kaashoek, David Karger, Robert Morris, Ion Stoica, Hari Balakrishnan MIT Laboratory.
1. Outline  Introduction  Different Mechanisms Broadcasting Multicasting Forward Pointers Home-based approach Distributed Hash Tables Hierarchical approaches.
1 An Efficient, Low-Cost Inconsistency Detection Framework for Data and Service Sharing in an Internet-Scale System Yijun Lu †, Hong Jiang †, and Dan Feng.
Plethora: Infrastructure and System Design. Introduction Peer-to-Peer (P2P) networks: –Self-organizing distributed systems –Nodes receive and provide.
By: Gang Zhou Computer Science Department University of Virginia 1 Medians and Beyond: New Aggregation Techniques for Sensor Networks CS851 Seminar Presentation.
Chapter 7: Consistency & Replication IV - REPLICATION MANAGEMENT By Jyothsna Natarajan Instructor: Prof. Yanqing Zhang Course: Advanced Operating Systems.
LOOKING UP DATA IN P2P SYSTEMS Hari Balakrishnan M. Frans Kaashoek David Karger Robert Morris Ion Stoica MIT LCS.
NCLAB 1 Supporting complex queries in a distributed manner without using DHT NodeWiz: Peer-to-Peer Resource Discovery for Grids Sujoy Basu, Sujata Banerjee,
09/13/04 CDA 6506 Network Architecture and Client/Server Computing Peer-to-Peer Computing and Content Distribution Networks by Zornitza Genova Prodanoff.
Tanenbaum & Van Steen, Distributed Systems: Principles and Paradigms, 2e, (c) 2007 Prentice-Hall, Inc. All rights reserved DISTRIBUTED SYSTEMS.
Group Communication A group is a collection of users sharing some common interest.Group-based activities are steadily increasing. There are many types.
Tanenbaum & Van Steen, Distributed Systems: Principles and Paradigms, 2e, (c) 2007 Prentice-Hall, Inc. All rights reserved DISTRIBUTED SYSTEMS.
The Tail at Scale Dean and Barroso, CACM Feb 2013 Presenter: Chien-Ying Chen (cchen140) CS538 - Advanced Computer Networks 1.
Chord: A Scalable Peer-to-Peer Lookup Service for Internet Applications * CS587x Lecture Department of Computer Science Iowa State University *I. Stoica,
Plethora: Infrastructure and System Design
Chapter 7: Consistency & Replication IV - REPLICATION MANAGEMENT -Sumanth Kandagatla Instructor: Prof. Yanqing Zhang Advanced Operating Systems (CSC 8320)
SCOPE: Scalable Consistency in Structured P2P Systems
CS5412: Using Gossip to Build Overlay Networks
5.2 FLAT NAMING.
Cloud computing mechanisms
CS5412: Using Gossip to Build Overlay Networks
Replica Placement Model: We consider objects (and don’t worry whether they contain just data or code, or both) Distinguish different processes: A process.
Dynamic Replica Placement for Scalable Content Delivery
Combinatorial Optimization of Multicast Key Management
Other Routing Protocols
Design and Implementation of OverLay Multicast Tree Protocol
CS5412: Using Gossip to Build Overlay Networks
Presentation transcript:

Distributed Monitoring and Management Presented by: Ahmed Khurshid Abdullah Al-Nayeem CS 525 Spring 2009 Advanced Distributed Systems

Large Distributed Systems Infrastructure PlanetLab: 971 nodes in 485 sites Application Hadoop at Yahoo!: 4000 nodes 2 Google probably has more than 450,000 servers worldwide (Wikipedia) Not only nodes, data processed in commercial systems, e.g. Facebook is enormous (over 10 billion picture uploaded). 3/12/2009Department of Computer Science, UIUC

Monitoring and Management Monitoring and management of both infrastructures and applications – Corrective measures against failure, attacks, etc. – Ensuring better performance, e.g. load balancing What resources are managed? – Distributed application processes, objects (log files, routing table, etc.) – System resources: CPU utilization, free disk space, bandwidth utilization 33/12/2009Department of Computer Science, UIUC

Management and Monitoring Operations Query current system status – CPU utilization, disk space.. – Process progress rate.. Push software updates – Install the query program Monitor dynamically changing state ? n1 n2 n3 n4 n5 n6 43/12/2009Department of Computer Science, UIUC

Challenges Managing today’s large-scale systems is difficult. – A centralized solution doesn’t scale (no in-network aggregation) – Self-organization capability is becoming a necessity – Responses are expected in seconds, not in minutes/hours. – Node failure causes inconsistent results (network partition) Brewer’s conjecture: – It is impossible for a web service to provide the following three guarantees: Consistency, Availability, Partition- tolerance (CAP Dilemma) 53/12/2009Department of Computer Science, UIUC

Astrolabe: A Robust and Scalable Technology for Distributed System Monitoring, Management, and Data Mining Presented by: Abdullah Al-Nayeem Robbert Van Renesse, Kenneth P. Birman, and Werner Vogels

Overview Astrolabe as an information management service. – Locates and collects the status of a set of servers. – Reports the summaries of this information (aggregation mechanism using SQL). – Automatically updates and reports any changed summaries. Design principles: – Scalability through hierarchy of resources – Robustness through gossip protocol (p2p) – Flexibility through customization of queries (SQL) – Security through certificates 73/12/2009Department of Computer Science, UIUC

n8 Astrolabe Zone Hierarchy n1 n2 n3 n4 n5 n7 n6 /berkeley/cornell /uiuc /uiuc/cs Zone: 83/12/2009Department of Computer Science, UIUC

Astrolabe Zone Hierarchy (2) /uiuc/ece/n1/uiuc/cs/n4/uiuc/cs/n6/cornell/n2/cornell/cs/n3 /berkeley /eecs/n5 /berkeley /eecs/n7 /berkeley /eecs/n8 /uiuc/ece /uiuc/cs /cornell/cs /berkeley/eecs /uiuc/cornell/berkeley / It’s a virtual hierarchy. Only the host in leaf zone runs an Astrolabe agent - Zone hierarchy is determined by the administrators (less flexibility). - Assumption: zone names are consistent with the physical topology. - Zone hierarchy is determined by the administrators (less flexibility). - Assumption: zone names are consistent with the physical topology. 93/12/2009Department of Computer Science, UIUC

Decentralized Hierarchical Database An attribute list is associated with each zone. – This attribute list is defined as Management Information Base (MIB) Attributes includes information on load, total free disk space, process information, etc. Each internal zone has a relational table of MIBs of its child zones. – The leaf zone is an exception (.. next slide) 103/12/2009Department of Computer Science, UIUC

Decentralized Hierarchical Database (2) /uiuc/ece/n1/uiuc/cs/n4/uiuc/cs/n6 /uiuc/ece /uiuc/cs /uiuc /cornell /berkeley / cs ece uiuc cornell berkeley 11 …… n4 n6 Load = 0.1 Load = 0.3 system process Load = 0.1 Disk = 1.2TB Service: A(1.1) progress = 0.7 files system process Load = 0.3 Disk = 0.6TB Service: A(1.0) progress = 0.5 files 3/12/2009Department of Computer Science, UIUC Agent (/uiuc/cs/n6) has its local copy of these management table of MIBs. Agent (/uiuc/cs/n6) has its local copy of these management table of MIBs.

State Aggregation uiuc cornell berkeley /uiuc/cs/n4 /uiuc/cs /uiuc / cs ece n4 n6 Load = 0.3 Load = 0.5 (Own) SELECT MIN(Load) as Load Load = 0.3 Time = 121 Time = 101 Time = 130 Other aggregation includes: MAX (attribute) SUM (attribute) AVG (attribute) FIRST(n, attribute) 12 Aggregates the result using SQL query 3/12/2009Department of Computer Science, UIUC

State Merge – Gossip Protocol uiuc cornell berkeley /uiuc/cs/n4 /uiuc/cs /uiuc / cs ece n4 n6 Load = 0.3 Load = 0.5 Load = 0.3 (Own) /uiuc/cs/n6 /uiuc/cs /uiuc / Time = 121 Time = 101 uiuc cornell berkeley cs ece n4 n6 Load = 0.5 (Own) Time = 101 Each agent periodically contacts some other agent and exchanges the state associated with MIB based on timestamp. Time = 130Time = 110 Time = 121 Load = /12/2009Department of Computer Science, UIUC

More about Astrolabe Gossip /uiuc/ece/n1/uiuc/cs/n4/uiuc/cs/n6/cornell/n2/cornell/cs/n3 /berkeley /eecs/n5 /berkeley /eecs/n7 /berkeley /eecs/n8 /uiuc/ece /uiuc/cs /cornell/cs /berkeley/eecs /uiuc/cornell/berkeley / uiuc cornell berkeley cs ece How does /uiuc/cs/n4 know the MIB of /cornell? Gossiped MIBs in /uiuc/cs 14 By gossiping with /cornell/n2 or /cornell/cs/n3 3/12/2009Department of Computer Science, UIUC

More about Astrolabe Gossip (2) Each zone dynamically elects a set of representative agents to gossip on behalf of this zone. – Election can be based on the load of the agents, their longevity. – The MIB contains the list of representative agents. – An agent can represent for multiple zones. Each agent periodically gossips for a zone it represents. – Randomly picks another sibling zones and one of its representative agents. – Gossips the MIBs of all their sibling zones. Gossip dissemination within a zone grows at O(logK). – K = no. of child zones. 153/12/2009Department of Computer Science, UIUC

More about Astrolabe Gossip (3) /uiuc/ece/n1/uiuc/cs/n4/uiuc/cs/n6/cornell/n2/cornell/cs/n3 /berkeley /eecs/n5 /berkeley /eecs/n7 /berkeley /eecs/n8 /uiuc/ece /uiuc/cs /cornell/cs /berkeley/eecs /uiuc/cornell/berkeley / /uiuc/cs/n4 /uiuc/ece/n1 /cornell/cs/n3 /berkeley/eecs/n8 /berkeley/eecs/n5 Gossips about the MIBs of the /berkeley, /uiuc 16 Representative agents for /uiuc uiuc cornell berkeley 3/12/2009Department of Computer Science, UIUC

Example: P2P Caching of Large Objects Query to locate a copy of the file, game.db – SELECT COUNT(*) AS file_count FROM files WHERE name = ‘game.db’ – SELECT FIRST(1, result) AS result SUM(file_count) AS file_count WHERE file_count > 0 The SQL query code is installed in an Astrolabe agent using an aggregation function certificate (AFC) 17 Each host installs an attribute ‘result’ with its host name in its leaf MIB. Aggregates the ‘result’ of each zone and picks one host in each zone. 3/12/2009Department of Computer Science, UIUC

Example: P2P Caching of Large Objects (2) The querying application introduces this new AFC at the management table of some Astrolabe agent. Query propagation and aggregation of output: 1.An Astrolabe agent automatically evaluates the AFC for the leaf zone and recursively updates the table of the ancestor zones. A copy of the AFC is also included along the result. This query is evaluated recursively till the root zone as long as the policy permits. 2.AFCs are also gossiped to other agents (as part of the MIB). An receiving agent scans the gossiped message and installs the new AFCs and in the leaf MIBs Each AFC has an expiration period. Till then, queries are evaluated frequently. 183/12/2009Department of Computer Science, UIUC

Membership Removing failed or disconnected nodes: – Astrolabe also gossips about the membership information. – If a process (or host) fails, its MIB will eventually expire and be deleted. Integrating new members: – Astrolabe relies on IP multicast to set up the initial contacts. Gossip message is also broadcast on the local LAN (occasionally). – Astrolabe agents also contact a set of its relatives (occasionally) 193/12/2009Department of Computer Science, UIUC

Simulation Results - Number of representative agents per zone = 1. - No failure. The smaller the branching of the zone hierarchy, the slower the gossip dissemination. 20 Effect of branching factor of the zone hierarchy on the gossip dissemination time. 3/12/2009Department of Computer Science, UIUC

Simulation Results (2) - Branching factor = 25 - No failure. The higher the representative agents per zone, the lower the time of gossip dissemination in the presence of failures. 21 Effect of the number of representative agents on the gossip dissemination time. 3/12/2009Department of Computer Science, UIUC

Discussion Astrolabe is not meant to provide routing features similar to DHTs. – How is Astrolabe different from DHTs? Astrolabe attributes are updated proactively and frequently. – Do you think this proactive management is better than the reactive one? 223/12/2009Department of Computer Science, UIUC

Moara: Flexible and Scalable Group-Based Querying System Steven Y. Ko 1, Praveen Yalagandula 2, Indranil Gupta 1, Vanish Talwar 2, Dejan Milojicic 2, Subu Iyer 2 1 University of Illinois at Urbana-Champaign 2 HP Labs, Palo Alto ACM/IFIP/USENIX Middleware, 2008 Presented by Ahmed Khurshid

Motivation 3/12/2009Department of Computer Science, UIUC24 Linux Apache MySQL What is the average memory utilization of machines running MySQL? Naïve approach Consumes extra bandwidth and adds delay

Motivation (cont.) 3/12/2009Department of Computer Science, UIUC25 Linux Apache MySQL What is the average memory utilization of machines running MySQL? Better approach Avoids unnecessary traffic

Two Approaches 3/12/2009Department of Computer Science, UIUC26 Single Tree (No Grouping) Query cost is high Group Based Trees Group maintenance cost is high MOARAMOARA

Moara Features Moara maintains aggregation trees for different groups – Uses FreePasrty DHT for this purpose Supports a query language having the following form – (query-attribute, aggregation function, group-predicate) – E.g. (Mem-Util, Average, MySQL = true) Supports composite queries that target multiple groups using unions and intersections Reduces bandwidth usage and response time by – Adaptively pruning branches of the DHT tree – Bypassing intermediate nodes that do not satisfy a given query – Only querying those nodes that are able to answer quickly without affecting the result of the query 3/12/2009Department of Computer Science, UIUC27

Common Queries 3/12/2009Department of Computer Science, UIUC28

Group Size and Dynamism 3/12/2009Department of Computer Science, UIUC29 Usage of PlanetLab nodes by different slices Usage of HP’s utility computing environment by different jobs Most slices have fewer than 10 nodes Number of machines to do a job varies

Moara: Data and Query Model Three part query – Query attribute – Aggregation function – Group-predicate Aggregation functions are partially aggregatable – To perform in-network aggregation Composite queries can be constructed using “and” and “or operators” – E.g. (Linux=true and Apache=true) 3/12/2009Department of Computer Science, UIUC30 (attribute 1, value 1 ) (attribute 2, value 2 ). (attribute n, value n ) Moara Agent

Scalable Aggregation Moara employs a peer-to-peer in-network aggregation approach Maintains separate DHT trees for each group predicate Hash of the group attribute is used to designate a node as the root of such tree Queries are first send to the root node that then propagates the query down the tree Data coming from child nodes are aggregated before sending results to parent node 3/12/2009Department of Computer Science, UIUC31

DHT Tree Steps Take the hash of the group predicate – Let, Hash(ServiceX) = 000 Select the root based on the hash Use Pastry’s routing mechanism to join the tree – Similar to SCRIBE 3/12/2009Department of Computer Science, UIUC

Optimizations Prune out branches of the tree the do not contain any node satisfying a given query – Need to balance the cost of maintaining group tree with the query resolution cost Bypass internal nodes that do not satisfy a given query While dealing with composite queries, select a minimal set of groups by rewriting the query into a more manageable form 3/12/2009Department of Computer Science, UIUC33

Dynamic Maintenance 3/12/2009Department of Computer Science, UIUC34 DEFGH BC A p=false prune (p) = binary local state variable at every node per attribute p=false p(D)=false p(E)=false p=false p(F)=false p(G)=false p=false p(B)=false p(C)=false p(H)=false NO-PRUNE message

Dynamic Maintenance 3/12/2009Department of Computer Science, UIUC35 DEFGH BC A p=false prune (p) = binary local state variable at every node p=falsep=truep=false p(D)=false p(E)=false p=false p(F)=true p(G)=false p=false p(B)=false p(C)=false p(H)=false PRUNE message

Dynamic Maintenance 3/12/2009Department of Computer Science, UIUC36 DEFGH BC A p=false prune (p) = binary local state variable at every node p=falsep=truep=false p(D)=false p(E)=false p=false p(F)=true p(G)=true p=true p(B)=false p(C)=true p(H)=false PRUNE message High group churn rate will cause more PRUNE/NO-PRUNE message and may be more expensive than forwarding query to all nodes

Adaptation Policy Maintain two additional state variables at every node – sat To track if a subtree rooted at this node should continue receiving queries for a given predicate – update Denotes whether a node will update its prune variable or not The following invariants are maintained – update = 1 AND sat = 1 => prune = 0 – update = 1 AND sat = 0 => prune = 1 – update = 0 => prune = 0 3/12/2009Department of Computer Science, UIUC37

Adaptation Policy (cont.) 3/12/2009Department of Computer Science, UIUC38 When overhead due to unrelated queries is higher compared to group maintenance messages When group maintenance messages consume more bandwidth than queries

Separate Query Plane Used to bypass intermediate nodes that do not satisfy a given query Reduces message complexity from O(mlogN) to O(m), where – N = total nodes in the system – m = number of query satisfying nodes Uses two locally maintained sets at each node – updateSet – a list of nodes forwarded to parent – qSet – list of children nodes to which queries are forwarded qSet is the union of all updateSets received from child nodes Based on the size of the qSet and SAT value, a node decides whether to remain in the tree or not 3/12/2009Department of Computer Science, UIUC39

Separate Query Plane (cont.) 3/12/2009Department of Computer Science, UIUC40 SAT NOSAT CDE B A CD B C,D updateSet qSet Remain in the tree if SAT is true and |qSet| ≥ threshold SAT NOSAT CDE B A CD C,D threshold = 2threshold = 3 (B) B Query (C,D) C,D

Composite Queries Moara does not maintain trees for composite queries Answers composite queries by contacting one or more simple predicate trees Example - 1 – (Free_Mem, Avg, ServiceX = true AND ServiceY = true) – Two trees – one for each service – Can be answered using a single tree (whichever promises to respond early) Example - 2 – (Free_Mem, Avg, ServiceX = true OR ServiceY = true) – Two trees – one for each service – Need to query both trees 3/12/2009Department of Computer Science, UIUC41

Composite Queries (cont.) Moara selects a small cover that contains required trees to answer a query For example – cover(Q=“A”) = {A} if A is a predefined group – cover(Q=“A or B”) = cover(A) U cover(B) – cover(Q=“A and B”) = cover(A), cover(B) or (cover(A) U cover(B)) Bandwidth is saved by – Rewriting a nested query to select a low cost cover – Estimating query costs for individual trees – Use semantic information supplied by the user 3/12/2009Department of Computer Science, UIUC42

Finding Low-Cost Covers 3/12/2009Department of Computer Science, UIUC43 - CNF = Conjunctive Normal Form - Gives minimal-cost cover

Performance Evaluation - Dynamic Maintenance FreePastry simulator 10,000 nodes 2000-sized group churn event 500 events Query for an attribute A with value Є {0, 1} 3/12/2009Department of Computer Science, UIUC 44 Bandwidth usage with various query-to churn ratios Moara performs better than both the extreme approaches

Performance Evaluation - Separate Query Plane 3/12/2009Department of Computer Science, UIUC 45 Bandwidth usage for different (group size, threshold) For higher threshold values, query cost does not depend on total number of nodes

Emulab Experiments 50 machines, 10 instances of Moara per machine Fixed query rate 3/12/2009Department of Computer Science, UIUC46 Latency and BW usage with static groups Average Latency of dynamically changing groups Average latency with a static group of the same size - Latency and message cost increases with group size - Moara performs well even with high group churn events

PlanetLab Experiments 200 PlanetLab nodes, one instance of Moara per node 500 queries injected 5 seconds apart 3/12/2009Department of Computer Science, UIUC47 Moara vs. Centralized Aggregator Moara responds faster than a centralized aggregator

Discussion Points Using DAG or Synopsis Diffusion instead of trees Handling group churn in the middle of a query – Moara ensures eventual consistency Effect of nodes that are unreachable State maintenance overhead for different attributes Computation overhead for maintaining DHT trees for different attributes Using Moara in ad-hoc mobile wireless networks 3/12/2009Department of Computer Science, UIUC48

Network Imprecision: A New Consistency Metric for Scalable Monitoring Navendu Jain †, Prince Mahajan ⋆, Dmitry Kit ⋆, Praveen Yalagandula ‡, Mike Dahlin ⋆, and Yin Zhang ⋆ † Microsoft Research ⋆ The University of Texas at Austin ‡ HP Labs

Motivation Providing a consistency metric suitable for large-scale monitoring system Safeguarding accuracy despite node and network failure Providing a level of confidence on the reported information Efficiently tracking the number of nodes that fail to report status or that report status multiple times 3/12/2009Department of Computer Science, UIUC50

Thanks Questions and Comments?