Networking Panel Jeannie Albrecht Williams College, Plush/Gush project Ivan Seskar Rutgers University, WINLAB/ORBIT project Steven Schwab Cobham Analytic.

Slides:



Advertisements
Similar presentations
Seungmi Choi PlanetLab - Overview, History, and Future Directions - Using PlanetLab for Network Research: Myths, Realities, and Best Practices.
Advertisements

Proposal by CA Technologies, IBM, SAP, Vnomic
PlanetLab: An Overlay Testbed for Broad-Coverage Services Bavier, Bowman, Chun, Culler, Peterson, Roscoe, Wawrzoniak Presented by Jason Waddle.
FIBRE-BR Meeting GENI I&M Marcelo Pinheiro. Agenda GENI Overview GENI User groups GENI I&M Use Cases GENI I&M Services.
Resource specification update for PlanetLab and VINI Andy Bavier Princeton University March 16, 2010.
Selected Project Highlights: An Overview of Gush Jeannie Albrecht Williams College
GENI Experiment Control Using Gush Jeannie Albrecht and Amin Vahdat Williams College and UC San Diego.
An Overview of Gush Jeannie Albrecht David Irwin
PlanetLab Operating System support* *a work in progress.
The Challenges of Repeatable Experiment Archiving – Lessons from DETER Stephen Schwab SPARTA, Inc. d.b.a. Cobham Analytic Solutions May 25, 2010.
MCTS GUIDE TO MICROSOFT WINDOWS 7 Chapter 10 Performance Tuning.
Toward Replayable Research in Networking and Systems Eric Eide University of Utah, School of Computing May 25, 2010.
1 In VINI Veritas: Realistic and Controlled Network Experimentation Jennifer Rexford with Andy Bavier, Nick Feamster, Mark Huang, and Larry Peterson
Distributed Application Management Using PLuSH Jeannie Albrecht, Christopher Tuttle, Alex C. Snoeren, and Amin Vahdat UC San Diego CSE {jalbrecht, ctuttle,
An Experimentation Workbench for Replayable Networking Research Eric Eide, Leigh Stoller, and Jay Lepreau University of Utah, School of Computing NSDI.
Integrated Scientific Workflow Management for the Emulab Network Testbed Eric Eide, Leigh Stoller, Tim Stack, Juliana Freire, and Jay Lepreau and Jay Lepreau.
MCDST : Supporting Users and Troubleshooting a Microsoft Windows XP Operating System Chapter 10: Collect and Analyze Performance Data.
1 PLuSH – Mesh Tree Fast and Robust Wide-Area Remote Execution Mikhail Afanasyev ‧ Jose Garcia ‧ Brian Lum.
1 Implementing the Emulab-PlanetLab Portal: Experiences and Lessons Learned Kirk Webb Mike Hibler Robert Ricci Austin Clements Jay Lepreau University of.
©Company confidential 1 Performance Testing for TM & D – An Overview.
Platform as a Service (PaaS)
11 MAINTAINING THE OPERATING SYSTEM Chapter 5. Chapter 5: MAINTAINING THE OPERATING SYSTEM2 CHAPTER OVERVIEW Understand the difference between service.
16.1 © 2004 Pearson Education, Inc. Exam Managing and Maintaining a Microsoft® Windows® Server 2003 Environment Lesson 16: Examining Software Update.
Enterprise Search. Search Architecture Configuring Crawl Processes Advanced Crawl Administration Configuring Query Processes Implementing People Search.
11 MAINTAINING THE OPERATING SYSTEM Chapter 5. Chapter 5: MAINTAINING THE OPERATING SYSTEM2 CHAPTER OVERVIEW  Understand the difference between service.
Load Test Planning Especially with HP LoadRunner >>>>>>>>>>>>>>>>>>>>>>
Chapter 17: Watching Your System BAI617. Chapter Topics Working With Event Viewer Performance Monitor Resource Monitor.
Hands-On Microsoft Windows Server 2008 Chapter 1 Introduction to Windows Server 2008.
Acceleratio Ltd. is a software development company based in Zagreb, Croatia, founded in We create innovative software solutions for SharePoint,
Shilpa Seth.  Centralized System Centralized System  Client Server System Client Server System  Parallel System Parallel System.
Web Based Applications
Hosting Virtual Networks on Commodity Hardware VINI Summer Camp.
Type of Software There are two main types of software They are System software Application software Hardware System Software (OS) Application Software.
Hands-On Microsoft Windows Server 2008 Chapter 1 Introduction to Windows Server 2008.
MCTS Guide to Microsoft Windows 7
Andy Bavier, PlanetWorks Scott Baker, SB-Software July 27, 2011.
CHAPTER 2 OPERATING SYSTEM OVERVIEW 1. Operating System Operating System Definition A program that controls the execution of application programs and.
1 Guide to Novell NetWare 6.0 Network Administration Chapter 13.
Chapter 3: Operating-System Structures System Components Operating System Services System Calls System Programs System Structure Virtual Machines System.
Cluster Reliability Project ISIS Vanderbilt University.
Module 7: Fundamentals of Administering Windows Server 2008.
1 Supporting the development of distributed systems CS606, Xiaoyan Hong University of Alabama.
Intel IT Overlay Jeff Sedayao PlanetLab Workshop at HPLABS May 11, 2006.
Lecture 4: Sun: 23/4/1435 Distributed Operating Systems Lecturer/ Kawther Abas CS- 492 : Distributed system & Parallel Processing.
1 Overview of IEPM-BW - Bandwidth Testing of Bulk Data Transfer Tools Connie Logg & Les Cottrell – SLAC/Stanford University Presented at the Internet 2.
Maintaining and Updating Windows Server Monitoring Windows Server It is important to monitor your Server system to make sure it is running smoothly.
Capabilities of Software. Object Linking & Embedding (OLE) OLE allows information to be shared between different programs For example, a spreadsheet created.
VO-Ganglia Grid Simulator Catalin Dumitrescu, Mike Wilde, Ian Foster Computer Science Department The University of Chicago.
1 Testbeds Breakout Tom Anderson Jeff Chase Doug Comer Brett Fleisch Frans Kaashoek Jay Lepreau Hank Levy Larry Peterson Mothy Roscoe Mehul Shah Ion Stoica.
LAMP: Bringing perfSONAR to ProtoGENI Martin Swany.
Large-scale Virtualization in the Emulab Network Testbed Mike Hibler, Robert Ricci, Leigh Stoller Jonathon Duerig Shashi Guruprasad, Tim Stack, Kirk Webb,
Shivkumar Kalyanaraman Rensselaer Polytechnic Institute 1 Based upon slides from Jay Lepreau, Utah Emulab Introduction Shiv Kalyanaraman
Copyright 2007, Information Builders. Slide 1 Machine Sizing and Scalability Mark Nesson, Vashti Ragoonath June 2008.
Sample Presentation Headline REPRESENTATIVE SUBHEAD TO SUPPORT SUBJECT Presenter’s Name Presenter’s Title Presentation Date DeterLab A Tool for Cybersecurity.
Globus and PlanetLab Resource Management Solutions Compared M. Ripeanu, M. Bowman, J. Chase, I. Foster, M. Milenkovic Presented by Dionysis Logothetis.
 Distributed Database Concepts  Parallel Vs Distributed Technology  Advantages  Additional Functions  Distribution Database Design  Data Fragmentation.
Page 1 Monitoring, Optimization, and Troubleshooting Lecture 10 Hassan Shuja 11/30/2004.
Unit 2 VIRTUALISATION. Unit 2 - Syllabus Basics of Virtualization Types of Virtualization Implementation Levels of Virtualization Virtualization Structures.
Spark on Entropy : A Reliable & Efficient Scheduler for Low-latency Parallel Jobs in Heterogeneous Cloud Huankai Chen PhD Student at University of Kent.
Towards a High Performance Extensible Grid Architecture Klaus Krauter Muthucumaru Maheswaran {krauter,
Platform as a Service (PaaS)
Platform as a Service (PaaS)
Clouds , Grids and Clusters
IoT at the Edge Technical guidance deck.
MCTS Guide to Microsoft Windows 7
Introduction to Operating System (OS)
Introduction to Operating Systems
IoT at the Edge Technical guidance deck.
Grid Means Business OGF-20, Manchester, May 2007
IP Control Gateway (IPCG)
Presentation transcript:

Networking Panel Jeannie Albrecht Williams College, Plush/Gush project Ivan Seskar Rutgers University, WINLAB/ORBIT project Steven Schwab Cobham Analytic Solutions, DETER project Eric Eide University of Utah, Emulab project

Achieving Experiment Repeatability on PlanetLab Jeannie Albrecht Williams College

Overview Archiving experiments on wide-area testbeds requires the ability to capture (i.e., measure and record): Network conditions (bandwidth, latency, etc) Machine properties (CPU usage, free memory, etc) Experiment characteristics (software/OS versions, etc) Repeating experiments on wide-area testbeds requires the ability to configure these same properties How can we achieve these goals on wide-area testbeds?

4 Network of Linux machines at 500+ sites in 25+ countries Allows researchers to run experiments “in the wild” (i.e., on machines spread around the world connected via “normal” Internet links) Each user gets an “account” (called a sliver) on each machine Resources are “allocated” via a proportional fair share scheduler Volatile network High contention for machines leads to high failure rates near deadlines Common problems: low disk space, clock skew, connection refused In April 2006, only 394/599 machines were actually usable

Experimenter Tools Many tools exist/have existed for coping with unpredictability of PlanetLab Monitoring services – measure machine/network usage in real-time CoMon ( S 3 ( Ganglia, iPerf, all-pairs-ping, Trumpet Resource discovery – find machines that meet specific criteria Sword ( Experiment management – simplify/automate tasks associated with running experiments Gush/Plush ( appmanager ( research.net/) research.net/

CoMon: Node Monitoring

S 3 : Network Monitoring

Node1 Node5 Node6 Node7 Node8 Node4 Node3Node2 Node10 Node9 Node11 CoMon+S 3 data Node5Node4Node2Node6Node7 Group 1Group 2 (i) Query Candidate nodes SWORD (ii) Logical Database & Query Processor XML (iii) Matcher & Optimizer PlanetLab Optimal resource groups Node5Node4 Node2No5e6Node7 Group 1 Group 2 Node6 Node4 Node3 SWORD: Resource Discovery

Gush: Experiment Management Allows users to describe, run, monitor, & visualize experiments XML-RPC interface for managing experiments programmatically

Capturing Live Conditions Machine properties CoMon is a centrally run service that satisfies this requirement Experiment characteristics Gush records information about software versions and machines used for experiment Network conditions S 3 mostly meets these requirements Other services have existed in the past—now mostly offline! S 3 is difficult to query (lacks “sensor” interface) and is only updated every 4 hours

Experiment Configuration Machine properties No resource isolation in PlanetLab Cannot specify machine properties Experiment characteristics Experiment management and resource discovery tools can help with this Cannot control OS version Network conditions Currently no way to specify underlying network topology characteristics

Possible Solutions 1.Create a reliable network measurement service (similar to S 3 +CoMon)! 2.Capture conditions in initial experiment; monitor live conditions until they “match” and then start experiment 3.Provide stronger resource isolation on PlanetLab (Vini?) 4.Use captured conditions to replay experiment in more controllable environment (Emulab, ORCA, etc)

Food For Thought Experiment archival on PlanetLab is difficult but can (almost) be accomplished Experiment repeatability is mostly impossible But is this necessarily bad? What does it mean for an experiment to be repeatable? Do all testbeds have to enable fully repeatable experiments? Does archival imply repeatability? Are both required? Some volatility/unpredictability is arguably a good thing (more “realistic”) Internet does not provide repeatability! Perhaps best approach is to use a combination of configurable and non-configurable testbeds Simulation/emulation + live deployment Best of both worlds?

Thanks!