HOW TO BUILD A BETTER TESTBED Fabien Hermenier Robert Ricci LESSONS FROM A DECADE OF NETWORK EXPERIMENTS ON EMULAB TridentCom ’ 2012 1.

Slides:



Advertisements
Similar presentations
S Licentiate course on Telecommunications Technology (4+1+3 cr.) Course Topic Spring 2000: Routing Algorithms in the DiffServ MPLS Networks Introduction.
Advertisements

System Area Network Abhiram Shandilya 12/06/01. Overview Introduction to System Area Networks SAN Design and Examples SAN Applications.
VCRIB: Virtual Cloud Rule Information Base Masoud Moshref, Minlan Yu, Abhishek Sharma, Ramesh Govindan HotCloud 2012.
Dynamic Topology Optimization for Supercomputer Interconnection Networks Layer-1 (L1) switch –Dumb switch, Electronic “patch panel” –Establishes hard links.
Estinet open flow network simulator and emulator. IEEE Communications Magazine 51.9 (2013): Wang, Shie-Yuan, Chih-Liang Chou, and Chun-Ming Yang.
* Mellanox Technologies LTD, + Technion - EE Department
The TickerTAIP Parallel RAID Architecture P. Cao, S. B. Lim S. Venkatraman, J. Wilkes HP Labs.
The Challenges of Repeatable Experiment Archiving – Lessons from DETER Stephen Schwab SPARTA, Inc. d.b.a. Cobham Analytic Solutions May 25, 2010.
Design Deployment and Use of the DETER Testbed Terry Benzel, Robert Braden, Dongho Kim, Clifford Informatino Sciences Institute
OpenFlow-Based Server Load Balancing GoneWild
CSC 450/550 Part 3: The Medium Access Control Sublayer More Contents on the Engineering Side of Ethernet.
1 Modeling and Emulation of Internet Paths Pramod Sanaga, Jonathon Duerig, Robert Ricci, Jay Lepreau University of Utah.
An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool.
1 GENI: Global Environment for Network Innovations Jennifer Rexford Princeton University
1 Cluster or Network? An Emulation Facility for Research Jay Lepreau Chris Alfeld David Andersen (MIT) Mac Newbold Rob Place Kristin Wright Dept. of Computer.
Disco Running Commodity Operating Systems on Scalable Multiprocessors.
Larger Site Networks Part2. 2 Ethernet Virtual LANs Hubs versus Switches –Hubs broadcast bits out all ports –Switches usually send a frame out a one port.
1 Implementing the Emulab-PlanetLab Portal: Experiences and Lessons Learned Kirk Webb Mike Hibler Robert Ricci Austin Clements Jay Lepreau University of.
1 A Large-Scale Network Testbed Jay Lepreau Chris Alfeld David Andersen Kevin Van Maren University of Utah September.
PortLand Presented by Muhammad Sadeeq and Ling Su.
1 A Large-Scale Network and Distributed Systems Testbed Jay Lepreau Chris Alfeld David Andersen (MIT) Kristin Wright University of Utah
A Scalable, Commodity Data Center Network Architecture Mohammad Al-Fares, Alexander Loukissas, Amin Vahdat Presented by Gregory Peaker and Tyler Maclean.
Testing Intrusion Detection Systems: A Critic for the 1998 and 1999 DARPA Intrusion Detection System Evaluations as Performed by Lincoln Laboratory By.
© Wiley Inc All Rights Reserved. CCNA: Cisco Certified Network Associate Study Guide CHAPTER 8: Virtual LANs (VLANs)
Practical TDMA for Datacenter Ethernet
Sybex CCNA Chapter 9: VLAN’s Instructor & Todd Lammle.
Network Support for Cloud Services Lixin Gao, UMass Amherst.
LAN Wiring & Physical Interfaces Lecture 6. Network Interfaces Every device has some form of a network interfaces. A PC has a Network Interface Card (NIC).
Mirror Mirror on the Ceiling: Flexible Wireless Links for Data Centers Presenter: Lu Gong.
Evaluating Centralized, Hierarchical, and Networked Architectures for Rule Systems Benjamin Craig University of New Brunswick Faculty of Computer Science.
Hosting Virtual Networks on Commodity Hardware VINI Summer Camp.
Dynamic Resource Allocation Using Virtual Machines for Cloud Computing Environment.
End-to-end resource management in DiffServ Networks –DiffServ focuses on singal domain –Users want end-to-end services –No consensus at this time –Two.
Capacity Scaling with Multiple Radios and Multiple Channels in Wireless Mesh Networks Oguz GOKER.
Network Aware Resource Allocation in Distributed Clouds.
This work was supported by the TRUST Center (NSF award number CCF ) Introduction Since public utilities must rely on the internet, they are vulnerable.
A Scalable Bandwidth Management Architecture for Supporting VoIP Applications Using Bandwidth Broker Zhenhai Duan, Zhi-Li Zhang University of Minnesota.
Improving Network I/O Virtualization for Cloud Computing.
E VALUATION OF F AIRNESS IN ICN X. De Foy, JC. Zuniga, B. Balazinski InterDigital
The University of Bolton School of Games Computing & Creative Technologies LCT2516 Network Architecture CCNA Exploration LAN Switching and Wireless Chapter.
Building a Parallel File System Simulator E Molina-Estolano, C Maltzahn, etc. UCSC Lab, UC Santa Cruz. Published in Journal of Physics, 2009.
LAN Switching and Wireless – Chapter 1
Introduction of CRON Lin Xue Feb What is CRON “cron.cct.lsu.edu” testbed project is based on the Emulab system in the University of Utah. Emulab:
LAN Switching and Wireless – Chapter 1 Vilina Hutter, Instructor
Selfishness, Altruism and Message Spreading in Mobile Social Networks September 2012 In-Seok Kang
Inter VLAN routing W.lilakiatsakun. What is inter VLAN routing.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
1 Testbeds Breakout Tom Anderson Jeff Chase Doug Comer Brett Fleisch Frans Kaashoek Jay Lepreau Hank Levy Larry Peterson Mothy Roscoe Mehul Shah Ion Stoica.
Large-scale Virtualization in the Emulab Network Testbed Mike Hibler, Robert Ricci, Leigh Stoller Jonathon Duerig Shashi Guruprasad, Tim Stack, Kirk Webb,
McGraw-Hill©The McGraw-Hill Companies, Inc., 2004 Connecting Devices CORPORATE INSTITUTE OF SCIENCE & TECHNOLOGY, BHOPAL Department of Electronics and.
Shivkumar Kalyanaraman Rensselaer Polytechnic Institute 1 Based upon slides from Jay Lepreau, Utah Emulab Introduction Shiv Kalyanaraman
1 Wide Area Network Emulation on the Millennium Bhaskaran Raman Yan Chen Weidong Cui Randy Katz {bhaskar, yanchen, wdc, Millennium.
Advanced Computer Networks Lecturer: E EE Eng. Ahmed Hemaid Office: I 114.
SHADOWSTREAM: PERFORMANCE EVALUATION AS A CAPABILITY IN PRODUCTION INTERNET LIVE STREAM NETWORK ACM SIGCOMM CING-YU CHU.
Multi-layer Network Virtualization with Resource Reservation based on SDN Nguyen Huu Thanh Tran Manh Nam Truong Thu Huong School of Electronics and Telecommunications.
Author Utility-Based Scheduling for Bulk Data Transfers between Distributed Computing Facilities Xin Wang, Wei Tang, Raj Kettimuthu,
Akhyari Nasir.  Router-on-a-stick is a type of router configuration in which a single physical interface routes traffic between.
GPFS: A Shared-Disk File System for Large Computing Clusters Frank Schmuck & Roger Haskin IBM Almaden Research Center.
Using Deduplicating Storage for Efficient Disk Image Deployment Xing Lin, Mike Hibler, Eric Eide, Robert Ricci University of Utah.
PACMan: Coordinated Memory Caching for Parallel Jobs Ganesh Ananthanarayanan, Ali Ghodsi, Andrew Wang, Dhruba Borthakur, Srikanth Kandula, Scott Shenker,
Network Topologies for Scalable Multi-User Virtual Environments Lingrui Liang.
Chapter 3: Link Aggregation
LESSON 2.1_A Networking Fundamentals Understand Switches.
Architecture and Algorithms for an IEEE 802
Chuanxiong Guo, et al, Microsoft Research Asia, SIGCOMM 2008
TRUST:Team for Research in Ubiquitous Secure Technologies
Cluster Resource Management: A Scalable Approach
ExaO: Software Defined Data Distribution for Exascale Sciences
Comparison to existing state of security experimentation
Towards Predictable Datacenter Networks
Presentation transcript:

HOW TO BUILD A BETTER TESTBED Fabien Hermenier Robert Ricci LESSONS FROM A DECADE OF NETWORK EXPERIMENTS ON EMULAB TridentCom ’

Network testbeds  support for experiments in networked systems  highly customizable networked environments  raw access to a variety of specific hardware  design and features to match researchers’ needs 2

Testbed physical design  no extensive studies for effective testbed design  assumption-based  budget-constrained  lack of data from real experimenters 3  bad design decisions have consequences  prevent support for certain experiments  over-commitment on un-needed hardware

How to build better testbeds ?  careful analysis of the Utah Emulab facility usage  one of the largest testbeds  used in production since 2001  > 4 dozen testbeds worldwide with a similar designs  several alternative testbed designs  evaluation of new designs using real workload 4

The Utah Emulab facility: not just another pretty cluster 5

Virtual topology in Emulab 6 Lan node Full bisection bandwidth between members Traffic shaping Implemented with a PC “delay node”

Making the virtual into reality 7

Mapping a virtual topology 8

The raw dataset 522 projects – 22,266 experiments – 619,504 topologies 9 -empty, non-instantiated and single node topologies-unconnected topologies-internal projects- non-cluster resources The working dataset 477 projects – 13,057 experiments – 504,226 topologies

Most experiments are small 10 80%4%  multiplexable topologies  to increase acceptance rate  to deploy larger experiments  prepare with small topologies  evaluate with larger

Attractive nodes are the bottleneck 11  upgrading is a necessity  to meet the demand  to support larger experiments  beauty is ephemeral  repeatability is essential

Most requests use few interfaces 12  network interfaces are wasted  homogeneous connectivity does not reflect the reality of usage  traffic shaping is a prime feature  multiple interfaces are required

LANs are common, but most are small 13  another prime feature  LANs dominate large experiments 90%  small LANs  a moderated usage of the interswitch bandwidth ?

Facts from experimenters data 14  the testbed size limits  its acceptance rate  experiments size  connectivity is overprovisioned  improved designs must provide  some nodes with multiple interfaces  non-blocking bandwidth between a few nodes

A cost model for testbeds  the impact of high connectivity is significant at scale  $100,000: 34 2-link nodes or 27 4-link nodes ?  $1,000,000: link nodes or link nodes ? 15

Heterogeneous node connectivity  the best nodes for a topology do not have extra links  from an ideal node connectivity distribution…  … to a testbed that minimizes link waste %

A cost model for testbeds  small switches lead to a 30% bigger testbed …  … but the reduced bisection bandwidth between the nodes limits maximum LAN size 17

The interswitch bandwidth myth  a simulation replayed the user requests for the pc3000 nodes  interswitch links are overprovisioned  an opportunity for smaller and cheaper switches Busted 97% of the topologies 18

Alternative for switch connectivity  interswitch bandwidth  limit LANs, experiments among switches  faster interconnect are expensive  link concentration limits direct communication 19

Striping links  direct communication between nodes  scalability limited by switch size  hard to mix with heterogeneous connectivity 20

Evaluating new designs  impact of relevant testbed designs on  completion time  rejection rate  the cost model as a testbed generator  the 15k topologies focusing pc3000 nodes as workload  a simulator to replay the mapping  FIFO scheduling policy  each topology runs for a day 21

Connectivity for nodes: good trade 2-link nodesNodesBandwidthRejectionsTime (days) 0% Gb % Gb % Mb33 (0.2%) 1487  Testbeds  $500,000 as funds  nodes with 2 or 4 links 22

Striping annihilates bandwidth requirements ConfigurationBandwidthRejectionsTime (days) 1.11 Gb Mb0996  Testbeds  148 nodes; 60% with 2-links  2 switches 23

Small switches, big testbed ConfigurationCostNodesBandwidthRejections Time (days) 264-ports$498, Gb ports$498, Gb138 (0.9%) ports$390, Mb142 (0.9%)1314  using large switches, support the 0.9% “hard” topologies  with 40% more time  with $108,000. $761 per “hard” topology vs. $33.5  we are the 99% 24

Conclusions  the testbed size is the bottleneck, not the network  facts lead to new design suggestions  lower connectivity  smaller switches  link striping  cost models and replays for a good insight into the testbed’s effectiveness  what to give up to support the outliers?  read the paper for more conclusions 25

Watson, I told you the problem wasn’t the interswitch bandwidth Watson, I told you the problem wasn’t the interswitch bandwidth you Sherlock How to Build a Better Testbed Lessons from a decade of network experiments on Emulab 26

Datacenter vs. network testbed physical design 27  datacenters  node centric  network as a support to maximize performance  non-explicit communications  network testbeds  network centric  explicit communication to reproduce  conservative allocation