Taming the Complexity of Artifact Reproducibility

Slides:



Advertisements
Similar presentations
The Challenges of Repeatable Experiment Archiving – Lessons from DETER Stephen Schwab SPARTA, Inc. d.b.a. Cobham Analytic Solutions May 25, 2010.
Advertisements

Chapter 14 Network Design and Implementation. 2 Network Analysis and Design Aspects of network analysis and design Understanding the requirements for.
SKELETON BASED PERFORMANCE PREDICTION ON SHARED NETWORKS Sukhdeep Sodhi Microsoft Corp Jaspal Subhlok University of Houston.
Integrated Scientific Workflow Management for the Emulab Network Testbed Eric Eide, Leigh Stoller, Tim Stack, Juliana Freire, and Jay Lepreau and Jay Lepreau.
GSJGD. Agenda Review of what we delivered in Visual Studio enhancements for Visual Studio 11 Paradigm Shift to Exploratory Testing Enhancements.
QWise software engineering – refactored! Testing, testing A first-look at the new testing capabilities in Visual Studio 2010 Mathias Olausson.
IT Project Management Cheng Li, Ph.D. August 2003.
GIG Software Integration: Area Overview TeraGrid Annual Project Review April, 2008.
Implementation of HUBzero as a Knowledge Management System in a Large Organization HUBBUB Conference 2012 September 24 th, 2012 Gaurav Nanda, Jonathan.
Flexible Performance Prediction of Data Center Networks using Automatically Generated Simulation Models Piotr Rygielski, Samuel Kounev, Phuoc Tran-Gia.
ICN Baseline Scenarios draft-pentikousis-icn-scenarios-04 K. Pentikousis (Ed.), B. Ohlman, D. Corujo, G. Boggia, G. Tyson, E. Davies, P. Mahadevan, S.
Algoval: Evaluation Server Past, Present and Future Simon Lucas Computer Science Dept Essex University 25 January, 2002.
Rational Unified Process Fundamentals Module 5: Implementing RUP.
ICN Baseline Scenarios draft-pentikousis-icn-scenarios-04 K. Pentikousis (Ed.), B. Ohlman, D. Corujo, G. Boggia, G. Tyson, E. Davies, P. Mahadevan, S.
SDN Management Layer DESIGN REQUIREMENTS AND FUTURE DIRECTION NO OF SLIDES : 26 1.
Shivkumar Kalyanaraman Rensselaer Polytechnic Institute 1 Based upon slides from Jay Lepreau, Utah Emulab Introduction Shiv Kalyanaraman
GraDS MacroGrid Carl Kesselman USC/Information Sciences Institute.
28 June 2000P. Mato, CERN1 Software Checklist Proposal for controlling and documenting all the steps in developing software Program of Work Meeting, June.
1 Evaluation of Cooperative Web Caching with Web Polygraph Ping Du and Jaspal Subhlok Department of Computer Science University of Houston presented at.
© THE UNIVERSITY OF WAIKATO TE WHARE WANANGA O WAIKATO 1 ns-2 TCP Simulations with The Network Simulation Cradle Sam Jansen and Anthony McGregor.
Interaction and Animation on Geolocalization Based Network Topology by Engin Arslan.
NFP: Enabling Network Function Parallelism in NFV
SDN controllers App Network elements has two components: OpenFlow client, forwarding hardware with flow tables. The SDN controller must implement the network.
ITEC 275 Computer Networks – Switching, Routing, and WANs
Mininet and Openflow Labs
Analysis Services in times of Continuous Integration
Prepared by: Ameer Sameer Hamood University of Babylon - Iraq
SDN challenges Deployment challenges
A Network-State Management Service
Scholarly Workflow: Federal Prototype and Preprints
The DPIaaS Controller Prototype
Heitor Moraes, Marcos Vieira, Italo Cunha, Dorgival Guedes
WSP quality assurance tool
ReproZip: Computational Reproducibility With Ease
CONFIGURING A MICROSOFT EXCHANGE SERVER 2003 INFRASTRUCTURE
IEEE 802 OmniRAN Study Group: SDN Use Case
Securing the Network Perimeter with ISA 2004
Learning Reproducibility with a Yearly Networking Contest
Streamlining the Statistical Production in TurkStat Metadata Studies in TURKSTAT High Level Seminar for Eastern Europe, Caucasus and Central Asia Countries.
The 2007 Winter Conference on Business Intelligence
Enterprise vCPE use case requirement
CIS 332 Course Experience Tradition / snaptutorial.com
Web Engineering.
Enhancing Scholarly Communication with ReproZip
SCTP v/s TCP – A Comparison of Transport Protocols for Web Traffic
API Documentation Guidelines
CAE-SCRUB for Incorporating Static Analysis into Peer Reviews
Enhancing Cloud Foundry with CLI Plugins
Cost To Develop Blockchain Wallet App?. It is difficult to define the exact price of the Bitcoin wallet application development, however, Here is a cost.
NFP: Enabling Network Function Parallelism in NFV
Northbound API Dan Shmidt | January 2017
Continuous Automated Chatbot Testing
Dev Test on Windows Azure Solution in a Box
Rational Unified Process
Software Defined Networking (SDN)
Standard Scripts Project 2
NFP: Enabling Network Function Parallelism in NFV
DocuMelt 2.0 An EmpFinesseTM Fundamentals Solution.
HEW Study Group Documentation
Helping a friend out Guidelines for better software
Managing Services with VMM and App Controller
Touchstone Testing Platform
Network Troubleshooting
Standard Scripts Project 2
敦群數位科技有限公司(vanGene Digital Inc.) 游家德(Jade Yu.)
Standard Scripts Project 2
In-network computation
Data + Research Elements What Publishers Can Do (and Are Doing) to Facilitate Data Integration and Attribution David Parsons – Lawrence, KS, 13th February.
ONAP Architecture Principle Review
Logical Architecture & UML Package Diagrams
Presentation transcript:

Taming the Complexity of Artifact Reproducibility Matthias Flittner1, Robert Bauer1, Amr Rizk2, Stefan Geißler3, Thomas Zinner3, Martina Zitterbart1 1Karlsruhe Institute of Technology, 2Technische Universität Darmstadt, 3University of Würzburg

Should be reproducible! A typical paper Evaluation Related Work Problem Design Should be reproducible! Taming the Complexity of Artifact Reproducibility

Reproducibility by Design General idea Reproducibility by Design Taming the Complexity of Artifact Reproducibility

Methodology Selected five of 2016‘s top SDN conferences CoNext, NSDI, OSDI, SIGCOMM, SOSR Identified a sample of 34 SDN papers Focused on evaluation of research artifacts Analysis of basic requirements for reproducibility Type of evaluation Tools Topologies Traffic Environment Documentation Taming the Complexity of Artifact Reproducibility

Observation 1 Taming the Complexity of Artifact Reproducibility

Observation 1 Proper documentation is difficult Version Version of mininet HW-OS Operation System HW-spec Hardware Specs (CPU, Memory, …) HW-dim Dimensioning of the hardware (#servers, …) HW-vswitch Virtual Switch (OVS, bmv2, …) HW-nic/net Network Setup (NIC, Latency) Topo Topology used with mininet Topo-size Dimensioning of the topology Traffic-tools Tool(s) used for traffic generation Traffic-param Parameters / Workflow of traffic generation Metrics What is evaluated? Iterations #experiments Example from paper study: 6 of 34 investigated papers make use of the tool „mininet“  What documentation is required for reproducibility? Taming the Complexity of Artifact Reproducibility

Proper documentation is difficult Observation 1 Proper documentation is difficult Example 1 Example 2 Example 3 Example 4 Example 5 Example 6 Version No Yes HW-OS HW-spec HW-dim HW-vswitch HW-nic/net Topo Topo-size Traffic-tools Traffic-param Partial Metrics Iterations Taming the Complexity of Artifact Reproducibility

Proper documentation is difficult Observation 1 Proper documentation is difficult Example 1 Example 2 Example 3 Example 4 Example 5 Example 6 Version No Yes HW-OS HW-spec HW-dim HW-vswitch HW-nic/net Topo Topo-size Traffic-tools Traffic-param Partial Metrics Iterations Taming the Complexity of Artifact Reproducibility

Proposal: Meta-Artifacts Common way of describing well-known aspects of the evaluation Select the tool Fill out a domain-specific template Provide access to Meta-Artifacts Paper author Portal / Publisher Taming the Complexity of Artifact Reproducibility

Example Step 1: select the tools Taming the Complexity of Artifact Reproducibility

Step 2: fill out the templates Example Step 2: fill out the templates Taming the Complexity of Artifact Reproducibility

Automatic / easy access Checklist for documentation Example Benefits? Automatic / easy access Checklist for documentation Sources for inspiration Taming the Complexity of Artifact Reproducibility

Observation 2 Taming the Complexity of Artifact Reproducibility

Similar terms but different realizations Observation 2 Similar terms but different realizations Taming the Complexity of Artifact Reproducibility

Similar terms but different realizations Observation 2 Similar terms but different realizations Some similarities, but different traffic, topologies, configurations, implementations … Taming the Complexity of Artifact Reproducibility

Proposal: Shared Evaluation Environment Separation of scenario, application and evaluation environment Allows reuse, comparability and reproducibility Taming the Complexity of Artifact Reproducibility

Example @KIT: one-year research project of three students  Results Based upon a shared simulative environment Elaboration of common tools, traffic, topologies, and implementations  Results Support for NS3, OMNeT++ and Mininet Unified data format for traffic and topologies SDN app integration (POX) Exchange / Share of apps, scenarios, and simulator Other examples Portable workflow framework (CK), http://cknowledge.org/ Pantheon of Congestion Control, http://pantheon.stanford.edu/overview/ Taming the Complexity of Artifact Reproducibility

Test with other scenarios Example @KIT: one-year research project of three students Based upon a shared simulative environment Elaboration of common tools, traffic, topologies, and implementations  Results Support for NS3, OMNeT++ and Mininet Unified data format for traffic and topologies SDN app integration (POX) Exchange / Share of apps, scenarios, and simulator Other examples Portable workflow framework (CK), http://cknowledge.org/ Pantheon of Congestion Control, http://pantheon.stanford.edu/overview/ Benefits? Comparability Advanced review Test with other scenarios Taming the Complexity of Artifact Reproducibility

Observation 3 Taming the Complexity of Artifact Reproducibility

Observation 3 use “complex” setups, e.g., multiple approaches Individual complex evaluation parts are hard to reproduce use “complex” setups, e.g., multiple approaches 35% provide a direct link to the code 52% No step-by-step documentations for reproducibility ~0% Taming the Complexity of Artifact Reproducibility

Proposal: Provisioning of the Evaluation Setup Provide self-deploying evaluation environments $ git clone <paper repository> myFolder $ cd myFolder $ vagrant up $ … wait … $ vagrant ssh $ ./run_all_experiments.sh $ cd /results Taming the Complexity of Artifact Reproducibility

Proposal: Provisioning of the Evaluation Setup Provide self-deploying evaluation environments $ git clone <paper repository> myFolder $ cd myFolder $ vagrant up $ … wait … $ vagrant ssh $ ./run_all_experiments.sh $ cd /results Benefits? Original experiment One click reproducibility Taming the Complexity of Artifact Reproducibility

Challenges Taming the Complexity of Artifact Reproducibility

Challenges Paradigm shift How to get started? Conflicting priorities high Every researcher has to fulfill the standard Overhead to establish BCPs Every researcher can do whatever he wants high Overhead for Reproducibility Taming the Complexity of Artifact Reproducibility

Scope Taming the Complexity of Artifact Reproducibility

Reproducibility Landscape Get Incentives Get BCPs - Badges - Contest - Fast Track - Traffic Features - Testbeds - Environments Change Paper Change Portals Change Review Change the paper Reproducibility considerations More details Enhance portals Artifact upload Change review Questions about reproducibility Review of artifacts New incentives Highlight reproducible papers (e.g. badges) Reproducibility challenge/ networking contest in reproducibility Requirement for submission Journal fast tracking Best current practice Recommended traffic features Testbeds (Responsebility, Load, Standardization) - More Details - Considerations - Upload - Meta-Artifacts - Artifacts - Questions Taming the Complexity of Artifact Reproducibility

Reproducibility by Design Summary Sharing Environment Set up Provision Docu- mentation Reproducibility by Design Taming the Complexity of Artifact Reproducibility

QuestioNS? Taming the Complexity of Artifact Reproducibility