Theophilus Benson*, Ashok Anand*, Aditya Akella*, Ming Zhang + *University of Wisconsin, Madison + Microsoft Research.

Slides:



Advertisements
Similar presentations
Optimizing Cost and Performance for Multihoming Nick Feamster CS 6250 Fall 2011.
Advertisements

Data Center Networking with Multipath TCP
Improving Datacenter Performance and Robustness with Multipath TCP
U NDERSTANDING D ATA C ENTER T RAFFIC C HARACTERISTICS Theophilus Benson 1, Ashok Anand 1, Aditya Akella 1, Ming Zhang 2 University Of Wisconsin – Madison.
Abhigyan, Aditya Mishra, Vikas Kumar, Arun Venkataramani University of Massachusetts Amherst 1.
Big Data + SDN SDN Abstractions. The Story Thus Far Different types of traffic in clusters Background Traffic – Bulk transfers – Control messages Active.
SDN Controller Challenges
Traffic Engineering with Forward Fault Correction (FFC)
Software-defined networking: Change is hard Ratul Mahajan with Chi-Yao Hong, Rohan Gandhi, Xin Jin, Harry Liu, Vijay Gill, Srikanth Kandula, Mohan Nanduri,
SDN + Storage.
Slick: A control plane for middleboxes Bilal Anwer, Theophilus Benson, Dave Levin, Nick Feamster, Jennifer Rexford Supported by DARPA through the U.S.
Improving Datacenter Performance and Robustness with Multipath TCP Costin Raiciu, Sebastien Barre, Christopher Pluntke, Adam Greenhalgh, Damon Wischik,
1 Advancing Supercomputer Performance Through Interconnection Topology Synthesis Yi Zhu, Michael Taylor, Scott B. Baden and Chung-Kuan Cheng Department.
1 EL736 Communications Networks II: Design and Algorithms Class8: Networks with Shortest-Path Routing Yong Liu 10/31/2007.
SIGCOMM 2003 Making Intra-Domain Routing Robust to Changing and Uncertain Traffic Demands: Understanding Fundamental Tradeoffs David Applegate Edith Cohen.
“ElasticTree: Saving energy in data center networks“ by Brandon Heller, Seetharaman, Mahadevan, Yiakoumis, Sharma, Banerjee, McKeown presented by Nicoara.
Network Architecture for Joint Failure Recovery and Traffic Engineering Martin Suchara in collaboration with: D. Xu, R. Doverspike, D. Johnson and J. Rexford.
Traffic Engineering With Traditional IP Routing Protocols
11 Quantifying Benefits of Traffic Information Provision under Stochastic Demand and Capacity Conditions: A Multi-day Traffic Equilibrium Approach Mingxin.
Rethinking Internet Traffic Management: From Multiple Decompositions to a Practical Protocol Jiayue He Princeton University Joint work with Martin Suchara,
Data Center Traffic and Measurements Hakim Weatherspoon Assistant Professor, Dept of Computer Science CS 5413: High Performance Systems and Networking.
ProActive Routing In Scalable Data Centers with PARIS Joint work with Dushyant Arora + and Jennifer Rexford* + Arista Networks *Princeton University Theophilus.
Demystifying and Controlling the Performance of Data Center Networks.
MATE: MPLS Adaptive Traffic Engineering Anwar Elwalid, et. al. IEEE INFOCOM 2001.
Practical TDMA for Datacenter Ethernet
ElasticTree: Saving Energy in Data Center Networks 許倫愷 2013/5/28.
Identifying and Using Energy Critical Paths Nedeljko Vasić with Dejan Novaković, Satyam Shekhar, Prateek Bhurat, Marco Canini, and Dejan Kostić EPFL, Switzerland.
Curbing Delays in Datacenters: Need Time to Save Time? Mohammad Alizadeh Sachin Katti, Balaji Prabhakar Insieme Networks Stanford University 1.
DaVinci: Dynamically Adaptive Virtual Networks for a Customized Internet Jennifer Rexford Princeton University With Jiayue He, Rui Zhang-Shen, Ying Li,
Lecture 15. IGP and MPLS D. Moltchanov, TUT, Spring 2008 D. Moltchanov, TUT, Spring 2015.
Shannon Lab 1AT&T – Research Traffic Engineering with Estimated Traffic Matrices Matthew Roughan Mikkel Thorup
VL2 – A Scalable & Flexible Data Center Network Authors: Greenberg et al Presenter: Syed M Irteza – LUMS CS678: 2 April 2013.
Network Aware Resource Allocation in Distributed Clouds.
David G. Andersen CMU Guohui Wang, T. S. Eugene Ng Rice Michael Kaminsky, Dina Papagiannaki, Michael A. Kozuch, Michael Ryan Intel Labs Pittsburgh 1 c-Through:
DARD: Distributed Adaptive Routing for Datacenter Networks Xin Wu, Xiaowei Yang.
On the Data Path Performance of Leaf-Spine Datacenter Fabrics Mohammad Alizadeh Joint with: Tom Edsall 1.
Budget-based Control for Interactive Services with Partial Execution 1 Yuxiong He, Zihao Ye, Qiang Fu, Sameh Elnikety Microsoft Research.
CloudNaaS: A Cloud Networking Platform for Enterprise Applications Theophilus Benson*, Aditya Akella*, Anees Shaikh +, Sambit Sahu + (*University of Wisconsin,
VL2: A Scalable and Flexible Data Center Network Albert Greenberg, James R. Hamilton, Navendu Jain, Srikanth Kandula, Changhoon Kim, Parantap Lahiri, David.
Demystifying and Controlling the Performance of Big Data Jobs Theophilus Benson Duke University.
DaVinci: Dynamically Adaptive Virtual Networks for a Customized Internet Jiayue He, Rui Zhang-Shen, Ying Li, Cheng-Yen Lee, Jennifer Rexford, and Mung.
Towards a More Fair and Robust Internet Backbone Year 1 Status Report Rene Cruz, Tara Javidi, Bill Lin Center for Networked Systems University of California,
Intradomain Traffic Engineering By Behzad Akbari These slides are based in part upon slides of J. Rexford (Princeton university)
6 December On Selfish Routing in Internet-like Environments paper by Lili Qiu, Yang Richard Yang, Yin Zhang, Scott Shenker presentation by Ed Spitznagel.
Symbiotic Routing in Future Data Centers Hussam Abu-Libdeh Paolo Costa Antony Rowstron Greg O’Shea Austin Donnelly MICROSOFT RESEARCH Presented By Deng.
1 An Arc-Path Model for OSPF Weight Setting Problem Dr.Jeffery Kennington Anusha Madhavan.
Network Traffic Characteristics of Data Centers in the Wild
1 Slides by Yong Liu 1, Deep Medhi 2, and Michał Pióro 3 1 Polytechnic University, New York, USA 2 University of Missouri-Kansas City, USA 3 Warsaw University.
Jiaxin Cao, Rui Xia, Pengkun Yang, Chuanxiong Guo,
DECOR: A Distributed Coordinated Resource Monitoring System Shan-Hsiang Shen Aditya Akella.
R2C2: A Network Stack for Rack-scale Computers Paolo Costa, Hitesh Ballani, Kaveh Razavi, Ian Kash Microsoft Research Cambridge EECS 582 – W161.
1 Scalability and Accuracy in a Large-Scale Network Emulator Nov. 12, 2003 Byung-Gon Chun.
VL2: A Scalable and Flexible Data Center Network
Resilient Datacenter Load Balancing in the Wild
How I Learned to Stop Worrying About the Core and Love the Edge
University of Maryland College Park
Data Center Network Architectures
OTCP: SDN-Managed Congestion Control for Data Center Networks
Hydra: Leveraging Functional Slicing for Efficient Distributed SDN Controllers Yiyang Chang, Ashkan Rezaei, Balajee Vamanan, Jahangir Hasan, Sanjay Rao.
ECE 544: Traffic engineering (supplement)
Improving Datacenter Performance and Robustness with Multipath TCP
Improving Datacenter Performance and Robustness with Multipath TCP
ElasticTree Michael Fruchtman.
ISP and Egress Path Selection for Multihomed Networks
Congestion-Aware Load Balancing at the Virtual Edge
VL2: A Scalable and Flexible Data Center Network
Centralized Arbitration for Data Centers
Congestion-Aware Load Balancing at the Virtual Edge
Towards Predictable Datacenter Networks
Data Center Traffic Engineering
Presentation transcript:

Theophilus Benson*, Ashok Anand*, Aditya Akella*, Ming Zhang + *University of Wisconsin, Madison + Microsoft Research

Why are Data Centers Important?  Host a variety of applications  Web portals  Web services  Multimedia applications  Chat/IM  DC performance  app performance  Congestion impacts app latency

Why are Data Centers Important?  Poor performance  loss of revenue  Traffic engineering is crucial

Contributions  Study effectiveness of current approach  Developed guidelines for ideal approach  Designed and implemented MicroTE  Showed effectiveness in mitigating congestion

Road Map  Background  Traffic Engineering in data centers  Design goals for ideal TE  MicroTE  Evaluation  Conclusion

Options for TE in Data Centers?  Current supported techniques  Equal Cost MultiPath (ECMP)  Spanning Tree Protocol (STP)  Proposed  Fat-Tree, VL2  Other existing WAN techniques  COPE,…, OSPF link tuning

How do we evaluate TE?  Simulator  Input: Traffic matrix, topology, traffic engineering  Output: Link utilization  Optimal TE  Route traffic using knowledge of future TM  Data center traces  Cloud data center (CLD)  Map-reduce app  ~1500 servers  University data center (UNV)  3-Tier Web apps  ~500 servers

Draw Backs of Existing TE  STP does not use multiple path  40% worst than optimal  ECMP does not adapt to burstiness  15% worst than optimal

Draw Backs of Proposed TE  Fat-Tree  Rehash flows  Local opt. != global opt.  VL2  ECMP on VLB architecture  Coarse grained flow assignment  VL2 & Fat-Tree do not adapt to burstiness

Draw Backs of Other Approaches  COPE …. OSPF link tuning Requires hours or days of predictable traffic Traffic unpredictable over 100 secs (Maltz. et al) Ingress Egress x

Design Goals for Ideal TE

Design Requirements for TE  Calculate paths & reconfigure network  Use all network paths  Use global view  Avoid local optimals  Must react quickly  React to burstiness  How predictable is traffic? ….

Is Data Center Traffic Predictable?  YES! 27% or more of traffic matrix is predictable  Manage predictable traffic more intelligently 99% 27%

How Long is Traffic Predictable?  Different patterns of predictability  1 second of historical data able to predict future 1.5 –

MicroTE

MicroTE: Architecture  Global view:  Created by network controller  React to predictable traffic:  Routing component tracks demand history  All N/W paths:  Routing component creates routes using all paths Monitoring Component Routing Component Network Controller

Architectural Questions  Efficiently gather network state?  Determine predictable traffic?  Generate and calculate new routes?  Install network state?

Architectural Questions  Efficiently gather network state?  Determine predictable traffic?  Generate and calculate new routes?  Install network state?

Monitoring Component  Efficiently gather TM  Only one server per ToR monitors traffic  Transfer changed portion of TM  Compress data  Tracking predictability  Calculate EWMA over TM (every second)  Empirically derived alpha of 0.2  Use time-bins of 0.1 seconds

Routing Component Install routes Calculate network routes for predictable traffic Set ECMP for unpredictable traffic Determine predictable ToRs New Global View

Routing Predictable Traffic  LP formulation  Constraints  Flow conservation  Capacity constraint  Use K-equal length paths  Objective  Minimize link utilization  Bin-packing heuristic  Sort flows in decreasing order  Place on link with greatest capacity

Implementation  Changes to data center  Switch  Install OpenFlow firmware  End hosts  Add kernel module  New component  Network controller  C++ NOX modules

Evaluation

Evaluation: Motivating Questions  How does MicroTE Compare to Optimal?  How does MicroTE perform under varying levels of predictability?  How does MicroTE scale to large DCN?  What overheard does MicroTE impose?

Evaluation: Motivating Questions  How does MicroTE Compare to Optimal?  How does MicroTE perform under varying levels of predictability?  How does MicroTE scale to large DCN?  What overheard does MicroTE impose?

How do we evaluate TE?  Simulator  Input: Traffic matrix, topology, traffic engineering  Output: Link utilization  Optimal TE  Route traffic using knowledge of future TM  Data center traces  Cloud data center (CLD)  Map-reduce app  ~1500 servers  University data center (UNV)  3-Tier Web apps  ~400 servers

Performing Under Realistic Traffic  Significantly outperforms ECMP  Slightly worse than optimal (1%-5%)  Bin-packing and LP of comparable performance

Performance Versus Predictability  Low predictability  performance is similar to ECMP

Performance Versus Predictability  Low predictability  performance is similar to ECMP  High predictability  performance is comparable to Optimal  MicroTE adjusts according to predictability

Scaling to Large DC  Reactiveness to traffic changes  Gathering network state  Bounded by network transfer time  Route computation  Flow installation  Bounded by switch installation time DC Size (# servers) LPHeuristic 2K0.04s0.007s 4K0.6s0.055s 8K6.85s0.25s 16K-0.98s

Other Related Works  SPAIN/MultiPath TCP  Use multiple paths  better DC utilization  Eliminating congestion isn’t goal  Helios/Flyways  Require augmenting existing DC  Hedera/Fat-Tree  Require forklift upgrade: full bisection

Conclusion  Study existing TE  Found them lacking (15-40%)  Study data center traffic  Discovered traffic predictability (27% for 2 secs)  Developed guidelines for ideal TE  Designed and implemented MicroTE  Brings state of the art within 1-5% of Ideal  Efficiently scales to large DC (16K servers)

Thank You  MicroTE adopts to low level properties of traffic.  Can we develop techniques that cooperate with applications?