Demystifying and Controlling the Performance of Data Center Networks.

Slides:



Advertisements
Similar presentations
Data Center Networking with Multipath TCP
Advertisements

U NDERSTANDING D ATA C ENTER T RAFFIC C HARACTERISTICS Theophilus Benson 1, Ashok Anand 1, Aditya Akella 1, Ming Zhang 2 University Of Wisconsin – Madison.
SDN Controller Challenges
Traffic Engineering with Forward Fault Correction (FFC)
SDN + Storage.
Data Center Fabrics. Forwarding Today Layer 3 approach: – Assign IP addresses to hosts hierarchically based on their directly connected switch. – Use.
Applying NOX to the Datacenter Arsalan Tavakoli, Martin Casado, Teemu Koponen, and Scott Shenker 10/22/2009Hot Topics in Networks Workshop 2009.
Improving Datacenter Performance and Robustness with Multipath TCP Costin Raiciu, Sebastien Barre, Christopher Pluntke, Adam Greenhalgh, Damon Wischik,
1 EL736 Communications Networks II: Design and Algorithms Class8: Networks with Shortest-Path Routing Yong Liu 10/31/2007.
Towards Virtual Routers as a Service 6th GI/ITG KuVS Workshop on “Future Internet” November 22, 2010 Hannover Zdravko Bozakov.
Scalable and Crash-Tolerant Load Balancing based on Switch Migration
“ElasticTree: Saving energy in data center networks“ by Brandon Heller, Seetharaman, Mahadevan, Yiakoumis, Sharma, Banerjee, McKeown presented by Nicoara.
Network Architecture for Joint Failure Recovery and Traffic Engineering Martin Suchara in collaboration with: D. Xu, R. Doverspike, D. Johnson and J. Rexford.
Reconfigurable Network Topologies at Rack Scale
Datacenter Network Topologies
Network Traffic Measurement and Modeling CSCI 780, Fall 2005.
Rethinking Internet Traffic Management: From Multiple Decompositions to a Practical Protocol Jiayue He Princeton University Joint work with Martin Suchara,
Measuring a (MapReduce) Data Center Srikanth KandulaSudipta SenguptaAlbert Greenberg Parveen Patel Ronnie Chaiken.
Data Center Traffic and Measurements Hakim Weatherspoon Assistant Professor, Dept of Computer Science CS 5413: High Performance Systems and Networking.
ProActive Routing In Scalable Data Centers with PARIS Joint work with Dushyant Arora + and Jennifer Rexford* + Arista Networks *Princeton University Theophilus.
Multipath Protocol for Delay-Sensitive Traffic Jennifer Rexford Princeton University Joint work with Umar Javed, Martin Suchara, and Jiayue He
Understanding Network Failures in Data Centers: Measurement, Analysis and Implications Phillipa Gill University of Toronto Navendu Jain & Nachiappan Nagappan.
A Scalable, Commodity Data Center Network Architecture.
Modeling and Evaluation of Fibre Channel Storage Area Networks Xavier Molero, Federico Silla, Vicente Santonia and Jose Duato.
MATE: MPLS Adaptive Traffic Engineering Anwar Elwalid, et. al. IEEE INFOCOM 2001.
Practical TDMA for Datacenter Ethernet
ElasticTree: Saving Energy in Data Center Networks 許倫愷 2013/5/28.
Distributed Quality-of-Service Routing of Best Constrained Shortest Paths. Abdelhamid MELLOUK, Said HOCEINI, Farid BAGUENINE, Mustapha CHEURFA Computers.
Identifying and Using Energy Critical Paths Nedeljko Vasić with Dejan Novaković, Satyam Shekhar, Prateek Bhurat, Marco Canini, and Dejan Kostić EPFL, Switzerland.
Department of Computer Science Engineering SRM University
DaVinci: Dynamically Adaptive Virtual Networks for a Customized Internet Jennifer Rexford Princeton University With Jiayue He, Rui Zhang-Shen, Ying Li,
1 Kommunikatsiooniteenuste arendus IRT0080 Loeng 7 Avo Ots telekommunikatsiooni õppetool, TTÜ raadio- ja sidetehnika inst.
VL2 – A Scalable & Flexible Data Center Network Authors: Greenberg et al Presenter: Syed M Irteza – LUMS CS678: 2 April 2013.
Network Aware Resource Allocation in Distributed Clouds.
A Distributed Energy Saving Approach for Ethernet Switches in Data Centers Weisheng Si 1, Javid Taheri 2, Albert Zomaya 2 1 School of Computing, Engineering,
David G. Andersen CMU Guohui Wang, T. S. Eugene Ng Rice Michael Kaminsky, Dina Papagiannaki, Michael A. Kozuch, Michael Ryan Intel Labs Pittsburgh 1 c-Through:
DARD: Distributed Adaptive Routing for Datacenter Networks Xin Wu, Xiaowei Yang.
On the Data Path Performance of Leaf-Spine Datacenter Fabrics Mohammad Alizadeh Joint with: Tom Edsall 1.
Budapest University of Technology and Economics Department of Telecommunications and Media Informatics Optimized QoS Protection of Ethernet Trees Tibor.
CloudNaaS: A Cloud Networking Platform for Enterprise Applications Theophilus Benson*, Aditya Akella*, Anees Shaikh +, Sambit Sahu + (*University of Wisconsin,
VL2: A Scalable and Flexible Data Center Network Albert Greenberg, James R. Hamilton, Navendu Jain, Srikanth Kandula, Changhoon Kim, Parantap Lahiri, David.
Demystifying and Controlling the Performance of Big Data Jobs Theophilus Benson Duke University.
The Only Constant is Change: Incorporating Time-Varying Bandwidth Reservations in Data Centers Di Xie, Ning Ding, Y. Charlie Hu, Ramana Kompella 1.
DaVinci: Dynamically Adaptive Virtual Networks for a Customized Internet Jiayue He, Rui Zhang-Shen, Ying Li, Cheng-Yen Lee, Jennifer Rexford, and Mung.
Intradomain Traffic Engineering By Behzad Akbari These slides are based in part upon slides of J. Rexford (Princeton university)
Network-Aware Scheduling for Data-Parallel Jobs: Plan When You Can
6 December On Selfish Routing in Internet-like Environments paper by Lili Qiu, Yang Richard Yang, Yin Zhang, Scott Shenker presentation by Ed Spitznagel.
GreenCloud: A Packet-level Simulator of Energy-aware Cloud Computing Data Centers Dzmitry Kliazovich ERCIM Fellow University of Luxembourg Apr 16, 2010.
Network Traffic Characteristics of Data Centers in the Wild
1 Slides by Yong Liu 1, Deep Medhi 2, and Michał Pióro 3 1 Polytechnic University, New York, USA 2 University of Missouri-Kansas City, USA 3 Warsaw University.
Internet Measurement and Analysis Vinay Ribeiro Shriram Sarvotham Rolf Riedi Richard Baraniuk Rice University.
Theophilus Benson*, Ashok Anand*, Aditya Akella*, Ming Zhang + *University of Wisconsin, Madison + Microsoft Research.
1 Internet Traffic Measurement and Modeling Carey Williamson Department of Computer Science University of Calgary.
R2C2: A Network Stack for Rack-scale Computers Paolo Costa, Hitesh Ballani, Kaveh Razavi, Ian Kash Microsoft Research Cambridge EECS 582 – W161.
VL2: A Scalable and Flexible Data Center Network
Chen Qian, Xin Li University of Kentucky
Modeling and Evaluation of Fibre Channel Storage Area Networks
How I Learned to Stop Worrying About the Core and Love the Edge
University of Maryland College Park
ECE 544: Traffic engineering (supplement)
Improving Datacenter Performance and Robustness with Multipath TCP
Improving Datacenter Performance and Robustness with Multipath TCP
Dingming Wu+, Yiting Xia+*, Xiaoye Steven Sun+,
Multi-hop Coflow Routing and Scheduling in Data Centers
NTHU CS5421 Cloud Computing
VL2: A Scalable and Flexible Data Center Network
Data Center Architectures
Building a Smart Cloud Strategy
Towards Predictable Datacenter Networks
Data Center Traffic Engineering
Presentation transcript:

Demystifying and Controlling the Performance of Data Center Networks

Why are Data Centers Important? Internal users –Line-of-Business apps –Production test beds External users –Web portals –Web services –Multimedia applications –Chat/IM

Why are Data Centers Important? Poor performance  loss of revenue Understanding traffic is crucial Traffic engineering is crucial

Road Map Understanding Data center traffic Improving network level performance Ongoing work

Canonical Data Center Architecture Core (L3) Edge (L2) Top-of-Rack Aggregation (L2) Application servers

Dataset: Data Centers Studied DC RoleDC Name LocationNumber Devices UniversitiesEDU1US-Mid22 EDU2US-Mid36 EDU3US-Mid11 Private Enterprise PRV1US-Mid97 PRV2US-West100 Commercial Clouds CLD1US-West562 CLD2US-West763 CLD3US-East612 CLD4S. America427 CLD5S. America427  10 data centers  3 classes  Universities  Private enterprise  Clouds  Internal users  Univ/priv  Small  Local to campus  External users  Clouds  Large  Globally diverse

Dataset: Collection SNMP –Poll SNMP MIBs –Bytes-in/bytes-out/discards – > 10 Days –Averaged over 5 mins Packet Traces –Cisco port span –12 hours Topology –Cisco Discovery Protocol DC Name SNMPPacket Traces Topology EDU1Yes EDU2Yes EDU3Yes PRV1Yes PRV2Yes CLD1YesNo CLD2YesNo CLD3YesNo CLD4YesNo CLD5YesNo

Canonical Data Center Architecture Core (L3) Edge (L2) Top-of-Rack Aggregation (L2) Application servers Packet Sniffers

Analyzing Packet Traces Transmission patterns of the applications Properties of packet crucial for –Understanding effectiveness of techniques ON-OFF traffic at edges –Binned in 15 and 100 m. secs –We observe that ON-OFF persists 9 Routing must react quickly to overcome bursts

Data-Center Traffic is Bursty Understanding arrival process –Range of acceptable models What is the arrival process? –Heavy-tail for the 3 distributions ON, OFF times, Inter-arrival, –Lognormal across all data centers Different from Pareto of WAN –Need new models 10 Data Center Off Period Dist ON periods Dist Inter-arrival Dist Prv2_1Lognormal Prv2_2Lognormal Prv2_3Lognormal Prv2_4Lognormal EDU1LognormalWeibull EDU2LognormalWeibull EDU3LognormalWeibull Need new models to generate traffic

Canonical Data Center Architecture Core (L3) Edge (L2) Top-of-Rack Aggregation (L2) Application servers

Intra-Rack Versus Extra-Rack Quantify amount of traffic using interconnect –Perspective for interconnect analysis Edge Application servers Extra-Rack Intra-Rack Extra-Rack = Sum of Uplinks Intra-Rack = Sum of Server Links – Extra-Rack

Intra-Rack Versus Extra-Rack Results Clouds: most traffic stays within a rack (75%) –Colocation of apps and dependent components Other DCs: > 50% leaves the rack –Un-optimized placement

Extra-Rack Traffic on DC Interconnect Utilization: core > agg > edge – Aggregation of many unto few Tail of core utilization differs –Hot-spots  links with > 70% util –Prevalence of hot-spots differs across data centers

Persistence of Core Hot-Spots Low persistence: PRV2, EDU1, EDU2, EDU3, CLD1, CLD3 High persistence/low prevalence: PRV1, CLD2 –2-8% are hotspots > 50% High persistence/high prevalence: CLD4, CLD5 –15% are hotspots > 50%

Prevalence of Core Hot-Spots Low persistence: very few concurrent hotspots High persistence: few concurrent hotspots High prevalence: < 25% are hotspots at any time Time (in Hours) 0.6% 0.0% 6.0% 24.0% Smart routing can better utilize core and avoid hotspots Smart routing can better utilize core and avoid hotspots

Observations from Interconnect Links utils low at edge and agg Core most utilized –Hot-spots exists (> 70% utilization) –< 25% links are hotspots –Loss occurs on less utilized links (< 70%) Implicating momentary bursts Time-of-Day variations exists –Variation an order of magnitude larger at core Apply these results to evaluate DC design requirements

Insights Gained 75% of traffic stays within a rack (Clouds) –Applications are not uniformly placed Traffic is bursty at the edge At most 25% of core links highly utilized –Effective routing algorithm to reduce utilization –Load balance across paths and migrate VMs

Road Map Understanding Data center traffic Improving network level performance Ongoing work

Options for TE in Data Centers? Current supported techniques –Equal Cost MultiPath (ECMP) –Spanning Tree Protocol (STP) Proposed –Fat-Tree, VL2 Other existing WAN techniques –COPE,…, OSPF link tuning

How do we evaluate TE? Simulator –Input: Traffic matrix, topology, traffic engineering –Output: Link utilization Optimal TE –Route traffic using knowledge of future TM Data center traces –Cloud data center (CLD) Map-reduce app ~1500 servers –University data center (UNV) 3-Tier Web apps ~500 servers

Draw Backs of Existing TE STP does not use multiple path –40% worst than optimal ECMP does not adapt to burstiness –15% worst than optimal

Design Goals for Ideal TE

Design Requirements for TE Calculate paths & reconfigure network –Use all network paths –Use global view Avoid local optimals –Must react quickly React to burstiness How predictable is traffic? ….

Is Data Center Traffic Predictable? YES! 27% or more of traffic matrix is predictable Manage predictable traffic more intelligently 99% 27%

How Long is Traffic Predictable? Different patterns of predictability 1 second of historical data able to predict future 1.5 –

MicroTE

MicroTE: Architecture Global view: –Created by network controller React to predictable traffic: –Routing component tracks demand history All N/W paths: –Routing component creates routes using all paths Monitoring Component Routing Component Network Controller

Architectural Questions Efficiently gather network state? Determine predictable traffic? Generate and calculate new routes? Install network state?

Architectural Questions Efficiently gather network state? Determine predictable traffic? Generate and calculate new routes? Install network state?

Monitoring Component Efficiently gather TM Only one server per ToR monitors traffic Transfer changed portion of TM Compress data Tracking predictability –Calculate EWMA over TM (every second) Empirically derived alpha of 0.2 Use time-bins of 0.1 seconds

Routing Component Install routes Calculate network routes for predictable traffic Set ECMP for unpredictable traffic Determine predictable ToRs New Global View

Routing Predictable Traffic LP formulation –Constraints Flow conservation Capacity constraint Use K-equal length paths –Objective Minimize link utilization Bin-packing heuristic –Sort flows in decreasing order –Place on link with greatest capacity

Implementation Changes to data center –Switch Install OpenFlow firmware –End hosts Add kernel module New component –Network controller C++ NOX modules

Evaluation

Evaluation: Motivating Questions How does MicroTE Compare to Optimal? How does MicroTE perform under varying levels of predictability? How does MicroTE scale to large DCN? What overheard does MicroTE impose?

Evaluation: Motivating Questions How does MicroTE Compare to Optimal? How does MicroTE perform under varying levels of predictability? How does MicroTE scale to large DCN? What overheard does MicroTE impose?

How do we evaluate TE? Simulator –Input: Traffic matrix, topology, traffic engineering –Output: Link utilization Optimal TE –Route traffic using knowledge of future TM Data center traces –Cloud data center (CLD) Map-reduce app ~1500 servers –University data center (UNV) 3-Tier Web apps ~400 servers

Performing Under Realistic Traffic Significantly outperforms ECMP Slightly worse than optimal (1%-5%) Bin-packing and LP of comparable performance

Performance Versus Predictability Low predictability  performance is similar to ECMP

Performance Versus Predictability Low predictability  performance is similar to ECMP High predictability  performance is comparable to Optimal MicroTE adjusts according to predictability

Conclusion Study existing TE –Found them lacking (15-40%) Study data center traffic –Discovered traffic predictability (27% for 2 secs) Developed guidelines for ideal TE Designed and implemented MicroTE –Brings state of the art within 1-5% of Ideal –Efficiently scales to large DC (16K servers)

Road Map Understanding Data center traffic Improving network level performance Ongoing work

Looking forward Stop treating the network as a carrier of bits Bits in the network have a meaning – Applications know this meaning. Can applications control networks? – E.g Map-reduce Scheduler performs network aware task placement and flow placement