Networking the Cloud Presenter: b97901184 電機三 姜慧如.

Slides:



Advertisements
Similar presentations
Data Center Networking
Advertisements

Data Center Networking with Multipath TCP
Improving Datacenter Performance and Robustness with Multipath TCP
All Rights Reserved © Alcatel-Lucent 2009 Enhancing Dynamic Cloud-based Services using Network Virtualization F. Hao, T.V. Lakshman, Sarit Mukherjee, H.
2  Industry trends and challenges  Windows Server 2012: Beyond virtualization  Complete virtualization platform  Improved scalability and performance.
Mobility Jennifer Rexford COS 461: Computer Networks Lectures: MW 10-10:50am in Architecture N101
NDN in Local Area Networks Junxiao Shi The University of Arizona
The Case for Enterprise Ready Virtual Private Clouds Timothy Wood, Alexandre Gerber *, K.K. Ramakrishnan *, Jacobus van der Merwe *, and Prashant Shenoy.
Jaringan Komputer Lanjut Packet Switching Network.
PortLand: A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric. Presented by: Vinuthna Nalluri Shiva Srivastava.
Data Center Networking Major Theme: What are new networking issues posed by large-scale data centers? Network Architecture? Topology design? Addressing?
Data Center Fabrics. Forwarding Today Layer 3 approach: – Assign IP addresses to hosts hierarchically based on their directly connected switch. – Use.
Data Center Network Topologies: VL2 (Virtual Layer 2) Hakim Weatherspoon Assistant Professor, Dept of Computer Science CS 5413: High Performance Systems.
Resource Pooling A system exhibits complete resource pooling if it behaves as if there was a single pooled resource. The Internet has many mechanisms for.
Datacenter Network Topologies
Virtual Layer 2: A Scalable and Flexible Data-Center Network Work with Albert Greenberg, James R. Hamilton, Navendu Jain, Srikanth Kandula, Parantap Lahiri,
CS335 Networking & Network Administration Tuesday, April 20, 2010.
ProActive Routing In Scalable Data Centers with PARIS Joint work with Dushyant Arora + and Jennifer Rexford* + Arista Networks *Princeton University Theophilus.
COS 461: Computer Networks
VL2: A Scalable and Flexible data Center Network
A Scalable, Commodity Data Center Network Architecture Mohammad Al-Fares, Alexander Loukissas, Amin Vahdat Presented by Gregory Peaker and Tyler Maclean.
Jennifer Rexford Princeton University MW 11:00am-12:20pm Data-Center Traffic Management COS 597E: Software Defined Networking.
Jennifer Rexford Fall 2010 (TTh 1:30-2:50 in COS 302) COS 561: Advanced Computer Networks Data.
A Scalable, Commodity Data Center Network Architecture.
Datacenter Networks Mike Freedman COS 461: Computer Networks
Data Center Networks Jennifer Rexford COS 461: Computer Networks Lectures: MW 10-10:50am in Architecture N101
WAN Technologies.
Virtual LANs. VLAN introduction VLANs logically segment switched networks based on the functions, project teams, or applications of the organization regardless.
InterVLAN Routing Design and Implementation. What Routers Do Intelligent, dynamic routing protocols for packet transport Packet filtering capabilities.
VL2 – A Scalable & Flexible Data Center Network Authors: Greenberg et al Presenter: Syed M Irteza – LUMS CS678: 2 April 2013.
Routing & Architecture
MDC-B350: Part 1 Room: You are in it Time: Now What we introduced in SP1 recap How to setup your datacenter networking from scratch What’s new in R2.
DARD: Distributed Adaptive Routing for Datacenter Networks Xin Wu, Xiaowei Yang.
1 Energy in Networks & Data Center Networks Department of EECS University of Tennessee, Knoxville Yanjun Yao.
LAN Switching and Wireless – Chapter 1
Cloud Scale Performance & Diagnosability Comprehensive SDN Core Infrastructure Enhancements vRSS Remote Live Monitoring NIC Teaming Hyper-V Network.
Floodless in SEATTLE : A Scalable Ethernet ArchiTecTure for Large Enterprises. Changhoon Kim, Matthew Caesar and Jenifer Rexford. Princeton University.
VL2: A Scalable and Flexible Data Center Network Albert Greenberg, James R. Hamilton, Navendu Jain, Srikanth Kandula, Changhoon Kim, Parantap Lahiri, David.
6.888: Lecture 2 Data Center Network Architectures Mohammad Alizadeh Spring 2016  Slides adapted from presentations by Albert Greenberg and Changhoon.
Ananta: Cloud Scale Load Balancing Presenter: Donghwi Kim 1.
@projectcalico Sponsored by Simple, Secure, Scalable networking for the virtualized datacentre UKNOF 33 Ed 19 th January 2016.
Data Center Networking Major Theme: What are new networking issues posed by large-scale data centers? Network Architecture? Topology design? Addressing?
VL2: A Scalable and Flexible Data Center Network
Data Center Architectures
Data Center Networking
Instructor Materials Chapter 1: LAN Design
CIS 700-5: The Design and Implementation of Cloud Networks
Lecture 2: Cloud Computing
Data Center Network Topologies II
Bentley Systems, Incorporated
Heitor Moraes, Marcos Vieira, Italo Cunha, Dorgival Guedes
Data Centers: Network Architecture
Scaling the Network: The Internet Protocol
Revisiting Ethernet: Plug-and-play made scalable and efficient
Data Center Network Architectures
Improving Datacenter Performance and Robustness with Multipath TCP
Virtual LANs.
GGF15 – Grids and Network Virtualization
湖南大学-信息科学与工程学院-计算机与科学系
NTHU CS5421 Cloud Computing
VL2: A Scalable and Flexible Data Center Network
Internet and Web Simple client-server model
Specialized Cloud Architectures
Data Center Architectures
Data Center Networks Mohammad Alizadeh Fall 2018
Scaling the Network: The Internet Protocol
IS-IS VPLS for Data Center Network draft-xu-l2vpn-vpls-isis-02
Lecture 8, Computer Networks (198:552)
Lecture 9, Computer Networks (198:552)
Data Center Traffic Engineering
Presentation transcript:

Networking the Cloud Presenter: b 電機三 姜慧如

 Data Centers holding tens to hundreds of thousands of servers.  Concurrently supporting a large number of distinct services.  Economies of scale.  Dynamically reallocate servers among services as workload pattern changes.  High utilization is needed.  Agility!

 Agility: Any machine to be able to play any role.  “Any Service, any Server.”

 Ugly Secret: 30% utilization is considered good.  Uneven application fit: -- Each server has CPU, memory, disk: most applications exhaust one resource, stranding the others.  Long provisioning timescales: -- New servers purchased quarterly at best.  Uncertainty in demand: -- Demand for a new service can spike quickly.  Risk management: -- Not having spare servers to meet demand brings failure just when success is at hand.  Session state and storage constraints: -- If the world were stateless servers….

 Workload management -- Means for rapidly installing a service’s code on a server.  Virtual Machines, disk images.  Storage management -- Means for a server to access persistent data.  Distributed filesystems.  Network -- Means for communicating with other servers, regardless of where they are in the data center. But:  Today’s data center network prevent agility in several ways.

 Static network assignment  Fragmentation of resource  Poor server to server connectivity  Traffics affects each other  Poor reliability and utilization CR AR S S S S S S S S A A A A A A … S S S S A A A A A A …... S S S S S S S S A A A A A A … S S S S A A A A A A … I want more I have spare ones, but… 1:5 1:80 1:240

 Achieve scale by assigning servers topologically related IP addresses and dividing servers among VLANs.   Limited utility of VMs, cannot migrate out the original VLAN while keeping the same IP address.   Fragmentation of address space.   Configuration needed when reassigned to different services.

9 1. L2 semantics 2. Uniform high capacity 3. Performance isolation A A A A A A … A A A A A A …... A A A A A A … A A A A A A … CR AR S S S S S S S S S S S S S S S S S S S S S S S S A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A...

Developers want a mental model where all their servers, and only their servers, are plugged into an Ethernet switch. 1. Layer-2 semantics -- Flat addressing, so any server can have any IP Address. -- Server configuration is the same as in a LAN. -- VM keeps the same IP address even after migration 2. Uniform high capacity -- Capacity between servers limited only by their NICs. -- No need to consider topology when adding servers. 3. Performance isolation -- Traffic of one service should be unaffected by others.

11 Solution Approach Objective 2. Uniform high capacity between servers Enforce hose model using existing mechanisms only Employ flat addressing 1. Layer-2 semantics 3. Performance Isolation Guarantee bandwidth for hose-model traffic Flow-based random traffic indirection (Valiant LB) Name-location separation & resolution service TCP “Hose”: each node has ingress/egress bandwidth constraints

 Ethernet switching (layer 2)  Cheaper switch equipment  Fixed addresses and auto-configuration  Seamless mobility, migration, and failover  IP routing (layer 3)  Scalability through hierarchical addressing  Efficiency through shortest-path routing  Multipath routing through equal-cost multipath  So, like in enterprises…  Data centers often connect layer-2 islands by IP routers 12

 Data-Center traffic analysis:  DC traffic != Internet traffic  Traffic volume between servers to entering/leaving data center is 4:1  Demand for bandwidth between servers growing faster  Network is the bottleneck of computation  Flow distribution analysis:  Majority of flows are small, biggest flow size is 100MB  The distribution of internal flows is simpler and more uniform  50% times of 10 concurrent flows, 5% greater than 80 concurrent flows

 Traffic matrix analysis:  Poor summarizing of traffic patterns  Instability of traffic patterns  Failure characteristics:  Pattern of networking equipment failures: 95% 10 days  No obvious way to eliminate all failures from the top of the hierarchy

 Flat Addressing: Allow service instances (ex. virtual machines) to be placed anywhere in the network.  Valiant Load Balancing: (Randomly) Spread network traffic uniformly across network paths.  End-system based address resolution: To scale to large server pools, without introducing complexity to the network control plane.

 Design principle:  Randomizing to cope with volatility: ▪ Using Valiant Load Balancing (VLB) to do destination independent traffic spreading across multiple intermediate nodes  Building on proven networking technology: ▪ Using IP routing and forwarding technologies available in commodity switches  Separating names from locators: ▪ Using directory system to maintain the mapping between names and locations  Embracing end systems: ▪ A VL2 agent at each server

18... TOR 20 Servers Int... Aggr K aggr switches with D ports 20*(DK/4) Servers Offer huge aggr capacity & multi paths at modest cost

20 xy payload T3T3 y z T5T5 z I ANY Cope with arbitrary TMs with very little overhead Links used for up paths Links used for down paths T1T1 T2T2 T3T3 T4T4 T5T5 T6T6 [ ECMP + IP Anycast ] Harness huge bisection bandwidth Obviate esoteric traffic engineering or optimization Ensure robustness to failures Work with switch mechanisms available today [ ECMP + IP Anycast ] Harness huge bisection bandwidth Obviate esoteric traffic engineering or optimization Ensure robustness to failures Work with switch mechanisms available today 1. Must spread traffic 2. Must ensure dst independence 1. Must spread traffic 2. Must ensure dst independence Equal Cost Multi Path Forwarding

xy payload T3T3 y z T5T5 z I ANY Links used for up paths Links used for down paths T1T1 T2T2 T3T3 T4T4 T5T5 T6T6

 How Smart servers use Dumb switches– Encapsulation.  Commodity switches have simple forwarding primitives.  Complexity moved to servers -- computing the headers.

RSM DS RSM DS RSM DS Agent... Agent... Directory Servers RSM Servers 2. Reply 1. Lookup “Lookup” 5. Ack 2. Set 4. Ack (6. Disseminate) 3. Replicate 1. Update “Update”

payload ToR 3... y x Servers use flat names Switches run link-state routing and maintain only switch-level topology yz payload ToR 4 z ToR 2 ToR 4 ToR 1 ToR 3 y, z payload ToR 3 z... Directory Service … x  ToR 2 y  ToR 3 z  ToR 4 … Lookup & Response … x  ToR 2 y  ToR 3 z  ToR 3 … LAs AAs

 Data center Oses already heavily modified for VMs, storage clouds, etc.  No change to apps or clients outside DC.

 Uniform high capacity:  All-to-all data shuffle stress test: ▪ 75 servers, deliver 500MB ▪ Maximal achievable goodput is 62.3 ▪ VL2 network efficiency as 58.8/62.3 = 94%

 Performance isolation:  Two types of services: ▪ Service one: 18 servers do single TCP transfer all the time ▪ Service two: 19 servers starts a 8GB transfer over TCP every 2 seconds ▪ Service two: 19 servers burst short TCP connections

 Convergence after link failures  75 servers  All-to-all data shuffle  Disconnect links between intermediate and aggregation switches