Network Sharing Issues Lecture 15 Aditya Akella. Is this the biggest problem in cloud resource allocation? Why? Why not? How does the problem differ wrt.

Slides:



Advertisements
Similar presentations
Virtual Network Diagnosis as a Service Wenfei Wu (UW-Madison) Guohui Wang (Facebook) Aditya Akella (UW-Madison) Anees Shaikh (IBM System Networking)
Advertisements

Towards Predictable Datacenter Networks
Winter 2004 UCSC CMPE252B1 CMPE 257: Wireless and Mobile Networking SET 3f: Medium Access Control Protocols.
Sharing Cloud Networks Lucian Popa, Gautam Kumar, Mosharaf Chowdhury Arvind Krishnamurthy, Sylvia Ratnasamy, Ion Stoica UC Berkeley.
Architectures for Congestion-Sensitive Pricing of Network Services Thesis Defense by Murat Yuksel CS Department, RPI July 3 rd, 2002.
1 Traffic Engineering (TE). 2 Network Congestion Causes of congestion –Lack of network resources –Uneven distribution of traffic caused by current dynamic.
Cloud Computing Resource provisioning Keke Chen. Outline  For Web applications statistical Learning and automatic control for datacenters  For data.
CS 268: Lecture 8 Router Support for Congestion Control Ion Stoica Computer Science Division Department of Electrical Engineering and Computer Sciences.
FairCloud: Sharing the Network in Cloud Computing Computer Communication Review(2012) Arthur : Lucian Popa Arvind Krishnamurthy Sylvia Ratnasamy Ion Stoica.
Reciprocal Resource Fairness: Towards Cooperative Multiple-Resource Fair Sharing in IaaS Clouds School of Computer Engineering Nanyang Technological University,
My Point of View about Bandwidth Sharing Bin Wang.
Course Name- CSc 8320 Advanced Operating Systems Instructor- Dr. Yanqing Zhang Presented By- Sunny Shakya Latest AOS techniques, applications and future.
Ion Stoica, Scott Shenker, and Hui Zhang SIGCOMM’98, Vancouver, August 1998 subsequently IEEE/ACM Transactions on Networking 11(1), 2003, pp Presented.
Datacenter Network Topologies
Buffer Sizing for Congested Internet Links Chi Yin Cheung Cs 395 Advanced Networking.
Congestion Control and Resource Allocation
CS 268: Lecture 8 (Router Support for Congestion Control) Ion Stoica February 19, 2002.
Jerry Chou and Bill Lin University of California, San Diego
1 Emulating AQM from End Hosts Presenters: Syed Zaidi Ivor Rodrigues.
Using Prices to Allocate Resources at Access Points Jimmy Shih, Randy Katz, Anthony Joseph One Administrative Domain Access Point A Access Point B Network.
Core Stateless Fair Queueing Stoica, Shanker and Zhang - SIGCOMM 98 Rigorous fair Queueing requires per flow state: too costly in high speed core routers.
Core Stateless Fair Queueing Stoica, Shanker and Zhang - SIGCOMM 98 Fair Queueing requires per flow state: too costly in high speed core routers Yet, some.
DRFQ: Multi-Resource Fair Queueing for Packet Processing Ali Ghodsi 1,3, Vyas Sekar 2, Matei Zaharia 1, Ion Stoica 1 1 UC Berkeley, 2 Intel ISTC/Stony.
Sharing the Data Center Network Alan Shieh, Srikanth Kandula, Albert Greenberg, Changhoon Kim, Bikas Saha Microsoft Research, Cornell University, Windows.
Bargaining Towards Maximized Resource Utilization in Video Streaming Datacenters Yuan Feng 1, Baochun Li 1, and Bo Li 2 1 Department of Electrical and.
DaVinci: Dynamically Adaptive Virtual Networks for a Customized Internet Jennifer Rexford Princeton University With Jiayue He, Rui Zhang-Shen, Ying Li,
1 EL736 Communications Networks II: Design and Algorithms Class2: Network Design Problems -- Notation and Illustrations Yong Liu 09/12/2007.
CS640: Introduction to Computer Networks Aditya Akella Lecture 20 - Queuing and Basics of QoS.
Department of Computer Science at Florida State LFTI: A Performance Metric for Assessing Interconnect topology and routing design Background ‒ Innovations.
DRFQ: Multi-Resource Fair Queueing for Packet Processing Ali Ghodsi 1,3, Vyas Sekar 2, Matei Zaharia 1, Ion Stoica 1 1 UC Berkeley, 2 Intel ISTC/Stony.
The Only Constant is Change: Incorporating Time-Varying Bandwidth Reservations in Data Centers Di Xie, Ning Ding, Y. Charlie Hu, Ramana Kompella 1.
Computer Networks Performance Metrics. Performance Metrics Outline Generic Performance Metrics Network performance Measures Components of Hop and End-to-End.
E VALUATION OF F AIRNESS IN ICN X. De Foy, JC. Zuniga, B. Balazinski InterDigital
© Copyright 2010 Hewlett-Packard Development Company, L.P. 1 Jayaram Mudigonda, HP Labs Praveen Yalagandula, HP Labs Mohammad Al-Fares, UCSD Jeff Mogul,
Link Scheduling & Queuing COS 461: Computer Networks
Dominant Resource Fairness: Fair Allocation of Multiple Resource Types Ali Ghodsi, Matei Zaharia, Benjamin Hindman, Andy Konwinski, Scott Shenker, Ion.
Scalable Multi-Class Traffic Management in Data Center Backbone Networks Amitabha Ghosh (UtopiaCompression) Sangtae Ha (Princeton) Edward Crabbe (Google)
VL2: A Scalable and Flexible Data Center Network Albert Greenberg, James R. Hamilton, Navendu Jain, Srikanth Kandula, Changhoon Kim, Parantap Lahiri, David.
The Only Constant is Change: Incorporating Time-Varying Bandwidth Reservations in Data Centers Di Xie, Ning Ding, Y. Charlie Hu, Ramana Kompella 1.
Interconnect simulation. Different levels for Evaluating an architecture Numerical models – Mathematic formulations to obtain performance characteristics.
Using Map-reduce to Support MPMD Peng
CS640: Introduction to Computer Networks Aditya Akella Lecture 20 - Queuing and Basics of QoS.
Sep. 1, SIGCOMM '99 Dan Rubenstein1 The Impact of Multicast Layering on Network Fairness Dan Rubenstein Jim Kurose Don Towsley.
Towards Economic Fairness for Big Data Processing in Pay-as-you-go Cloud Computing Nanyang Technological University Shanjiang Tang, Bu-Sung Lee, Bingsheng.
Research Unit in Networking - University of Liège A Distributed Algorithm for Weighted Max-Min Fairness in MPLS Networks Fabian Skivée
Packet Scheduling: SCFQ, STFQ, WF2Q Yongho Seok Contents Review: GPS, PGPS SCFQ( Self-clocked fair queuing ) STFQ( Start time fair queuing ) WF2Q( Worst-case.
Shanjiang Tang, Bu-Sung Lee, Bingsheng He, Haikun Liu School of Computer Engineering Nanyang Technological University Long-Term Resource Fairness Towards.
A Case for Performance- Centric Network Allocation Gautam Kumar, Mosharaf Chowdhury, Sylvia Ratnasamy, Ion Stoica UC Berkeley.
Scheduling Determines which packet gets the resource. Enforces resource allocation to each flows. To be “Fair”, scheduling must: –Keep track of how many.
1 Fair Queuing Hamed Khanmirza Principles of Network University of Tehran.
XCP: eXplicit Control Protocol Dina Katabi MIT Lab for Computer Science
Towards Predictable Data Centers Why Johnny can’t use the cloud and what we can do about it? Hitesh Ballani, Paolo Costa, Thomas Karagiannis, Greg O’Shea.
Spring Computer Networks1 Congestion Control Sections 6.1 – 6.4 Outline Preliminaries Queuing Discipline Reacting to Congestion Avoiding Congestion.
Theophilus Benson*, Ashok Anand*, Aditya Akella*, Ming Zhang + *University of Wisconsin, Madison + Microsoft Research.
Univ. of TehranIntroduction to Computer Network1 An Introduction Computer Networks An Introduction to Computer Networks University of Tehran Dept. of EE.
Presented by Qifan Pu With many slides from Ali’s NSDI talk Ali Ghodsi, Matei Zaharia, Benjamin Hindman, Andy Konwinski, Scott Shenker, Ion Stoica.
Dominant Resource Fairness: Fair Allocation of Multiple Resource Types Ali Ghodsi, Matei Zaharia, Benjamin Hindman, Andy Konwinski, Scott Shenker, Ion.
6.888 Lecture 6: Network Performance Isolation Mohammad Alizadeh Spring
Shuihai Hu, Wei Bai, Kai Chen, Chen Tian (NJU), Ying Zhang (HP Labs), Haitao Wu (Microsoft) Sing Hong Kong University of Science and Technology.
1 Transport Bandwidth Allocation 3/29/2012. Admin. r Exam 1 m Max: 65 m Avg: 52 r Any questions on programming assignment 2 2.
CS244 Packet Scheduling (Generalized Processor Sharing) Some slides created by: Ion Stoica and Mohammad Alizadeh
VL2: A Scalable and Flexible Data Center Network
Data Center Network Architectures
Chris Cai, Shayan Saeed, Indranil Gupta, Roy Campbell, Franck Le
Improving Datacenter Performance and Robustness with Multipath TCP
Congestion Control and Resource Allocation
TCP, XCP and Fair Queueing
VL2: A Scalable and Flexible Data Center Network
Congestion Control and Resource Allocation
Towards Predictable Datacenter Networks
Presentation transcript:

Network Sharing Issues Lecture 15 Aditya Akella

Is this the biggest problem in cloud resource allocation? Why? Why not? How does the problem differ wrt allocating other resources? FairCloud: Sharing the Network in Cloud Computing, Sigcomm 2012 What are the assumptions? Drawbacks?

Motivation Network?

Context Networks are more difficult to share than other resources X

Context Several proposals that share network differently, e.g.: – proportional to # source VMs (Seawall [NSDI11]) – statically reserve bandwidth (Oktopus [Sigcomm12]) –…–… Provide specific types of sharing policies Characterize solution space and relate policies to each other?

FairCloud Paper 1.Framework for understanding network sharing in cloud computing – Goals, tradeoffs, properties 2.Solutions for sharing the network – Existing policies in this framework – New policies representing different points in the design space

Goals 1.Minimum Bandwidth Guarantees – Provides predictable performance – Example: file transfer finishes within time limit A1A1 A2A2 Time max = Size / B min B min

Goals 1.Minimum Bandwidth Guarantees 2.High Utilization – Do not leave useful resources unutilized – Requires both work-conservation and proper incentives A BBB Both tenants activeNon work-conservingWork-conserving

Goals 1.Minimum Bandwidth Guarantees 2.High Utilization 3.Network Proportionality – As with other services, network should be shared proportional to payment – Currently, tenants pay a flat rate per VM  network share should be proportional to #VMs (assuming identical VMs)

Goals 1.Minimum Bandwidth Guarantees 2.High Utilization 3.Network Proportionality – Example: A has 2 VMs, B has 3 VMs A1A1 A2A2 Bw A B1B1 B3B3 B2B2 Bw B Bw A = 2 3 When exact sharing is not possible use max-min

Goals 1.Minimum Bandwidth Guarantees 2.High Utilization 3.Network Proportionality Not all goals are achievable simultaneously!

Tradeoffs Min Guarantee High Utilization Network Proportionality Not all goals are achievable simultaneously!

Tradeoffs Min Guarantee Network Proportionality

Tradeoffs Bw B Bw A AB Access Link L Capacity C Bw B = 11/13 C Bw A = 2/13 C 10 VMs Network Proportionality Bw A ≈ C/N T  0 #VMs in the network Bw B = 1/2 C Bw A = 1/2 C Minimum Guarantee Min Guarantee Network Proportionality

Tradeoffs High Utilization Network Proportionality

Tradeoffs L B1B1 B3B3 B2B2 B4B4 A1A1 A3A3 A2A2 A4A4 High Utilization Network Proportionality

Tradeoffs L B1B1 B3B3 B2B2 B4B4 A1A1 A3A3 A2A2 A4A4 Bw B = 1/2 C Bw A = 1/2 C Network Proportionality High Utilization Network Proportionality

Tradeoffs L B1B1 B3B3 B2B2 B4B4 A1A1 A3A3 A2A2 A4A4 P High Utilization Network Proportionality Uncongested path

Tradeoffs L B1B1 B3B3 B2B2 B4B4 A1A1 A3A3 A2A2 A4A4 Bw B Bw A < Network Proportionality L L Tenants can be disincentivized to use free resources If A values A 1  A 2 or A 3  A 4 more than A 1  A 3 Bw A +Bw A = Bw B L L P P High Utilization Network Proportionality Uncongested path

Network Proportionality Tradeoffs L B1B1 B3B3 B2B2 B4B4 A1A1 A3A3 A2A2 A4A4 P Network proportionality applied only for flows traversing congested links shared by multiple tenants High Utilization Uncongested path Congestion Proportionality

Network Proportionality Tradeoffs L B1B1 B3B3 B2B2 B4B4 A1A1 A3A3 A2A2 A4A4 Uncongested path P Bw B Bw A = Congestion Proportionality L L High Utilization Congestion Proportionality

Network Proportionality Tradeoffs Still conflicts with high utilization High Utilization Congestion Proportionality

Tradeoffs B1B1 B3B3 A1A1 A3A3 B2B2 B4B4 A2A2 A4A4 L2L2 C 1 = C 2 = C Network Proportionality High Utilization L1L1 Congestion Proportionality

Tradeoffs B1B1 B3B3 A1A1 A3A3 B2B2 B4B4 A2A2 A4A4 L1L1 L2L2 Bw B Bw A = Congestion Proportionality L1 Bw B Bw A = L2 Network Proportionality High Utilization C 1 = C 2 = C Congestion Proportionality

Tradeoffs B1B1 B3B3 A1A1 A3A3 B2B2 B4B4 A2A2 A4A4 L1L1 L2L2 Demand drops to ε Network Proportionality High Utilization C 1 = C 2 = C Congestion Proportionality

Tradeoffs B1B1 B3B3 A1A1 A3A3 B2B2 B4B4 A2A2 A4A4 ε C - ε ε Tenants incentivized to not fully utilize resources Network Proportionality High Utilization C 1 = C 2 = C Congestion Proportionality L1L1 L2L2

Tradeoffs B1B1 B3B3 A1A1 A3A3 B2B2 B4B4 A2A2 A4A4 ε C - 2 ε Uncongested ε C - ε L1L1 L2L2 C 1 = C 2 = C Network Proportionality High Utilization Tenants incentivized to not fully utilize resources Congestion Proportionality

L2L2 Tradeoffs B1B1 B3B3 A1A1 A3A3 B2B2 B4B4 A2A2 A4A4 ε C - 2 ε C/2 Uncongested C 1 = C 2 = C Network Proportionality High Utilization Congestion Proportionality Tenants incentivized to not fully utilize resources L1L1

Tradeoffs Proportionality applied to each link independently Congestion Proportionality Link Proportionality Network Proportionality High Utilization L2L2 B1B1 B3B3 A1A1 A3A3 B2B2 B4B4 A2A2 A4A4 L1L1

L2L2 Tradeoffs B1B1 B3B3 A1A1 A3A3 B2B2 B4B4 A2A2 A4A4 Full incentives for high utilization Congestion Proportionality Link Proportionality Network Proportionality High Utilization L1L1

Goals and Tradeoffs Min Guarantee Link Proportionality High Utilization Congestion Proportionality Network Proportionality

Guiding Properties Min Guarantee Link Proportionality High Utilization Congestion Proportionality Network Proportionality Break down goals into lower-level necessary properties

Properties Min Guarantee Link Proportionality High Utilization Congestion Proportionality Network Proportionality

Work Conservation Bottleneck links are fully utilized Static reservations do not have this property

Properties Min Guarantee Link Proportionality High Utilization Congestion Proportionality Network Proportionality Work Conservation

Utilization Incentives Tenants are not incentivized to lie about demand to leave links underutilized Network and congestion proportionality do not have this property Allocating links independently provides this property

Properties Min Guarantee Link Proportionality High Utilization Congestion Proportionality Network Proportionality Work Conservation Utilization Incentives

Communication-pattern Independence Allocation does not depend on communication pattern Per flow allocation does not have this property – (per flow = give equal shares to each flow)  Same Bw 

Properties Min Guarantee Link Proportionality High Utilization Congestion Proportionality Network Proportionality Work Conservation Utilization Incentives Comm-Pattern Independence

Symmetry Swapping demand directions preserves allocation Per source allocation lacks this property – (per source = give equal shares to each source)  Same Bw 

Goals, Tradeoffs, Properties Min Guarantee Link Proportionality High Utilization Congestion Proportionality Network Proportionality Work Conservation Utilization Incentives Comm-Pattern Independence Symmetry

Outline 1.Framework for understanding network sharing in cloud computing – Goals, tradeoffs, properties 2.Solutions for sharing the network – Existing policies in this framework – New policies representing different points in the design space

Per Flow (e.g. today) Min Guarantee Link Proportionality High Utilization Congestion Proportionality Network Proportionality Work Conservation Utilization Incentives Comm-Pattern Independence Symmetry

Per Source (e.g., Seawall [NSDI’11]) Min Guarantee Link Proportionality High Utilization Congestion Proportionality Network Proportionality Work Conservation Utilization Incentives Comm-Pattern Independence Symmetry

Static Reservation (e.g., Oktopus [Sigcomm’11]) Min Guarantee Link Proportionality High Utilization Congestion Proportionality Network Proportionality Work Conservation Utilization Incentives Comm-Pattern Independence Symmetry

New Allocation Policies 3 new allocation policies that take different stands on tradeoffs

Proportional Sharing at Link-level (PS-L) Min Guarantee Link Proportionality High Utilization Congestion Proportionality Network Proportionality Work Conservation Utilization Incentives Comm-Pattern Independence Symmetry

Proportional Sharing at Link-level (PS-L) Per tenant WFQ where weight = # tenant’s VMs on link A B WQ A = #VMs A on L Bw B Bw A = #VMs A on L #VMs B on L Can easily be extended to use heterogeneous VMs (by using VM weights)

Proportional Sharing at Network-level (PS-N) Min Guarantee Link Proportionality High Utilization Congestion Proportionality Network Proportionality Work Conservation Utilization Incentives Comm-Pattern Independence Symmetry

Proportional Sharing at Network-level (PS-N) Congestion proportionality in severely restricted context Per source-destination WFQ, total tenant weight = # VMs

Proportional Sharing at Network-level (PS-N) WQ A1  A2 = 1/N A1 + 1/N A2 A1A1 A2A2 N A2 N A1 Total WQ A = #VMs A Congestion proportionality in severely restricted context Per source-destination WFQ, total tenant weight = # VMs

Proportional Sharing at Network-level (PS-N) WQ B WQ A = #VMs A #VMs B Congestion proportionality in severely restricted context Per source-destination WFQ, total tenant weight = # VMs

Proportional Sharing on Proximate Links (PS-P) Min Guarantee Link Proportionality High Utilization Congestion Proportionality Network Proportionality Work Conservation Utilization Incentives Comm-Pattern Independence Symmetry

Assumes a tree-based topology: traditional, fat-tree, VL2 (currently working on removing this assumption) Proportional Sharing on Proximate Links (PS-P)

Assumes a tree-based topology: traditional, fat-tree, VL2 (currently working on removing this assumption) Min guarantees – Hose model – Admission control Proportional Sharing on Proximate Links (PS-P) A1A1 Bw A1 A2A2 Bw A2 AnAn Bw An …

Assumes a tree-based topology: traditional, fat-tree, VL2 (currently working on removing this assumption) Min guarantees – Hose model – Admission control High Utilization – Per source fair sharing towards tree root Proportional Sharing on Proximate Links (PS-P)

Assumes a tree-based topology: traditional, fat-tree, VL2 (currently working on removing this assumption) Min guarantees – Hose model – Admission control High Utilization – Per source fair sharing towards tree root – Per destination fair sharing from tree root Proportional Sharing on Proximate Links (PS-P)

Deploying PS-L, PS-N and PS-P Full Switch Support – All allocations can use hardware queues (per tenant, per VM or per source-destination) Partial Switch Support – PS-N and PS-P can be deployed using CSFQ [Sigcomm’98] No Switch Support – PS-N can be deployed using only hypervisors – PS-P could be deployed using only hypervisors, we are currently working on it

Evaluation Small Testbed + Click Modular Router – 15 servers, 1Gbps links Simulation + Real Traces – 3200 nodes, flow level simulator, Facebook MapReduce traces

Many to one Bw B Bw A AB N One link, testbed PS-P offers guarantees Bw A N

MapReduce One link, testbed Bw B Bw A 5 R5 M M+R = 10 M Bw B (Mbps) PS-L offers link proportionality

MapReduce Network, simulation, Facebook trace

MapReduce Network, simulation, Facebook trace PS-N is close to network proportionality

MapReduce Network, simulation, Facebook trace PS-N and PS-P reduce shuffle time of small jobs by 10-15X

Summary Sharing cloud networks is not trivial First step towards a framework to analyze network sharing in cloud computing – Key goals (min guarantees, high utilization and proportionality), tradeoffs and properties New allocation policies, superset properties from past work – PS-L: link proportionality + high utilization – PS-N: restricted network proportional – PS-P: min guarantees + high utilization What are the assumptions, drawbacks?