2018/5/13 CoSwitch: A Cooperative Switching Design for Software Defined Data Center Networking Author: Yue ZhangKai ,Zheng, Chengchen Hu, Kai Chen, Yi.

Slides:



Advertisements
Similar presentations
1 Exploring Efficient and Scalable Multicast Routing in Future Data Center Networks Dan Li, Jiangwei Yu, Junbiao Yu, Jianping Wu Tsinghua University Presented.
Advertisements

SDN and Openflow.
A Scalable, Commodity Data Center Network Architecture.
OpenFlow Switch Limitations. Background: Current Applications Traffic Engineering application (performance) – Fine grained rules and short time scales.
A User Experience-based Cloud Service Redeployment Mechanism KANG Yu.
1 Route Table Partitioning and Load Balancing for Parallel Searching with TCAMs Department of Computer Science and Information Engineering National Cheng.
OpenFlow-Based Server Load Balancing GoneWild Author : Richard Wang, Dana Butnariu, Jennifer Rexford Publisher : Hot-ICE'11 Proceedings of the 11th USENIX.
StriD 2 FA: Scalable Regular Expression Matching for Deep Packet Inspection Author: Xiaofei Wang, Junchen Jiang, Yi Tang, Bin Liu, and Xiaojun Wang Publisher:
SDN AND OPENFLOW SPECIFICATION SPEAKER: HSUAN-LING WENG DATE: 2014/11/18.
SDN Management Layer DESIGN REQUIREMENTS AND FUTURE DIRECTION NO OF SLIDES : 26 1.
CellSDN: Software-Defined Cellular Core networks Xin Jin Princeton University Joint work with Li Erran Li, Laurent Vanbever, and Jennifer Rexford.
Network Virtualization in Multi-tenant Datacenters Author: VMware, UC Berkeley and ICSI Publisher: 11th USENIX Symposium on Networked Systems Design and.
Research on TCAM-based OpenFlow Switch Author: Fei Long, Zhigang Sun, Ziwen Zhang, Hui Chen, Longgen Liao Conference: 2012 International Conference on.
Selective Packet Inspection to Detect DoS Flooding Using Software Defined Networking Author : Tommy Chin Jr., Xenia Mountrouidou, Xiangyang Li and Kaiqi.
OpenFlow MPLS and the Open Source Label Switched Router Department of Computer Science and Information Engineering, National Cheng Kung University, Tainan,
Shadow MACs: Scalable Label- switching for Commodity Ethernet Author: Kanak Agarwal, John Carter, Eric Rozner and Colin Dixon Publisher: HotSDN 2014 Presenter:
Updating Designed for Fast IP Lookup Author : Natasa Maksic, Zoran Chicha and Aleksandra Smiljani´c Conference: IEEE High Performance Switching and Routing.
Packet Classification Using Dynamically Generated Decision Trees
1 IP Routing table compaction and sampling schemes to enhance TCAM cache performance Author: Ruirui Guo, Jose G. Delgado-Frias Publisher: Journal of Systems.
Scalable Multi-match Packet Classification Using TCAM and SRAM Author: Yu-Chieh Cheng, Pi-Chung Wang Publisher: IEEE Transactions on Computers (2015) Presenter:
Network Virtualization Ben Pfaff Nicira Networks, Inc.
Data Center Architectures
2018/4/23 Dynamic Load-balanced Path Optimization in SDN-based Data Center Networks Author: Yuan-Liang Lan , Kuochen Wang and Yi-Huai Hsu Presenter: Yi-Hsien.
Minimizing latency of critical traffic through SDN
Software defined networking: Experimental research on QoS
2018/5/8 An approach for detecting encrypted insider attacks on OpenFlow SDN Networks Author: Charles V. Neu , Avelino F. Zorzox , Alex M. S. Orozcoy and.
The DPIaaS Controller Prototype
Heitor Moraes, Marcos Vieira, Italo Cunha, Dorgival Guedes
A Survey of Data Center Network Architectures By Obasuyi Edokpolor
A DFA with Extended Character-Set for Fast Deep Packet Inspection
Distributed Processors
2018/6/5 Centralized Retransmission Management with SDN in Multihop Wireless Access Network Author: Bong-Hwan Oh , Jaiyoung Lee Presenter: Yi-Hsien Wu.
2018/6/26 An Energy-efficient TCAM-based Packet Classification with Decision-tree Mapping Author: Zhao Ruan, Xianfeng Li , Wenjun Li Publisher: 2013.
2018/7/4 QoS-aware Multipathing in datacenters using effective bandwidth estimation and SDN Author: Runxin Wang, Simone Mangiante, Alan Davy, Lei Shi,
Migration Strategies – Business Desktop Deployment (BDD) Overview
Indigo Doyoung Lee Dept. of CSE, POSTECH
Dingming Wu+, Yiting Xia+*, Xiaoye Steven Sun+,
Software Defined Networking (SDN)
Statistical Optimal Hash-based Longest Prefix Match
2018/11/19 Source Routing with Protocol-oblivious Forwarding to Enable Efficient e-Health Data Transfer Author: Shengru Li, Daoyun Hu, Wenjian Fang and.
SoftRing: Taming the Reactive Model for Software Defined Networks
SPEAKER: Yu-Shan Chou ADVISOR: DR. Kai-Wei Ke
Dynamic Packet-filtering in High-speed Networks Using NetFPGAs
2018/12/10 Energy Efficient SDN Commodity Switch based Practical Flow Forwarding Method Author: Amer AlGhadhban and Basem Shihada Publisher: 2016 IEEE/IFIP.
Virtual TCAM for Data Center Switches
Cloud Computing Architecture
A Small and Fast IP Forwarding Table Using Hashing
Scalable Multi-Match Packet Classification Using TCAM and SRAM
Data Center Architectures
EMOMA- Exact Match in One Memory Access
2019/5/2 Using Path Label Routing in Wide Area Software-Defined Networks with OpenFlow ICNP = International Conference on Network Protocols Presenter:Hung-Yen.
2019/5/3 A De-compositional Approach to Regular Expression Matching for Network Security Applications Author: Eric Norige Alex Liu Presenter: Yi-Hsien.
2019/5/5 A Flexible Wildcard-Pattern Matching Accelerator via Simultaneous Discrete Finite Automata Author: Hsiang-Jen Tsai, Chien-Chih Chen, Yin-Chi Peng,
2019/5/13 A Weighted ECMP Load Balancing Scheme for Data Centers Using P4 Switches Presenter:Hung-Yen Wang Authors:Peng Wang, George Trimponias, Hong Xu,
SDN-Guard: DoS Attacks Mitigation in SDN Networks
Virtualization Dr. S. R. Ahmed.
Fast Testing Network Data Plane with RuleChecker
QoS Constrained Path Optimization Algorithm in NFV/SDN Environment
Fast Network Congestion Detection And Avoidance Using P4
OpenSec:Policy-Based Security Using Software-Defined Networking
Design principles for packet parsers
2019/7/26 OpenFlow-Enabled User Traffic Profiling in Campus Software Defined Networks Presenter: Wei-Li,Wang Date: 2016/1/4 Author: Taimur Bakhshi and.
An Improved Wu-Manber Multiple Patterns Matching Algorithm
A SRAM-based Architecture for Trie-based IP Lookup Using FPGA
2019/10/9 A Weighted ECMP Load Balancing Scheme for Data Centers Using P4 Switches Presenter:Hung-Yen Wang Authors:Jin-Li Ye, Yu-Huang Chu, Chien Chen.
Authors: Ding-Yuan Lee, Ching-Che Wang, An-Yeu Wu Publisher: 2019 VLSI
Elmo Muhammad Shahbaz Lalith Suresh, Jennifer Rexford, Nick Feamster,
2019/10/19 Efficient Software Packet Processing on Heterogeneous and Asymmetric Hardware Architectures Author: Eva Papadogiannaki, Lazaros Koromilas, Giorgos.
MEET-IP Memory and Energy Efficient TCAM-based IP Lookup
Efficient Migration of Large-memory VMs Using Private Virtual Memory
Presentation transcript:

2018/5/13 CoSwitch: A Cooperative Switching Design for Software Defined Data Center Networking Author: Yue ZhangKai ,Zheng, Chengchen Hu, Kai Chen, Yi Wang Presenter: Yi-Hsien Wu Date: 2016/10/19 Department of Computer Science and Information Engineering National Cheng Kung University, Taiwan R.O.C. CSIE CIAL Lab 1

Outline Introduction Problem Statement Coswitch Architecture Evaluation Related Work Conclusion National Cheng Kung University CSIE Computer & Internet Architecture Lab

Introduction Data Center NetWork (DCN) : 2018/5/13 Introduction Data Center NetWork (DCN) : Data center is a pool of resources(computational, storage, network) interconnected using a communication network. Data Center Network (DCN) holds a pivotal role in a data center, as it interconnects all of the data center resources together. DCNs need to be scalable and efficient to connect tens or even hundreds of thousands of servers to handle the growing demands of Cloud computing. Today’s data centers are constrained by the interconnection network. 近期因為ruleset size快速成長,因此ruleset複雜度造成一般封包分類方法memory表現很差 Swintop是一種將ruleset去分類的方式 National Cheng Kung University CSIE Computer & Internet Architecture Lab CSIE CIAL Lab

Type of DCN Three-tier DCN : 2018/5/13 Type of DCN Three-tier DCN : three-tier DCN architecture follows a multi-rooted tree based network topology composed of three layers of network switches, namely access, aggregate, and core layers. The servers in the lowest layers are connected directly to Access layer switches. The aggregate layer switches interconnects multiple access layer switches together. All of the aggregate layer switches are connected to each other by core layer switches. Core layer switches are also responsible for connecting the data center to the Internet. 近期因為ruleset size快速成長,因此ruleset複雜度造成一般封包分類方法memory表現很差 Swintop是一種將ruleset去分類的方式 National Cheng Kung University CSIE Computer & Internet Architecture Lab CSIE CIAL Lab

Type of DCN Fat tree DCN : 2018/5/13 Type of DCN Fat tree DCN : Employing commodity network switches based architecture using Clos topology. The network elements in fat tree topology also follows hierarchical organization of network switches in edge, aggregate, and core layers. 近期因為ruleset size快速成長,因此ruleset複雜度造成一般封包分類方法memory表現很差 Swintop是一種將ruleset去分類的方式 National Cheng Kung University CSIE Computer & Internet Architecture Lab CSIE CIAL Lab

Clos Networks / Fat-Trees Also called “K-ary FAT tree” Each pod consists of (k/2)2 servers & 2 layers of k/2 k-port switches. Each edge switch connects to k/2 sever &k/2 aggr. switches. Each aggr. switch connects to k/2 edge &k/2 core (k/2)2 core switches : each connects to k pods. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Problem Statement Data Center Workload The Cumulative Distribution Function (CDF) of flow number and packet number (in Figure1) is demonstrated together, with the increasing of flow size . The rate of the flow number curve is obviously bigger than the rate of packet number curve, when the flow size (per flow packet) is smaller than 18. And the opposite case turns out, when the flow size is bigger than 36. In other words, there are more mice flows containing fewer packets, and fewer elephant flows containing more packets. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Problem Statement Data Center Workload National Cheng Kung University CSIE Computer & Internet Architecture Lab

Problem Statement Data Center Workload Figure 2 depicts the comparisons between mice flows and elephant flows with regard to their population and total traffic volume (indicated by total packet number). It can be seen that the mice flows take over 80% of flows population, and contain only less than 15% number of the total traffic volume. While the small numbers (10%) of elephant flows contain more than 75% of the total traffic volume. The polarized traffic patterns will bring unbalanced workload to the DCN. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Problem Statement Data Center Workload The processing of mice flows requires frequent flow table operations on the control plane, such as flow add/delete/lookup actions. The traffic volume in the mice flows is relatively small, so the cost for packet processing in data plane is actually not very significant. Thus we call the mice flows as “control plane intensive workload”. On the other hand, in the case of elephant flows, since the flow population is relatively much smaller, the flow operations on control plane are therefore much fewer. However, there are huge amount of packet data in the elephant flows, leading to high data plane operational overhead. So we call the elephant flows as “data plane intensive workload”. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Problem Statement Performance Analysis of Network Device Two types of OpenFlow enabled switches , One is dedicated physical switch, the other is software-based virtual switch. Figure 3 depicts the hierarchical networking infrastructure in data center. The virtual switches deployed on server connect to VMs directly. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Problem Statement Performance Analysis of Network Device When the packet comes into the switch ports, it will be processed and redirected to the destination port based on the flow table rules. Otherwise, if there is no entry match in the flow table, the packet will be directed to the controller through control plane. After getting the responding flow rule from the controller, the switch stores it into the flow table. From this point of view, OpenFlow actually introduces two kinds of potential performance bottlenecks to the switch. One is the embedded control CPU, because it is usually too weak to afford the frequent flow table update under heavy control plane workload. The other is the limitation of flow table size. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Problem Statement OpenFlow Networking bottleneck Data Plane Performance: Figure 4 shows that the physical switch can achieve line rate (1Gbps), and the embedded control plane CPU on the switch stays idle. By zooming in the input load, we find that the server CPU is proportionate growth with the traffic amount, and finally reaches the processing bottleneck under the extremely heavy load, about 80k pps. This constrains the general-purpose CPU resource of the server. The extra CPU cost will always waste valuable computing resource for the tenants. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Problem Statement OpenFlow Networking bottleneck National Cheng Kung University CSIE Computer & Internet Architecture Lab

Problem Statement OpenFlow Networking bottleneck Control Plane Performance: For each new incoming flow, the switch has to get a flow rule from the controller first. The rule is also need to be removed when the flow entry is timeout. Switch resource cost under the control plane intensive workload is compared in Figure 5. Since there is no pre-installed rule in this test, the switch needs to spend more CPU time on the control plane to update flow table for each new coming flow, before handling it on data plane. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Problem Statement OpenFlow Networking bottleneck It can be seen that the control plane workload dose not cause performance bottleneck on virtual switch’s CPU (Xeon E5420) and memory (RAM), with the input load growing over 10k pps. However, for the physical switch, the CPU (MPC8541EE500) is too weak to afford frequent flow table updating. The embedded CPU reaches saturation when the input load is over 7.7k pps. The on-chip memory (TCAM) is also so limited (e.g. 2k) that will be used up when the input load is over 6kpps, even under very short flow entry timeout (3s). National Cheng Kung University CSIE Computer & Internet Architecture Lab

CoSwitch Architecture By observed , the physical switch is good at handling large amount of traffic with small flow numbers, i.e. the elephant flows. On the other hand, the virtual switch with more control plane capacity is better to handle the mice flows. The cooperative switching scheme, CoSwitch, is proposed to strip the control plane intensive mice flow processing from the physical switches, and leverage the complementary virtual switches on the server side to handle these large-number-small volume mice flows, in a distributed and adaptive manner. In this way, both the physical switches and the virtual switches exert their strengths, respectively, and complement each others’ weakness in performance, leading to an adaptive balance of the system. National Cheng Kung University CSIE Computer & Internet Architecture Lab

CoSwitch Architecture National Cheng Kung University CSIE Computer & Internet Architecture Lab

CoSwitch Architecture 1. Physical switch with OpenFlow supporting (i.e. Pronto 3290 ) 2. Virtual switch deployed in the host server (i.e. OVS. ) 3. A close loop workload distribution mechanism is introduced in CoSwitch, supported by two modules. (1)Flow-Arbiter deployed on the server side, and (2)Flow-Prediction deployed as a OpenFlow controller module. It is worth that both of the modules are software implementation, which means no particular hardware modification is needed. National Cheng Kung University CSIE Computer & Internet Architecture Lab

CoSwitch Architecture Flow-Arbiter is deployed between VM network and hypervisor virtual switch. The traffic of VMs is filtered by Flow-Arbiter, where the elephant flows are picked out of mice flows. A feasible way is to pick out the elephant flows with the flow size bigger than a threshold. With the division of mice flows and elephant flows, the workload can be distributed to virtual switch and physical switch, respectively. National Cheng Kung University CSIE Computer & Internet Architecture Lab

CoSwitch Architecture In most cases the processing power of virtual switch and physical switch do not stay steady, either because of the inherent physical difference among networking devices, or the dynamic traffic workload. So the flow size threshold should be dynamically computed at run time. Flow-Prediction deployed on the OpenFlow controller, compute out the flow prediction threshold. The threshold is not obtained from the flow features. Instead, it is based on the resource utilization periodically feedback from the virtual and physical switch. If either of the switches is heavy-loaded and the other is not, the Flow-Prediction module will make proper change of the threshold to shift some workload to the light loaded switch. National Cheng Kung University CSIE Computer & Internet Architecture Lab

CoSwitch Architecture More specific , Flow-Arbiter divides the traffic into mice and elephant flows. The large numbers of mice flows are processed locally, by the OVS within each server. The elephant flows will only be marked and forwarded to the physical switch. The marking method is flexible, either a special VLAN or TOS tag can be used. At physical switch side, the mark of each packet is checked. For the marked elephant flows, the processing is carried by the switch hardware. In this way, the virtual switch and physical switch are cooperated to leverage their complementary capabilities to handle the data center workload. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Evaluation Testbed Setup : 1. Commodity OpenFlow enabled switches (Pronto 3290) are used for ToR, as the physical switches with OpenFlow support. 2. The switches are interconnected with each other via 10GE link, and with 16 IBM system x3550 servers via 1GE link. 3. On each server, (1)OpenVSwitch is deployed to replace the switch of KVM hypervisor, as the virtual switch supporting OpenFlow. (2) Flow-Arbiter is deployed as a net filter module to intercept the VM traffic before they come into OVS. 4. The OpenFlow controller (Floodlight) with Flow-Prediction module is deployed in a standalone server, with connectivity of all virtual and physical switches. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Evaluation – Data Plane Performance Figure 7 shows the particular CPU core utilization of virtual switch. Through CPU time sampling, it can be seen that most of the CPU time is cost by the packet processing of elephant flows, which is significantly larger than the processing of mice flows. This is because the packet processing on virtual switch takes up the general CPU, the elephant flows with more total packets and bytes will cost more CPU resource. Another part of overhead is the unavoidable network I/O for packet forwarding. Thus , CoSwitch put the handling of elephant flows to the physical switch, where the packet processing is performed by specialized hardware with little data plane overhead. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Evaluation – Data Plane Performance National Cheng Kung University CSIE Computer & Internet Architecture Lab

Evaluation – Data Plane Performance In Figure 8, the CPU overhead on the server is compared under the approaches of CoSwitch and handling all flows on virtual switch. For CoSwitch, since most of elephant flows packets are not processed on the server side, the corresponding CPU core cost is much lower. Only required by mice flow packet processing and network I/O. It can be seen that when the input load is over 100MB/s, the virtual switch will meet the CPU bottleneck for processing all the packets. However the core usage in CoSwitch can be reduced to 37%, by merely handling the mice flows. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Evaluation – Control Plane Performance Two control plane performance issues of OpenFlow are all on the physical switch. One is the storaging of flow rules, which comes from the limitation of on-chip flow table size. The run time flow entry requirement is depicts in Figure 9. With the growing of flow timeout, the number of flow entries required is increasing fast to overflow the hardware capability (e.g., 2k on Pronto 3290). By contrast in CoSwitch, only not more than 200 elephant flow rules need to be cached on physical switch, for which is more than sufficient to handle. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Evaluation – Control Plane Performance National Cheng Kung University CSIE Computer & Internet Architecture Lab

Evaluation – Control Plane Performance The other serious bottleneck on control plane is the capability of embedded switch CPU. In Figure 10, the switch CPU utilization is compared between CoSwitch and physical switch approaches. It can be seen that the physical switch CPU will soon get overloaded when the input exceeds 7.7k pps, because huge amount of mice flows require lots of flow table updating. In CoSwitch, since the flow updating only triggered by the small numbers of elephant flows, the increasing rate of switch CPU cost is much lower, means that the design is more scalable and cost-efficient. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Related Work Data center networking has been well studied. Some recent researches provide some level of agility based on fat-tree like topology. PortLand presents a big layer-2 network that takes advantages of layer-2 features. However, it needs special switches to support new forwarding behaviors, limiting its compatibility and deployment practicability. National Cheng Kung University CSIE Computer & Internet Architecture Lab

Conclusion The enablement of OpenFlow on DCN encounters several bottlenecks . The most serious one is the handling of mice flows on physical switches, which has limited resource on the control plane. Through a cooperative switching design of virtual switch and physical switch, CoSwitch is presented to mitigate the bottlenecks of OpenFlow deployment, without specialized hardware requirement. The powerful control plane of virtual switch provides scalable processing capability of the mice flows. And the high speed data plane of physical switch preserves the processing performance of elephant flows. Reallife testbed experiments prove that CoSwitch is a scalable SDN design to provide agile traffic control for DCN. National Cheng Kung University CSIE Computer & Internet Architecture Lab