Testing LOSSLESS Network for RDMA

Slides:



Advertisements
Similar presentations
© 2007 Cisco Systems, Inc. All rights reserved. Valašské Meziříčí Connecting to the Network.
Advertisements

RoCEv2 Update from the IBTA
Computer Networking Macedonia VLAN’s, VTP, InterVLAN Routing, (And if there is enough time - STP)
Towards Virtual Networks for Virtual Machine Grid Computing Ananth I. Sundararaj Peter A. Dinda Prescience Lab Department of Computer Science Northwestern.
Transmitting and Tracking Packets of Data Through The TCP and UDP Network Protocols Todd Deshane Ashwin Venkatraman McNair Program Clarkson University.
5/12/05CS118/Spring051 A Day in the Life of an HTTP Query 1.HTTP Brower application Socket interface 3.TCP 4.IP 5.Ethernet 2.DNS query 6.IP router 7.Running.
Design and Implementation of SIP-aware DDoS Attack Detection System.
\\fs\share File Server SMB Direct Client Application NIC RDMA NIC TCP/ IP SMB Direct Ethernet and/or InfiniBand TCP/ IP Unchanged.
(ITI310) By Eng. BASSEM ALSAID SESSIONS 8: Network Load Balancing (NLB)
Networking in a Linux Environment Pete Eby Dan Thomas Robert Zurawski.
Practical TDMA for Datacenter Ethernet
© 2013 Mellanox Technologies 1 NoSQL DB Benchmarking with high performance Networking solutions WBDB, Xian, July 2013.
Lab How to Use WANem Last Update Copyright 2011 Kenneth M. Chipps Ph.D. 1.
1 Enabling Large Scale Network Simulation with 100 Million Nodes using Grid Infrastructure Hiroyuki Ohsaki Graduate School of Information Sci. & Tech.
Remote Access Chapter 4. Learning Objectives Understand implications of IEEE 802.1x and how it is used Understand VPN technology and its uses for securing.
1 March 2010 A Study of Hardware Assisted IP over InfiniBand and its Impact on Enterprise Data Center Performance Ryan E. Grant 1, Pavan Balaji 2, Ahmad.
Towards a Common Communication Infrastructure for Clusters and Grids Darius Buntinas Argonne National Laboratory.
Extreme-scale computing systems – High performance computing systems Current No. 1 supercomputer Tianhe-2 at petaflops Pushing toward exa-scale computing.
InfiniSwitch Company Confidential. 2 InfiniSwitch Agenda InfiniBand Overview Company Overview Product Strategy Q&A.
Dynamic Host Configuration Protocol Engr. Mehran Mamonai.
Chapter 8: Virtual LAN (VLAN)
Team Member: Xinjie Guan, University of Missouri Kansas City; Xili Wan, University of Missouri Kansas City; Ruiling Gao, Tufts University; Date: June 2,
Data Center Bridging
By: Aleksandr Movsesyan Advisor: Hugh Smith. OSI Model.
LAN Switching and Wireless – Chapter 1 Vilina Hutter, Instructor
Infiniband Bart Taylor. What it is InfiniBand™ Architecture defines a new interconnect technology for servers that changes the way data centers will be.
Congestion Notification Mechanisms in 802 networks Manoj Wadekar IEEE Interim Meeting January 12, 2005.
Ethernet. Ethernet standards milestones 1973: Ethernet Invented 1983: 10Mbps Ethernet 1985: 10Mbps Repeater 1990: 10BASE-T 1995: 100Mbps Ethernet 1998:
Mellanox Connectivity Solutions for Scalable HPC Highest Performing, Most Efficient End-to-End Connectivity for Servers and Storage April 2010.
Efficiency of small size tasks calculation in grid clusters using parallel processing.. Olgerts Belmanis Jānis Kūliņš RTU ETF Riga Technical University.
Mellanox Connectivity Solutions for Scalable HPC Highest Performing, Most Efficient End-to-End Connectivity for Servers and Storage September 2010 Brandon.
High Performance Networks and Long-Distance Data Transfers Tom Ammon Network Engineer University Of Utah Center for High Performance Computing.
PATH DIVERSITY WITH FORWARD ERROR CORRECTION SYSTEM FOR PACKET SWITCHED NETWORKS Thinh Nguyen and Avideh Zakhor IEEE INFOCOM 2003.
Ethernet Packet Filtering – Part 2 Øyvind Holmeide 10/28/2014 by.
Disruptive Storage Workshop Lustre Hardware Primer
Enhancements for Voltaire’s InfiniBand simulator
Advanced Network Tap application for
Balazs Voneki CERN/EP/LHCb Online group
How I Learned to Stop Worrying About the Core and Love the Edge
Data Center Network Architectures
PTPv1 and PTPv2 translation IN FTI Systems
Kiyoshi Kodama, SE Japan 07-Oct-2008
Network Load Balancing Topology
Top-Down Network Design Chapter Nine Developing Network Management Strategies Copyright 2010 Cisco Press & Priscilla Oppenheimer.
Introduction to Networks
Virtual Local Area Network
2018 Huawei H Real Questions Killtest
Cisco Real Exam Dumps IT-Dumps
Advanced Network Training
Roaming Interval Measurements
Marrying OpenStack and Bare-Metal Cloud
PPPoE Internet Point to Point Protocol over Ethernet
Authors Mach Chen (Huawei) Xuesong Geng (Huawei) Zhenqiang Li (CMCC)
NTHU CS5421 Cloud Computing
Providing QoS through Active Domain Management
Big-Data around the world
Development & Evaluation of Network Test-beds
Indiana University, Bloomington
Specialized Cloud Architectures
RDMA over Commodity Ethernet at Scale
Peer-to-Peer Client-server.
Project-2 (20%) – DiffServ and TCP Congestion Control
Performance And Scalability In Oracle9i And SQL Server 2000
© 2016 Global Market Insights, Inc. USA. All Rights Reserved Ethernet Storage Market Size Growth During Forecast Period.
In-network computation
LOSSLESS Network for RDMA in ODCC
Towards Hyperscale HPC & RDMA
DetNet Bounded Latency-04
Top-Down Network Design Chapter Nine Developing Network Management Strategies Copyright 2010 Cisco Press & Priscilla Oppenheimer.
Presentation transcript:

Testing LOSSLESS Network for RDMA Liang Guo, ODCC(Chair of test WG) Feng Gao, Baidu Liyang Sun, Huawei Jun Liu, Cisco Qingchun Song, Mellanox IEEE 802 Nendica Vienna, Austria July 2019

Achievements The overall achievements of the Lossless network project Typical Scenario Testing Method for Lossless Network in Data Center General Technical Requirements for Lossless Network in Data Center Technology and Application White Paper Three ODCC specifications IEEE Nendica Lossless Network white paper

Participants of Lossless Network Tests

Lossless Network tests for HPC Test steps: 1、Install OpenFoam (version V1606+) and OFED (version 3.4-1.0.0.0) on each server (Huawei 2288 v3, operating system redhat7.1), and configure all nodes to log in without password. 2、Select one server as master (the leftmost one in the figure above) and the other servers as slave. IB network, lossless network and traditional Ethernet network are respectively configured. The IB and lossless networking configurations for 16 sets are as follows. 3、Pre-execution processing; Execute. / run 1. sh on Master to complete preprocessing. 4、Perform operations; Execute. / run 2. sh on Master for parallel computing 5、Record execution time Expected results: There was no packet loss and the task completion time was comparable. Test1:Lossless Network VS. IB Test2:Lossless Network VS. Traditional Eth

Test Results 0 packet loss, the lossless network helps HPC JCT reduced by 31% than traditional Eth, equal to IB

Lossless Network tests for NVMe over Fabric Configurations: 1、Test the IOPS and tail latency of NOF services. The network topology, port rate, and node configuration are the same. 2、Configure target and client servers. For example, Huawei RH2288v3 server can be used. 3、Using standard tool FIO to test, client accesses NVMe SSD under target and carries out reading and writing tests respectively. 4、Statistical FIO test results. 5、Recording IOPS and tail delay performance. Expected results: Lossless network IOPS greatly improved, no packet loss, low tail delay Test 1:Lossless Network vs IB Test 2:Lossless Network vs traditional Eth

Test Results Traditional Eth Lossless Network Infiniband write@128K, 3v1 US IB Traditional Eth Lossless network Traditional Eth Lossless Network Infiniband Lossless network: zero packet loss and low latency, and IOPS improved by 42% compared with traditional Ethernet. The tail delay is better than that of the IB, and the IB can be replaced.

To be continued: More considerations Network scalability for more latency sensitive RDMA applications deployed in data center networks, such as HPC/AI computing. More effective congestion control mechanisms for RDMA switching when step into the era of 100G servers with 400G switches or higher. Easy to deploy new congestion control algorithms on the server. Effects of the mixture of RDMA traffics and normal TCP traffics. More study should be done for better RDMA switching over IP networks!

THANKS!