1 Network Performance Optimisation and Load Balancing Wulf Thannhaeuser.

Slides:



Advertisements
Similar presentations
DataTAG CERN Oct 2002 R. Hughes-Jones Manchester Initial Performance Measurements With DataTAG PCs Gigabit Ethernet NICs (Work in progress Oct 02)
Advertisements

Ethernet “dominant” LAN technology: cheap $20 for 100Mbs!
The Internet Protocol.
Transmission Control Protocol (TCP)
CdL was here DataTAG/WP7 Amsterdam June 2002 R. Hughes-Jones Manchester 1 EU DataGrid - Network Monitoring Richard Hughes-Jones, University of Manchester.
1 NETWORKING TECHNOLOGIES (II) BUS Abdou Illia, Fall 2010 (October 20, 2010)
1 NETWORKING TECHNOLOGIES (II) BUS Abdou Illia, Spring 2007 (Week 9, Thursday 3/8/2007)
CMPE 80N - Introduction to Networks and the Internet 1 CMPE 80N Winter 2004 Lecture 8 Introduction to Networks and the Internet.
Circuit Switching (a) Circuit switching. (b) Packet switching.
Service Providers & Data Link & Physical layers Week 4 Lecture 1.
Review on Networking Technologies Linda Wu (CMPT )
 The Open Systems Interconnection model (OSI model) is a product of the Open Systems Interconnection effort at the International Organization for Standardization.
1.  A protocol is a set of rules that governs the communications between computers on a network.  Functions of protocols:  Addressing  Data Packet.
Process-to-Process Delivery:
Architecture and Dataflow Overview LHCb Data-Flow Review September 2001 Beat Jost Cern / EP.
1 Computer Communication & Networks Lecture 13 Datalink Layer: Local Area Network Waleed Ejaz
A TCP/IP transport layer for the DAQ of the CMS Experiment Miklos Kozlovszky for the CMS TriDAS collaboration CERN European Organization for Nuclear Research.
Huda AL_Omairl - Network 71 Protocols and Network Software.
Introduction network Protocol
Chapter 2 – X.25, Frame Relay & ATM. Switched Network Stations are not connected together necessarily by a single link Stations are typically far apart.
“L1 farm: some naïve consideration” Gianluca Lamanna (CERN) & Riccardo Fantechi (CERN/Pisa)
University of the Western Cape Chapter 12: The Transport Layer.
1 CHAPTER 8 TELECOMMUNICATIONSANDNETWORKS. 2 TELECOMMUNICATIONS Telecommunications: Communication of all types of information, including digital data,
CSE 6590 Department of Computer Science & Engineering York University 111/9/ :26 AM.
An Architecture and Prototype Implementation for TCP/IP Hardware Support Mirko Benz Dresden University of Technology, Germany TERENA 2001.
Lecture (Mar 23, 2000) H/W Assignment 3 posted on Web –Due Tuesday March 28, 2000 Review of Data packets LANS WANS.
Networking Basics CCNA 1 Chapter 11.
Lecture 4 Overview. Ethernet Data Link Layer protocol Ethernet (IEEE 802.3) is widely used Supported by a variety of physical layer implementations Multi-access.
Institute of Technology Sligo - Dept of Computing Chapter 12 The Transport Layer.
IEEE (Token Ring) LAN Dr. Sanjay P. Ahuja, Ph.D. Fidelity National Financial Distinguished Professor of CIS School of Computing, UNF.
1 DAQ System Realization DAQ Data Flow Review Sep th, 2001 Niko Neufeld CERN, EP.
Connectionless
1 OSI and TCP/IP Models. 2 TCP/IP Encapsulation (Packet) (Frame)
LHCb front-end electronics and its interface to the DAQ.
LHCb DAQ system LHCb SFC review Nov. 26 th 2004 Niko Neufeld, CERN.
BASIC NETWORK PROTOCOLS AND THEIR FUNCTIONS Created by: Ghadeer H. Abosaeed June 23,2012.
Computer Network Overview Ikjun Yeom. 2 What is a computer network? a collection of computing nodes and links connecting the nodes why we need it? –to.
LNL 1 SADIRC2000 Resoconto 2000 e Richieste LNL per il 2001 L. Berti 30% M. Biasotto 100% M. Gulmini 50% G. Maron 50% N. Toniolo 30% Le percentuali sono.
نظام المحاضرات الالكترونينظام المحاضرات الالكتروني.
Slide #1 CIT 380: Securing Computer Systems TCP/IP.
Protocol Layering Chapter 11.
TCP transfers over high latency/bandwidth networks & Grid DT Measurements session PFLDnet February 3- 4, 2003 CERN, Geneva, Switzerland Sylvain Ravot
DAQ interface + implications for the electronics Niko Neufeld LHCb Electronics Upgrade June 10 th, 2010.
1 Farm Issues L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
SRB data transmission Vito Palladino CERN 2 June 2014.
DAQ Overview + selected Topics Beat Jost Cern EP.
McGraw-Hill Chapter 23 Process-to-Process Delivery: UDP, TCP Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
L1/HLT trigger farm Bologna setup 0 By Gianluca Peco INFN Bologna Genève,
Lecture 4 unit 1. categories of network A network can be categorized as a 1.Local area network. 2.Wide area network. 3.Metropolitan area network. 4.Personal.
Introduction to DAQ Architecture Niko Neufeld CERN / IPHE Lausanne.
Introduction to Communication Lecture (10) 1. 2 Wired LANs: Ethernet IEEE Project 802 defines the LLC and MAC sublayers for all LANs including Ethernet.
R&D on data transmission FPGA → PC using UDP over 10-Gigabit Ethernet Domenico Galli Università di Bologna and INFN, Sezione di Bologna XII SuperB Project.
16 th IEEE NPSS Real Time Conference 2009 IHEP, Beijing, China, 12 th May, 2009 High Rate Packets Transmission on 10 Gbit/s Ethernet LAN Using Commodity.
Computer Networks 1000-Transport layer, TCP Gergely Windisch v spring.
1 NETWORKING TECHNOLOGIES (II) BUS Abdou Illia, Fall 2008 (November 4, 2008)
CHEP 2010, October 2010, Taipei, Taiwan 1 18 th International Conference on Computing in High Energy and Nuclear Physics This research project has.
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
Lec 2: Protocols.
The LHCb Event Building Strategy
Transport Layer Unit 5.
Data Link Issues Relates to Lab 2.
LHCb Trigger and Data Acquisition System Requirements and Concepts
John Harvey CERN EP/LBC July 24, 2001
Event Building With Smart NICs
Computer Network Overview
TCP/IP Protocol Suite: Review
CPEG514 Advanced Computer Networkst
LHCb Trigger, Online and related Electronics
Network Processors for a 1 MHz Trigger-DAQ System
Presentation transcript:

1 Network Performance Optimisation and Load Balancing Wulf Thannhaeuser

2 Network Performance Optimisation

3 Network Optimisation: Where? Read-out Network (RN) RU Control & Monitoring RU 2-4 GB/s 4 GB/s 20 MB/s LAN Read-out units (RU) Timing & Fast Control Front-End Electronics VDET TRACK ECAL HCAL MUON RICH LHC-B Detector L0 L1 Level 0 Trigger Level 1 Trigger 40 MHz 1 MHz 40 kHz Fixed latency 4.0  s Variable latency <1 ms Data rates 40 TB/s 1 TB/s Front-End Multiplexers (FEM) 1 MHz Front End Links Trigger Level 2 & 3 Event Filter SFC CPU Sub-Farm Controllers (SFC) Storage Throttle 60 SFCs with ca. 16 “off the shelf” PCs each

4 Network Optimisation: Why? Ethernet Speed: 10 Mb/s Fast Ethernet Speed:100 Mb/s (Fast) Ethernet Gigabit Ethernet Gigabit Ethernet Speed:1000 Mb/s (considering full-duplex:2000 Mb/s)

5 Network Optimisation: Why? An “average” CPU might not be able to process such a huge amount of data packets per second: -TCP/IP Overhead -Context Switching -Packet Checksums An “average” PCI Bus is 33 MHz, 32-bit wide. Theory: 1056 Mbit/s Actually: ca. 850 Mbit/s (PCI overhead, burstsize)

6 Network Optimisation: How? An “average” CPU might not be able to process such a huge amount of data packets per second: -TCP/IP Overhead -Context Switching -Packet Checksums An “average” PCI Bus is 33 MHz, 32-bit wide. Theory: 1056 Mbit/s Actually: ca. 850 Mbit/s (PCI overhead, burstsize) Reduce per packet Overhead: Replace TCP with UDP

7 TCP / UDP Comparison TCP (Transfer Control Protocol): - connection-oriented protocol - full-duplex - messages received in order, no loss or duplication  reliable but with overheads UDP (User Datagram Protocol): - messages called “datagrams” - messages may be lost or duplicated - messages may be received out of order  unreliable but potentially faster

8 Network Optimisation: How? An “average” CPU might not be able to process such a huge amount of data packets per second: -TCP/IP Overhead -Context Switching -Packet Checksums An “average” PCI Bus is 33 MHz, 32-bit wide. Theory: 1056 Mbit/s Actually: ca. 850 Mbit/s (PCI overhead, burstsize) Reduce per packet Overhead: Replace TCP with UDP Reduce number of packets: Jumbo Frames

9 Normal Ethernet Maximum Transmission Unit (MTU): 1500 bytes Ethernet with Jumbo Frames MTU: 9000 bytes

10 Test set-up Netperf is a benchmark for measuring network performance The systems tested were 800 and 1800 MHz Pentium PCs using (optical as well as copper) Gbit Ethernet NICs. The network set-up was always a simple point-to-point connection with a crossed twisted pair or optical cable. Results were not always symmetric: With two PCs of different performance, the benchmark results were usually better if data was sent from the slow PC to the fast PC, i.e. the receiving process is more expensive.

11 Results with the optimisations so far

12 Network Optimisation: How? An “average” CPU might not be able to process such a huge amount of data packets per second: -TCP/IP Overhead -Context Switching -Packet Checksums An “average” PCI Bus is 33 MHz, 32-bit wide. Theory: 1056 Mbit/s Actually: ca. 850 Mbit/s (PCI overhead, burstsize) Reduce per packet Overhead: Replace TCP with UDP Reduce number of packets: Jumbo Frames Reduce context switches: Interrupt coalescence

13 Interrupt Interrupt Coalescence Packet Processing without Interrupt coalescence: Memory Interrupt NIC CPU NIC Memory Packet Processing with Interrupt coalescence: CPU

14 Interrupt Coalescence: Results

15 Network Optimisation: How? An “average” CPU might not be able to process such a huge amount of data packets per second: -TCP /IP Overhead -Context Switching -Packet Checksums An “average” PCI Bus is 33 MHz, 32-bit wide. Theory: 1056 Mbit/s Actually: ca. 850 Mbit/s (PCI overhead, burstsize) Reduce per packet Overhead: Replace TCP with UDP Reduce number of packets: Jumbo Frames Reduce context switches: Interrupt coalescence Reduce context switches: Checksum Offloading

16 Checksum Offloading A checksum is a number calculated from the data transmitted and attached to the tail of each TCP/IP packet. Usually the CPU has to recalculate the checksum for each received TCP/IP packet in order to compare it with the checksum in the tail of the packet to detect transmission errors. With checksum offloading, the NIC performs this task. Therefore the CPU does not have to calculate the checksum and can perform other operations in the meanwhile.

17 Network Optimisation: How? An “average” CPU might not be able to process such a huge amount of data packets per second: -TCP/IP Overhead -Context Switching -Packet Checksums An “average” PCI Bus is 33 MHz, 32-bit wide. Theory: 1056 Mbit/s Actually: ca. 850 Mbit/s (PCI overhead, burstsize) Reduce per packet Overhead: Replace TCP with UDP Reduce number of packets: Jumbo Frames Reduce context switches: Interrupt coalescence Reduce context switches: Checksum Offloading Or buy a faster PC with a better PCI bus…

18 Load Balancing

19 Load Balancing: Where? Read-out Network (RN) RU Control & Monitoring RU 2-4 GB/s 4 GB/s 20 MB/s LAN Read-out units (RU) Timing & Fast Control Front-End Electronics VDET TRACK ECAL HCAL MUON RICH LHC-B Detector L0 L1 Level 0 Trigger Level 1 Trigger 40 MHz 1 MHz 40 kHz Fixed latency 4.0  s Variable latency <1 ms Data rates 40 TB/s 1 TB/s Front-End Multiplexers (FEM) 1 MHz Front End Links Trigger Level 2 & 3 Event Filter SFC CPU Sub-Farm Controllers (SFC) Storage Throttle 60 SFCs with ca. 16 “off the shelf” PCs each

20 Fast Ethernet Gigabit Ethernet … … SFC 2134 Load Balancing with round-robin Event Problem: The SFC doesn’t know if the node it wants to send the event to is ready to process it yet.

21 Fast Ethernet Gigabit Ethernet … … SFC 2134 Load Balancing with control-tokens Event With control tokens, nodes who are ready send a token, and every event is forwarded to the sender of a token. Token 2131

LHC Comp. Grid Testbed Structure SFC