TCP Performance over Gigabit Networks Arne Øslebø NORDUnet2002.

Slides:



Advertisements
Similar presentations
Tuning and Evaluating TCP End-to-End Performance in LFN Networks P. Cimbál* Measurement was supported by Sven Ubik**
Advertisements

Diagnosing Performance Overheads in the Xen Virtual Machine Environment Aravind Menon Willy Zwaenepoel EPFL, Lausanne Jose Renato Santos Yoshio Turner.
What is Hardware? Hardware is the physical parts of the computer system – the parts that you can touch and see. A motherboard, a CPU, a keyboard and a.
DataTAG CERN Oct 2002 R. Hughes-Jones Manchester Initial Performance Measurements With DataTAG PCs Gigabit Ethernet NICs (Work in progress Oct 02)
Berlin – November 10th, 2011 NetFPGA Programmable Networking for High-Speed Network Prototypes, Research and Teaching Presented by: Andrew W. Moore (University.
1 Trace collection in the UNC-CH DiRT lab The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
1 Web Server Performance in a WAN Environment Vincent W. Freeh Computer Science North Carolina State Vsevolod V. Panteleenko Computer Science & Engineering.
Identifying Performance Bottlenecks in CDNs through TCP-Level Monitoring Peng Sun Minlan Yu, Michael J. Freedman, Jennifer Rexford Princeton University.
Intro to computers Hardware Operating System, Communication Computers PowerHardware Software OS Application COMMUNICATION.
CdL was here DataTAG/WP7 Amsterdam June 2002 R. Hughes-Jones Manchester 1 EU DataGrid - Network Monitoring Richard Hughes-Jones, University of Manchester.
R. Hughes-Jones Video_TAG Meeting Oct 1999 Some Studies of H.323 VideoConference Traffic on LANs and WANs Studies done for PPNCG & PIPVIC2 Lab Tests of.
DataTAG Meeting CERN 7-8 May 03 R. Hughes-Jones Manchester 1 High Throughput: Progress and Current Results Lots of people helped: MB-NG team at UCL MB-NG.
Hosted VMM Architecture Advantages: –Installs and runs like an application –Portable – host OS does I/O access –Coexists with applications running on.
CdL was here DataTAG CERN Sep 2002 R. Hughes-Jones Manchester 1 European Topology: NRNs & Geant SuperJANET4 CERN UvA Manc SURFnet RAL.
Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck Pavan Balaji  Hemal V. Shah ¥ D. K. Panda 
System Architecture A Reconfigurable and Programmable Gigabit Network Interface Card Jeff Shafer, Hyong-Youb Kim, Paul Willmann, Dr. Scott Rixner Rice.
Sven Ubik, CESNET TNC2004, Rhodos, 9 June 2004 Performance monitoring of high-speed networks from NREN perspective.
Sven Ubik, Petr Žejdl CESNET TNC2008, Brugges, 19 May 2008 Passive monitoring of 10 Gb/s lines with PC hardware.
1.  Project Goals.  Project System Overview.  System Architecture.  Data Flow.  System Inputs.  System Outputs.  Rates.  Real Time Performance.
Implementation and Performance Analysis of a Delay Based Packet Scheduling Algorithm for an Embedded Open Source Router Master’s Thesis Presentation June.
Large File Transfer on 20,000 km - Between Korea and Switzerland Yusung Kim, Daewon Kim, Joonbok Lee, Kilnam Chon
High Performance User-Level Sockets over Gigabit Ethernet Pavan Balaji Ohio State University Piyush Shivam Ohio State University.
RiceNIC: A Reconfigurable and Programmable Gigabit Network Interface Card Jeff Shafer, Dr. Scott Rixner Rice Computer Architecture:
Comparison of Public End-to-End Bandwidth Estimation tools on High-Speed Links Alok Shriram, Margaret Murray, Young Hyun, Nevil Brownlee, Andre Broido,
Srihari Makineni & Ravi Iyer Communications Technology Lab
High-speed TCP  FAST TCP: motivation, architecture, algorithms, performance (by Cheng Jin, David X. Wei and Steven H. Low)  Modifying TCP's Congestion.
Operation System Support for Multi-User, Remote, Graphical Interaction Alexander Ya-li Wong, Margo Seltzer Harvard University, Division of Engineering.
TCP performance Sven Ubik FTP throughput capacity load ftp.uninett.no 12.3 Mb/s 1.2 Gb/s 80 Mb/s (6.6%) ftp.stanford.edu 1.3 Mb/s 600.
1 Network Performance Optimisation and Load Balancing Wulf Thannhaeuser.
Harnessing Multicore Processors for High Speed Secure Transfer Raj Kettimuthu Argonne National Laboratory.
Pavel Cimbál, Sven Ubik CESNET TNC2005, Poznan, 9 June 2005 Tools for TCP performance debugging.
Masaki Hirabaru NICT Koganei 3rd e-VLBI Workshop October 6, 2004 Makuhari, Japan Performance Measurement on Large Bandwidth-Delay Product.
1 Presented By: Eyal Enav and Tal Rath Eyal Enav and Tal Rath Supervisor: Mike Sumszyk Mike Sumszyk.
Vladimír Smotlacha CESNET High-speed Programmable Monitoring Adapter.
TCP Offload Through Connection Handoff Hyong-youb Kim and Scott Rixner Rice University April 20, 2006.
Wide Area Network Performance Analysis Methodology Wenji Wu, Phil DeMar, Mark Bowden Fermilab ESCC/Internet2 Joint Techs Workshop 2007
TERENA Networking Conference, Zagreb, Croatia, 21 May 2003 High-Performance Data Transport for Grid Applications T. Kelly, University of Cambridge, UK.
Debugging end-to-end performance in commodity operating system Pavel Cimbál, CTU, Sven Ubik, CESNET,
An Efficient Gigabit Ethernet Switch Model for Large-Scale Simulation Dong (Kevin) Jin.
TCP continued. Discussion – TCP Throughput TCP will most likely generate the saw tooth type of traffic. – A rough estimate is that the congestion window.
BMTS 242: Computer and Systems Lecture 2: Memory, and Software Yousef Alharbi Website
Visual Flow Analysis: What do real-world problems look like? Brent Draney NERSC Center Division, LBNL 2/07/06.
Exploiting Task-level Concurrency in a Programmable Network Interface June 11, 2003 Hyong-youb Kim, Vijay S. Pai, and Scott Rixner Rice Computer Architecture.
System Programming Basics Cha#2 H.M.Bilal. Operating Systems An operating system is the software on a computer that manages the way different programs.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
RAM (Memory) Written by Tim Keyser Georgia CTAE Resource Network 2010.
L1/HLT trigger farm Bologna setup 0 By Gianluca Peco INFN Bologna Genève,
Sven Ubik, Aleš Friedl CESNET TNC 2009, Malaga, Spain, 11 June 2009 Experience with passive monitoring deployment in GEANT2 network.
Studies of LHCb Trigger Readout Network Design Karol Hennessy University College Dublin Karol Hennessy University College Dublin.
16 th IEEE NPSS Real Time Conference 2009 IHEP, Beijing, China, 12 th May, 2009 High Rate Packets Transmission on 10 Gbit/s Ethernet LAN Using Commodity.
Recent experience with PCI-X 2.0 and PCI-E network interfaces and emerging server systems Yang Xia Caltech US LHC Network Working Group October 23, 2006.
DataGrid WP7 Meeting Jan 2002 R. Hughes-Jones Manchester Initial Performance Measurements Gigabit Ethernet NICs 64 bit PCI Motherboards (Work in progress)
Wide Area Network Performance Analysis Methodology
Network Tools and Utilities
Speaker : Che-Wei Chang
R. Hughes-Jones Manchester
Informational Tutorial
نماذج من تطبيقات شبكات الحساب مكونات شبكة الحاسب
Event Building With Smart NICs
Sven Ubik TCP performance Sven Ubik
MB – NG SuperJANET4 Development Network
Network Processors for a 1 MHz Trigger-DAQ System
IP Control Gateway (IPCG)
High-Performance Data Transport for Grid Applications
ECE 671 – Lecture 8 Network Adapters.
Achieving reliable high performance in LFNs (long-fat networks)
Cluster Computers.
Using NetLogger and Web100 for TCP analysis
Presentation transcript:

TCP Performance over Gigabit Networks Arne Øslebø NORDUnet2002

Modern Network Interface Cards ● Programmable network processors ● Hardware checksuming ● Offloads main CPU ● Interrupt coalescing ● Less interrupts ● On board memory ● Zero-copy ● Mainly operating system issue

Test tools ● netperf – used for measuring throughput between two systems. – ● vmstat / Microsoft Management Console – gives information about CPU load and interrupts/s. ● netstat – gives information about packet loss, retransmitions etc. ● One testrun consists of: – running netperf using various TCP window sizes. – for each TCP window size, netperf is run three or more times. Results presented are average values. – Before running netperf, information is collected from netstat – While running netperf, information is collected from vmstat.

Laboratory testing ● Hardware: ● Motherboard: Supermicro 370DE6 ● Memory: 1GB SDRAM ● CPU: 1GHz Pentium III ● NIC: 3COM 3C985B-SX ● Operating systems: ● Linux: ● NetBSD: current ● Windows 2000: service pack 2 ● RTT=0.227ms

Point to point - throughput

Point to point - CPU load

Point to point - interrupts/s

Analyzing results ● Problems with using tcpdump ● Negative effect on throughput ● Drops many packets ● Interrupt Coalescing ● Gives inaccurate time stamps ● Low performance using Windows 2000 probably related to driver.

Real world experiments MTU size: n=normal(1500) j=jumboframe(4470) Naming: nn country: c=Czech republic f=Finland n=Norway s=Switzerland These tests have been possible thanks to the good help from: ● Jari Miettinen, CSC ● Tuomas Toropainen, CSC ● Simon Leinen, SWITCH ● Sven Ubik, CESNET RTT:

Finland - Norway

Finland - Switzerland

Norway - Switzerland

Variations over a day

Czech-Finland-Norway

Analyzing results ● Good performance using jumboframes. ● Problems with large TCP window sizes ● Asymetric results – Probably driver and hardware related ● Variance in results to Switzerland caused by SACKs and retransmits From To MTU=1500MTU=4470 To From

Conclusions and further work ● Conclusions: – Diffucult to achieve high performance – Still possible if you use: ● A good PC ● A good network card ● A good driver for the card – Large TCP windows leads to performance degradation. – Default maximum TCP window size too small for high speed networks. ● Further work: – Will investigate asymetric results in more details. ● Use DAG cards for packet capture to get accurate TCP traces. ● Full report will be posted on