TCP transfers over high latency/bandwidth network & Grid TCP Sylvain Ravot

Slides:



Advertisements
Similar presentations
Helping TCP Work at Gbps Cheng Jin the FAST project at Caltech
Advertisements

Appropriateness of Transport Mechanisms in Data Grid Middleware Rajkumar Kettimuthu 1,3, Sanjay Hegde 1,2, William Allcock 1, John Bresnahan 1 1 Mathematics.
LOGO Transmission Control Protocol 12 (TCP) Data Flow.
TCP Vegas: New Techniques for Congestion Detection and Control.
1 End to End Bandwidth Estimation in TCP to improve Wireless Link Utilization S. Mascolo, A.Grieco, G.Pau, M.Gerla, C.Casetti Presented by Abhijit Pandey.
Restricted Slow-Start for TCP William Allcock 1,2, Sanjay Hegde 3 and Rajkumar Kettimuthu 1,2 1 Argonne National Laboratory 2 The University of Chicago.
Presentation by Joe Szymanski For Upper Layer Protocols May 18, 2015.
1 TCP - Part II. 2 What is Flow/Congestion/Error Control ? Flow Control: Algorithms to prevent that the sender overruns the receiver with information.
Congestion Control Tanenbaum 5.3, /12/2015Congestion Control (A Loss Based Technique: TCP)2 What? Why? Congestion occurs when –there is no reservation.
Explicit Congestion Notification ECN Tilo Hamann Technical University Hamburg-Harburg, Germany.
1 Spring Semester 2007, Dept. of Computer Science, Technion Internet Networking recitation #7 TCP New Reno Vs. Reno.
High-performance bulk data transfers with TCP Matei Ripeanu University of Chicago.
MULTIMEDIA TRAFFIC MANAGEMENT ON TCP/IP OVER ATM-UBR By Dr. ISHTIAQ AHMED CH.
1 Chapter 3 Transport Layer. 2 Chapter 3 outline 3.1 Transport-layer services 3.2 Multiplexing and demultiplexing 3.3 Connectionless transport: UDP 3.4.
1 Internet Networking Spring 2004 Tutorial 10 TCP NewReno.
All rights reserved © 2006, Alcatel Accelerating TCP Traffic on Broadband Access Networks  Ing-Jyh Tsang 
Introduction 1 Lecture 14 Transport Layer (Congestion Control) slides are modified from J. Kurose & K. Ross University of Nevada – Reno Computer Science.
Grid simulation (AliEn) Network data transfer model Eugen Mudnić Technical university Split -FESB.
TCP: flow and congestion control. Flow Control Flow Control is a technique for speed-matching of transmitter and receiver. Flow control ensures that a.
KEK Network Qi Fazhi KEK SW L2/L3 Switch for outside connections Central L2/L3 Switch A Netscreen Firewall Super Sinet Router 10GbE 2 x GbE IDS.
TCP Throughput Collapse in Cluster-based Storage Systems
Raj Jain The Ohio State University R1: Performance Analysis of TCP Enhancements for WWW Traffic using UBR+ with Limited Buffers over Satellite.
Implementing High Speed TCP (aka Sally Floyd’s) Yee-Ting Li & Gareth Fairey 1 st October 2002 DataTAG CERN (Kinda!)
1 Transport Protocols (continued) Relates to Lab 5. UDP and TCP.
NORDUnet 2003, Reykjavik, Iceland, 26 August 2003 High-Performance Transport Protocols for Data-Intensive World-Wide Grids T. Kelly, University of Cambridge,
27th, Nov 2001 GLOBECOM /16 Analysis of Dynamic Behaviors of Many TCP Connections Sharing Tail-Drop / RED Routers Go Hasegawa Osaka University, Japan.
High TCP performance over wide area networks Arlington, VA May 8, 2002 Sylvain Ravot CalTech HENP Working Group.
HighSpeed TCP for High Bandwidth-Delay Product Networks Raj Kettimuthu.
Scavenger performance Cern External Network Division - Caltech Datagrid WP January, 2002.
Data Transport Challenges for e-VLBI Julianne S.O. Sansa* * With Arpad Szomoru, Thijs van der Hulst & Mike Garret.
Internet data transfer record between CERN and California Sylvain Ravot (Caltech) Paolo Moroni (CERN)
1 TCP - Part II Relates to Lab 5. This is an extended module that covers TCP data transport, and flow control, congestion control, and error control in.
TERENA Networking Conference, Zagreb, Croatia, 21 May 2003 High-Performance Data Transport for Grid Applications T. Kelly, University of Cambridge, UK.
What is TCP? Connection-oriented reliable transfer Stream paradigm
Transport Layer 3-1 Chapter 3 Transport Layer Computer Networking: A Top Down Approach 6 th edition Jim Kurose, Keith Ross Addison-Wesley March
Compound TCP in NS-3 Keith Craig 1. Worcester Polytechnic Institute What is Compound TCP? As internet speeds increased, the long ‘ramp’ time of TCP Reno.
CS640: Introduction to Computer Networks Aditya Akella Lecture 15 TCP – III Reliability and Implementation Issues.
TCP: Transmission Control Protocol Part II : Protocol Mechanisms Computer Network System Sirak Kaewjamnong Semester 1st, 2004.
1 CS 4396 Computer Networks Lab TCP – Part II. 2 Flow Control Congestion Control Retransmission Timeout TCP:
1 TCP Timeout And Retransmission Chapter 21 TCP sets a timeout when it sends data and if data is not acknowledged before timeout expires it retransmits.
CS640: Introduction to Computer Networks Aditya Akella Lecture 15 TCP – III Reliability and Implementation Issues.
Transport Layer3-1 Chapter 3 outline r 3.1 Transport-layer services r 3.2 Multiplexing and demultiplexing r 3.3 Connectionless transport: UDP r 3.4 Principles.
TCP transfers over high latency/bandwidth networks Internet2 Member Meeting HENP working group session April 9-11, 2003, Arlington T. Kelly, University.
1 TCP - Part II. 2 What is Flow/Congestion/Error Control ? Flow Control: Algorithms to prevent that the sender overruns the receiver with information.
Performance Engineering E2EpiPEs and FastTCP Internet2 member meeting - Indianapolis World Telecom Geneva October 15, 2003
Janey C. Hoe Laboratory for Computer Science at MIT 노상훈, Pllab.
TCP Congestion Control 컴퓨터공학과 인공지능 연구실 서 영우. TCP congestion control2 Contents 1. Introduction 2. Slow-start 3. Congestion avoidance 4. Fast retransmit.
Chapter 11.4 END-TO-END ISSUES. Optical Internet Optical technology Protocol translates availability of gigabit bandwidth in user-perceived QoS.
Development of a QoE Model Himadeepa Karlapudi 03/07/03.
TCP continued. Discussion – TCP Throughput TCP will most likely generate the saw tooth type of traffic. – A rough estimate is that the congestion window.
TCP transfers over high latency/bandwidth networks & Grid DT Measurements session PFLDnet February 3- 4, 2003 CERN, Geneva, Switzerland Sylvain Ravot
Final EU Review - 24/03/2004 DataTAG is a project funded by the European Commission under contract IST Richard Hughes-Jones The University of.
Peer-to-Peer Networks 13 Internet – The Underlay Network
TCP as a Reliable Transport. How things can go wrong… Lost packets Corrupted packets Reordered packets …Malicious packets…
Increasing TCP's CWND based on Throughput draft-you-iccrg-throughput-based-cwnd-increasing-00 Jianjie You IETF92 Dallas.
Transmission Control Protocol (TCP) TCP Flow Control and Congestion Control CS 60008: Internet Architecture and Protocols Department of CSE, IIT Kharagpur.
@Yuan Xue A special acknowledge goes to J.F Kurose and K.W. Ross Some of the slides used in this lecture are adapted from their.
Karn’s Algorithm Do not use measured RTT to update SRTT and SDEV Calculate backoff RTO when a retransmission occurs Use backoff RTO for segments until.
Chapter 3 outline 3.1 transport-layer services
Transport Protocols over Circuits/VCs
Generalizing The Network Performance Interference Problem
TCP - Part II Relates to Lab 5. This is an extended module that covers TCP flow control, congestion control, and error control in TCP.
TCP Performance over a 2.5 Gbit/s Transatlantic Circuit
Transport Layer: Congestion Control
TCP: Transmission Control Protocol Part II : Protocol Mechanisms
High-Performance Data Transport for Grid Applications
Review of Internet Protocols Transport Layer
Presentation transcript:

TCP transfers over high latency/bandwidth network & Grid TCP Sylvain Ravot

Tests configuration POS 155 MbpsGbE Pcgiga-gbe.cern.ch (Geneva) Plato.cacr.caltech.edu (California) Ar1-chicago Cernh 9 Calren2 / Abilene Lxusa-ge.cern.ch (Chicago) GbE u CERN (Geneva) Caltech (California) u RTT : 175 ms u Bandwith-delay product : 3,4 MBytes. u CERN Chicago u RTT : 110 ms u Bandwidth-delay-product : 1.9 MBytes. u Tcp flows were generated by Iperf. u Tcpdump was used to capture packets flows u Tcptrace and xplot were used to plot and summarize tcpdump data set.

TCP overview: Slow Start and congestion Avoidance Example Cwnd average of the last 10 samples. Cwnd average over the life of the connection to that point Slow start Congestion Avoidance SSTHRESH Here is an estimation of the cwnd (Output of TCPtrace): Slow start : fast increase of the cwnd Congestion Avoidance : slow increase of the window size

u During congestion avoidance and without any loss, the cwnd increases by one segment each RTT. In our case, we have no loss, so the window increases by 1460 bytes each 175 ms. If the cwnd is equal to 730 kbyte, it takes more than 5 minutes to have a cwnd larger than the bandwidth delay product (3,4 MByte). In other words, we have to wait almost 5 minutes to use the whole capacity of the link (155 Mbps)!!! SSTHRESH = 730Kbyte SSTHRESH =1460Kbyte Slow start Congestion avoidance Cwnd=f(time) ( Throughput = 33 Mbit/s)Cwnd=f(time) ( Throughput= 63 Mbit/s) Influence of the initial SSTHRESH on TCP performance

Reactivity u TCP reactivity r Time to recover a 200 Mbps throughput after a loss is larger than 50 seconds for a connection between Chicago and CERN. u A single loss is disastrous r TCP is much more sensitive to packet loss in WANs than in LANs 53 sec

Linux Patch GRID TCP u Parameter tuning è New parameter to better start a TCP transfer r Set the value of the initial SSTHRESH u Modifications of the TCP algorithms (RFC 2001) è Modification of the well-know congestion avoidance algorithm r During congestion avoidance, for every useful acknowledgement received, cwnd increases by M * (segment size) * (segment size) / cwnd. Its equivalent to increase cwnd by M segments each RTT. M is called congestion avoidance increment è Modification of the slow start algorithm r During slow start, for every useful acknowledgement received, cwnd increases by N segments. N is called slow start increment. è Note: N=1 and M=1 in common TCP implementations. u Smaller backoff (Not implemented yet) è Reduce the strong penalty imposed by a loss è Reproduce the behavior of a Multi-streams TCP connection. u Only the senders TCP stack need to be modified u Alternative to Multi-streams TCP transfers

TCP tuning by modifying the slow start increment Slow start, 0.8s Slow start, 1.2s Congestion window (cwnd) as function of the time Slow start increment = 1, throughput = 98 Mbit/s Congestion window (cwnd) as function of the time Slow start increment = 3, throughput = 116 Mbit/s Slow start, 2.0s Cwnd of the last 10 samples. Cwnd average over the life of the connection to that point Slow start, 0.65s Congestion window (cwnd) as function of the time Slow start increment = 2, throughput = 113 Mbit/s Congestion window (cwnd) as function of the time Slow start increment = 5, throughput = 119 Mbit/s

TCP tuning by modifying the congestion avoidance increment (1) Congestion window (cwnd) as function of the time – Congestion avoidance increment = 1, throughput = 37.5 Mbit/s Congestion window (cwnd) as function of the time – Congestion avoidance increment = 10, throughput = 61.5 Mbit/s SSTHREH = Mbyte Cwnd is increased by 1200 bytes in 27 sec. Cwnd is increased by bytes(10*1200) in 27 sec.

Benefice of larger congestion avoidance increment when losses occur u When a loss occur, the cwnd is divided by two. The performance is determined by the speed at which the cwnd increases after the loss. So higher is the congestion avoidance increment, better is the performance. u We simulate losses by using a program which drops packets according to a configured loss rate. For the next two plots, the program drop one packet every packets. 3) cwnd:=cwnd/2 2) Fast Recovery (Temporary state until the loss is repaired) 1) A packet is lost Congestion window (cwnd) as function of the time – Congestion avoidance increment = 1, throughput = 8 Mbit/s Congestion window (cwnd) as function of the time – Congestion avoidance increment = 10, throughput = 20 Mbit/s

TCP Performance Improvement Memory to memory transfers Without any tuning By tuning TCP buffers TCP Grid on 155 Mbps US-CERN Link New bottlenecks è Iperf is not able to perform long transfers è Linux station with 32 bit 33 Mhz PCI bus (Will replace with modern server) TCP Grid on 2 X 155 Mbps US-CERN Link TCP Grid on 622 Mbps US- CERN Link

Conclusion u To achieve high throughput over high latency/bandwidth network, we need to : è Set the initial slow start threshold (ssthresh) to an appropriate value for the delay and bandwidth of the link. è Avoid loss r by limiting the max cwnd size. è Recover fast if loss occurs: r Larger cwnd increment => we increase faster the cwnd after a loss r Smaller window reduction after a loss è …..