The TCP-ESTATS-MIB Matt Mathis John Heffner Raghu Reddy Pittsburgh Supercomputing Center Rajiv Raghunarayan Cisco Systems J. Saperia JDS Consulting, Inc.

Slides:



Advertisements
Similar presentations
Martin Suchara, Ryan Witt, Bartek Wydrowski California Institute of Technology Pasadena, U.S.A. TCP MaxNet Implementation and Experiments on the WAN in.
Advertisements

Measuring IP Performance Geoff Huston Telstra. What are you trying to measure? User experience –Responsiveness –Sustained Throughput –Application performance.
Tuning and Evaluating TCP End-to-End Performance in LFN Networks P. Cimbál* Measurement was supported by Sven Ubik**
Umut Girit  One of the core members of the Internet Protocol Suite, the set of network protocols used for the Internet. With UDP, computer.
Congestion Control Tanenbaum 5.3 Tanenbaum 6.5. Congestion Control Network Layer – Congestion control point to point Transport Layer – Congestion control.
24-1 Chapter 24. Congestion Control and Quality of Service (part 1) 23.1 Data Traffic 23.2 Congestion 23.3 Congestion Control 23.4 Two Examples.
Fundamentals of Computer Networks ECE 478/578 Lecture #20: Transmission Control Protocol Instructor: Loukas Lazos Dept of Electrical and Computer Engineering.
Transport Layer3-1 TCP. Transport Layer3-2 TCP: Overview RFCs: 793, 1122, 1323, 2018, 2581 r full duplex data: m bi-directional data flow in same connection.
BZUPAGES.COM 1 User Datagram Protocol - UDP RFC 768, Protocol 17 Provides unreliable, connectionless on top of IP Minimal overhead, high performance –No.
The TCP-ESTATS-MIB Matt Mathis John Heffner Pittsburgh Supercomputing Center Rajiv Raghunarayan Cisco Systems IETF 64, Nov 2005 draft-ietf-tsvwg-tcp-mib-extension-08.txt.
Restricted Slow-Start for TCP William Allcock 1,2, Sanjay Hegde 3 and Rajkumar Kettimuthu 1,2 1 Argonne National Laboratory 2 The University of Chicago.
Practice Questions: Congestion Control and Queuing
Identifying Performance Bottlenecks in CDNs through TCP-Level Monitoring Peng Sun Minlan Yu, Michael J. Freedman, Jennifer Rexford Princeton University.
Profiling Network Performance in Multi-tier Datacenter Applications
Leveraging Multiple Network Interfaces for Improved TCP Throughput Sridhar Machiraju, Prof. Randy Katz.
CSEE W4140 Networking Laboratory Lecture 7: TCP flow control and congestion control Jong Yul Kim
Leveraging Multiple Network Interfaces for Improved TCP Throughput Sridhar Machiraju SAHARA Retreat, June 10-12, 2002.
High-performance bulk data transfers with TCP Matei Ripeanu University of Chicago.
1 Chapter 3 Transport Layer. 2 Chapter 3 outline 3.1 Transport-layer services 3.2 Multiplexing and demultiplexing 3.3 Connectionless transport: UDP 3.4.
User-level Internet Path Diagnosis R. Mahajan, N. Spring, D. Wetherall and T. Anderson.
Performance Diagnostic Research at PSC Matt Mathis John Heffner Ragu Reddy 5/12/05 PathDiag ppt.
The Effects of Systemic Packets Loss on Aggregate TCP Flows Thomas J. Hacker May 8, 2002 Internet 2 Member Meeting.
Process-to-Process Delivery:
Peter O’Neil Executive Director November 29, 2007 MAX Fall Member Meeting.
Courtesy: Nick McKeown, Stanford 1 TCP Congestion Control Tahir Azim.
3: Transport Layer3b-1 Principles of Congestion Control Congestion: r informally: “too many sources sending too much data too fast for network to handle”
Pathdiag: Automatic TCP Diagnosis Matt Mathis John Heffner Ragu Reddy 8/01/08 PathDiag ppt.
COMT 4291 Communications Protocols and TCP/IP COMT 429.
3: Transport Layer3b-1 TCP: Overview RFCs: 793, 1122, 1323, 2018, 2581 r full duplex data: m bi-directional data flow in same connection m MSS: maximum.
Wireless TCP Prasun Dewan Department of Computer Science University of North Carolina
1 Transport Protocols (continued) Relates to Lab 5. UDP and TCP.
Transport Layer: TCP and UDP. Overview of TCP/IP protocols Comparing TCP and UDP TCP connection: establishment, data transfer, and termination Allocation.
Chapter 12 Transmission Control Protocol (TCP)
1 © 2003, Cisco Systems, Inc. All rights reserved. CCNA 1 v3.0 Module 11 TCP/IP Transport and Application Layers.
1 Internet Control Message Protocol (ICMP) Used to send error and control messages. It is a necessary part of the TCP/IP suite. It is above the IP module.
TAAD - A Tool for Traffic Analysis and Automatic Diagnosis Kathy L. Benninger NLANR/Pittsburgh Supercomputing Center.
Network Path and Application Diagnostics Matt Mathis John Heffner Ragu Reddy 4/24/06 PathDiag ppt.
Networking Fundamentals. Basics Network – collection of nodes and links that cooperate for communication Nodes – computer systems –Internal (routers,
National Center for Atmospheric Research Pittsburgh Supercomputing Center National Center for Supercomputing Applications Web100 Basil Irwin & George Brett.
Wide Area Network Performance Analysis Methodology Wenji Wu, Phil DeMar, Mark Bowden Fermilab ESCC/Internet2 Joint Techs Workshop 2007
1 Evaluating NGI performance Matt Mathis
1 Microsoft Windows 2000 Network Infrastructure Administration Chapter 4 Monitoring Network Activity.
Web100 Basil Irwin National Center for Atmospheric Research Matt Mathis Pittsburgh Supercomputing Center Halloween, 2000.
National Center for Atmospheric Research Pittsburgh Supercomputing Center National Center for Supercomputing Applications Web100 Roll Out I2 Members Meeting.
Winter 2008CS244a Handout 71 CS244a: An Introduction to Computer Networks Handout 7: Congestion Control Nick McKeown Professor of Electrical Engineering.
Advance Computer Networks Lecture#09 & 10 Instructor: Engr. Muhammad Mateen Yaqoob.
Prentice HallHigh Performance TCP/IP Networking, Hassan-Jain Chapter 13 TCP Implementation.
DoS/DDoS attack and defense
Development of a QoE Model Himadeepa Karlapudi 03/07/03.
TCP continued. Discussion – TCP Throughput TCP will most likely generate the saw tooth type of traffic. – A rough estimate is that the congestion window.
TCP transfers over high latency/bandwidth networks & Grid DT Measurements session PFLDnet February 3- 4, 2003 CERN, Geneva, Switzerland Sylvain Ravot
IP Configuration API. Network Interface Configuration NAIfconfigIsDeviceUp() NAIfconfigDeviceFromInterface() NAIfconfigBringDeviceUp() NAIfconfigSetIpAddress()
National Center for Atmospheric Research Pittsburgh Supercomputing Center National Center for Supercomputing Applications Web100 and Logistical Networking.
Network Path and Application Diagnostics Matt Mathis John Heffner Ragu Reddy 7/19/05 PathDiag ppt.
© 2015 Pittsburgh Supercomputing Center Opening the Black Box Using Web10G to Uncover the Hidden Side of TCP CC PI Meeting Austin, TX September 29, 2015.
TCP as a Reliable Transport. How things can go wrong… Lost packets Corrupted packets Reordered packets …Malicious packets…
11 CS716 Advanced Computer Networks By Dr. Amir Qayyum.
Connect communicate collaborate Performance Metrics & Basic Tools Robert Stoy, DFN EGI TF, Madrid September 2013.
3. END-TO-END PROTOCOLS (PART 1) Rocky K. C. Chang Department of Computing The Hong Kong Polytechnic University 22 March
By, Nirnimesh Ghose, Master of Science,
Topics discussed in this section:
The Transport Layer (TCP)
Network Path and Application Diagnostics
5. End-to-end protocols (part 1)
Chapter 5 TCP Transmission Control
Automatic TCP Buffer Tuning
Process-to-Process Delivery:
Process-to-Process Delivery: UDP, TCP
Anant Mudambi, U. Virginia
Presentation transcript:

The TCP-ESTATS-MIB Matt Mathis John Heffner Raghu Reddy Pittsburgh Supercomputing Center Rajiv Raghunarayan Cisco Systems J. Saperia JDS Consulting, Inc IETF 62, March 2005

The TCP Extended Statistics MIB Use TCP’s ideal diagnostic vantage point –Observe what the path is doing to segments –Observe what the application is doing to TCP TCP already measures many path properties –RTT, RTT variance, MTU, window size Easily instrumented to measure other properties –Reordering, loss rate, congestion signals Instrument why tcp_output() stops sending –Receiver window, congestion window or the sender? Per-connection controls to support workarounds

Example: a hard diagnostic problem Most symptoms scale with RTT –TCP Buffer Space, Network loss and reordering, etc –On a short path TCP compensates for the flaw Local Client to Server: all applications work –Including all standard diagnostics Remote Client to Server: all applications fail –Leading to faulty implication of other components This is the essence of the “End-to-end Problem”

How extended TCP statistics can help Without TCP instrumentation –Symptoms are reduced on short sections of long paths –Nearly all diagnostics yield a false pass on short paths With TCP ESTATS –Measure key properties of a short section of the path –Extrapolate to the full path to pass judgment –Tools get more sensitive as you test shorter sections –Example uses Web100.org instrumented TCP Target is a simple TCP discard service

Example diagnostic tool output End-to-end goal: 4 Mb/s over a 200 ms path including this section Tester at IP address: xxx.xxx Target at IP address: xxx.xxx Warning: TCP connection is not using SACK Fail: Received window scale is 0, it should be 2. Diagnosis: TCP on the test target is not properly configured for this path. > See TCP tuning instructions at Pass data rate check: maximum data rate was Mb/s Fail: loss event rate: % (3960 pkts between loss events) Diagnosis: there is too much background (non-congested) packet loss. The events averaged losses each, for a total loss rate of % FYI: To get 4 Mb/s with a 1448 byte MSS on a 200 ms path the total end-to-end loss budget is % (9733 pkts between losses). Warning: could not measure queue length due to previously reported bottlenecks Diagnosis: there is a bottleneck in the tester itself or test target (e.g insufficient buffer space or too much CPU load) > Correct previously identified TCP configuration problems > Localize all path problems by testing progressively smaller sections of the full path. FYI: This path may pass with a less strenuous application: Try rate=4 Mb/s, rtt=106 ms Or if you can raise the MTU: Try rate=4 Mb/s, rtt=662 ms, mtu=9000 Some events in this run were not completely diagnosed.

Changes with -06 draft Overhauled listen table –Designed to instrument generic SYN flood defenses Restructured per connection tables –Required “perf” table Expose TCP state variables (no memory footprint) Basic performance instrumentation First tier diagnostic instrumentation –3 optional tables for more detailed diagnosis Path (loss, reordering, duplication, etc) Stack (impact and state of control algorithms) Application (Is the data motion timely?) –1 table of writeable controls for workarounds LimCwnd, LinRwnd, and LimSsthresh Cleanup descriptions, references

Open Technical Issues Too much required stuff? –There may be further juggling between tables The proper SMI type for “Duration”, we want: –microsecond resolution for short flows –days (or months?) scale for exit stats –meaningful deltas at all scales What have we forgotten?

Next steps Last call for input from implementers, researchers MIB doctor review WG last call sometime this summer