CENIC Road to Ten Gigabit: Biggest Fastest in the West

Slides:



Advertisements
Similar presentations
TCP transfers over high latency/bandwidth network & Grid TCP Sylvain Ravot
Advertisements

Cheng Jin David Wei Steven Low FAST TCP: design and experiments.
High Speed Total Order for SAN infrastructure Tal Anker, Danny Dolev, Gregory Greenman, Ilya Shnaiderman School of Engineering and Computer Science The.
1 Internet End-to-end Monitoring Project at SLAC Les Cottrell, Connie Logg, Jerrod Williams, Gary Buhrmaster Site visit to SLAC by DoE program managers.
1 Testbeds Les Cottrell Site visit to SLAC by DoE program managers Thomas Ndousse & Mary Anne Scott April 27,
FAST TCP in Linux Cheng Jin David Wei
PFLDNet Argonne Feb 2004 R. Hughes-Jones Manchester 1 UDP Performance and PCI-X Activity of the Intel 10 Gigabit Ethernet Adapter on: HP rx2600 Dual Itanium.
ESLEA Bedfont Lakes Dec 04 Richard Hughes-Jones Network Measurement & Characterisation and the Challenge of SuperComputing SC200x.
Presented by Anshul Kantawala 1 Anshul Kantawala FAST TCP: From Theory to Experiments C. Jin, D. Wei, S. H. Low, G. Buhrmaster, J. Bunn, D. H. Choe, R.
1 High Performance WAN Testbed Experiences & Results Les Cottrell – SLAC Prepared for the CHEP03, San Diego, March 2003
1 Characterization and Evaluation of TCP and UDP-based Transport on Real Networks Les Cottrell Sun SuperG Spring 2005, April,
Hardware and Software Basics. Computer Hardware  Central Processing Unit - also called “The Chip”, a CPU, a processor, or a microprocessor  Memory (RAM)
The Effects of Systemic Packets Loss on Aggregate TCP Flows Thomas J. Hacker May 8, 2002 Internet 2 Member Meeting.
Is Lambda Switching Likely for Applications? Tom Lehman USC/Information Sciences Institute December 2001.
ICOM 6115©Manuel Rodriguez-Martinez ICOM 6115 – Computer Networks and the WWW Manuel Rodriguez-Martinez, Ph.D. Lecture 8.
Large File Transfer on 20,000 km - Between Korea and Switzerland Yusung Kim, Daewon Kim, Joonbok Lee, Kilnam Chon
A TCP/IP transport layer for the DAQ of the CMS Experiment Miklos Kozlovszky for the CMS TriDAS collaboration CERN European Organization for Nuclear Research.
Technology for Using High Performance Networks or How to Make Your Network Go Faster…. Robin Tasker UK Light Town Meeting 9 September.
Network Tests at CHEP K. Kwon, D. Han, K. Cho, J.S. Suh, D. Son Center for High Energy Physics, KNU, Korea H. Park Supercomputing Center, KISTI, Korea.
FAST TCP in Linux Cheng Jin David Wei Steven Low California Institute of Technology.
High TCP performance over wide area networks Arlington, VA May 8, 2002 Sylvain Ravot CalTech HENP Working Group.
High-speed TCP  FAST TCP: motivation, architecture, algorithms, performance (by Cheng Jin, David X. Wei and Steven H. Low)  Modifying TCP's Congestion.
Internet data transfer record between CERN and California Sylvain Ravot (Caltech) Paolo Moroni (CERN)
Iperf Quick Mode Ajay Tirumala & Les Cottrell. Sep 12, 2002 Iperf Quick Mode at LBL – Les Cottrell & Ajay Tirumala Iperf QUICK Mode Problem – Current.
Project Results Thanks to the exceptional cooperation spirit between the European and North American teams involved in the DataTAG project,
1 Achieving High Throughput on Fast Networks (Bandwidth Challenges and World Records) Les Cottrell & Yee-Ting Li Stanford Linear Accelerator Center Presented.
TCP transfers over high latency/bandwidth networks Internet2 Member Meeting HENP working group session April 9-11, 2003, Arlington T. Kelly, University.
Performance Engineering E2EpiPEs and FastTCP Internet2 member meeting - Indianapolis World Telecom Geneva October 15, 2003
30 June Wide Area Networking Performance Challenges Olivier Martin, CERN UK DTI visit.
GNEW2004 CERN March 2004 R. Hughes-Jones Manchester 1 Lessons Learned in Grid Networking or How do we get end-2-end performance to Real Users ? Richard.
TCP transfers over high latency/bandwidth networks & Grid DT Measurements session PFLDnet February 3- 4, 2003 CERN, Geneva, Switzerland Sylvain Ravot
FAST Protocols for High Speed Network David netlab, Caltech For HENP WG, Feb 1st 2003.
Final EU Review - 24/03/2004 DataTAG is a project funded by the European Commission under contract IST Richard Hughes-Jones The University of.
1 eVLBI Developments at Jodrell Bank Observatory Ralph Spencer, Richard Hughes- Jones, Simon Casey, Paul Burgess, The University of Manchester.
1 FAST TCP for Multi-Gbps WAN: Experiments and Applications Les Cottrell & Fabrizio Coccetti– SLAC Prepared for the Internet2, Washington, April 2003
1 Achieving Record Speed Trans-Atlantic End-to-end TCP Throughput Les Cottrell – SLAC Prepared for the NANOG meeting, Salt Lake City, June 2003
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Canadian Bioinformatics Workshops
Joint Genome Institute
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
TESTBED (Technical)
Computer Organization
R. Hughes-Jones Manchester
Introduction to Networking & Telecommunications
Introduction to networks
Prepared by Les Cottrell & Hadrien Bullot, SLAC & EPFL, for the
Networking between China and Europe
Drill Translate the following message:
High Speed File Replication
Using Netflow data for forecasting
TCP Performance over a 2.5 Gbit/s Transatlantic Circuit
Prepared by Les Cottrell & Hadrien Bullot, SLAC & EPFL, for the
Wide Area Networking at SLAC, Feb ‘03
High Performance Active End-to-end Network Monitoring
short term and long term speed, capacity, compression formats, access
Characterization and Evaluation of TCP and UDP-based Transport on Real Networks Les Cottrell, Saad Ansari, Parakram Khandpur, Ruchi Gupta, Richard Hughes-Jones,
Computer networking In this section of notes you will learn the rudiments of networking, the components of a network and how to secure a network.
My Experiences, results and remarks to TCP BW and CT Measurements Tools Jiří Navrátil SLAC.
Web Server Administration
Event Building With Smart NICs
Presentation at University of Twente, The Netherlands
Breaking the Internet2 Land Speed Record: Twice
Guinness Book of Records 2004
SUPPORTING MULTIMEDIA COMMUNICATION OVER A GIGABIT ETHERNET NETWORK
Wide-Area Networking at SLAC
LO2 – Understand Computer Software
Link from SLAC to SC2003 Les Cottrell, SLAC.
High-Performance Data Transport for Grid Applications
Presentation transcript:

CENIC Road to Ten Gigabit: Biggest Fastest in the West Les Cottrell – SLAC Prepared for the CENIC meeting, UCSB, May 2003 http://www.slac.stanford.edu/grp/scs/net/talk/cenic-may03.html Partially funded by DOE/MICS Field Work Proposal on Internet End-to-end Performance Monitoring (IEPM), by the SciDAC base program.

Outline Breaking the Internet2 Land Speed Record Who did it Not be confused with: Rocket-powered sled travels about 6,400 mph to break 1982 world land speed record, San Francisco Chronicle May 1, 2003 Who did it What was done How was it done? What was special about this anyway? Who needs it? So what’s next? Where do I find out more? Want to answer criticisms such as (from the NANOG list): Give 'em a million dollars, plus fiber from here to anywhere and let me muck with the TCP algorithm, and I can move a GigE worth of traffic too What am I missing here, theres OC48=2.4Gb, OC192=10Gb ... Production commercial networks ... Blow away these speeds on a regular basis. So, you turn down/off all the parts of TCP that allow you to share bandwidth ... I'm going to launch a couple DAT tapes across the parking lot with a spud gun and see if I can achieve 923 Mb/s! What kind of production environment needs a single TCP stream of data at 1Gbits/s over a 150ms latency link? High speeds are not important. High speeds at a *reasonable* cost are important. What you are describing is a high speed at an *unreasonable* cost. When a reporter from the New York Times asked Everest pioneer George Mallory why he wanted to scale the imposing peak, Mallory's answer was simple: "Because it is there."

Who did it: Collaborators and sponsors Caltech: Harvey Newman, Steven Low, Sylvain Ravot, Cheng Jin, Xiaoling Wei, Suresh Singh, Julian Bunn SLAC: Les Cottrell, Gary Buhrmaster, Fabrizio Coccetti LANL: Wu-chun Feng, Eric Weigle, Gus Hurwitz, Adam Englehart CERN: Olivier Martin, Paolo Moroni ANL: Linda Winkler DataTAG, StarLight, TeraGrid, SURFnet, NetherLight, Deutsche Telecom, Information Society Technologies Cisco, Level(3), Intel DoE, European Commission, NSF

What was done? Set a new Internet2 TCP land speed record, 10,619 Tbit-meters/sec (see http://lsr.internet2.edu/) With 10 streams achieved 8.6Gbps across US Beat the Gbps limit for a single TCP stream across the Atlantic – transferred a TByte in an hour One Terabyte transferred in less than one hour When From To Bottle-neck MTU Streams TCP Thru-put Nov ’02 (SC02) Amsterdam Sunny-vale 1 Gbps 9000B 1 Standard 923 Mbps Balti-more 10 Gbps 1500 10 FAST 8.6 Gbps Feb ‘03 Geneva 2.5 Gbps 2.38 Gbps

How was it done: Typical testbed 6*2cpu servers 12*2cpu servers 7609 T640 GSR 4 disk servers 4 disk servers OC192/POS (10Gbits/s) Chicago Sunnyvale 2.5Gbits/s (bottleneck) (EU+US) 7609 Sunnyvale section first deployed for SC2002 (Nov 02) 6*2cpu servers SNV Geneva CHI AMS > 10,000 km GVA

Typical Components Disk servers CPU Routers Compute servers Earthquake strap Typical Components Disk servers CPU Pentium 4 (Xeon) with 2.4GHz cpu For GE used Syskonnect NIC For 10GE used Intel NIC Linux 2.4.19 or 20 Routers Cisco GSR 12406 with OC192/POS & 1 and 10GE server interfaces (loaned, list > $1M) Cisco 760x Juniper T640 (Chicago) Level(3) OC192/POS fibers (loaned SNV-CHI monthly lease cost ~ $220K) Compute servers Heat sink GSR Note bootees

Challenges PCI bus limitations (66MHz * 64 bit = 4.2Gbits/s at best) At 2.5Gbits/s and 180msec RTT requires 120MByte window Some tools (e.g. bbcp) will not allow a large enough window – (bbcp limited to 2MBytes) Slow start problem at 1Gbits/s takes about 5-6 secs for 180msec link, i.e. if want 90% of measurement in stable (non slow start), need to measure for 60 secs need to ship >700MBytes at 1Gbits/s After a loss it can take over an hour for stock TCP (Reno) to recover to maximum throughput at 1Gbits/s i.e. loss rate of 1 in ~ 2 Gpkts (3Tbits), or BER of 1 in 3.6*1012 Sunnyvale-Geneva, 1500Byte MTU, stock TCP

What was special? 1/2 End-to-end application-to-application, single and multi-streams (not just internal backbone aggregate speeds) TCP has not run out of stream yet, scales from modem speeds into multi-Gbits/s region TCP well understood, mature, many good features: reliability etc. Friendly on shared networks New TCP stacks only need to be deployed at sender Often just a few data sources, many destinations No modifications to backbone routers etc No need for jumbo frames Used Commercial Off The Shelf (COTS) hardware and software

What was Special 2/2 Raise the bar on expectations for applications and users Some applications can use Internet backbone speeds Provide planning information The network is looking less like a bottleneck and more like a catalyst/enabler Reduce need to colocate data and cpu No longer ship literally truck or plane loads of data around the world Worldwide collaborations of people working with large amounts of data become increasingly possible

Who needs it? HENP – current driver Data intensive science: Multi-hundreds Mbits/s and Multi TByte files/day transferred across Atlantic today SLAC BaBar experiment already has almost a PByte stored Tbits/s and ExaBytes (1018) stored in a decade Data intensive science: Astrophysics, Global weather, Bioinformatics, Fusion, seismology… Industries such as aerospace, medicine, security … Future: Media distribution Gbits/s=2 full length DVD movies/minute 2.36Gbits/s is equivalent to Transferring a full CD in 2.3 seconds (i.e. 1565 CDs/hour) Transferring 200 full length DVD movies in one hour (i.e. 1 DVD in 18 seconds) Will sharing movies be like sharing music today? For example in medicine: Radiation oncology for each patient visit (e.g. a mammogram, 2 views, 2 breasts, 4Kx4Kx16bits / picture) between 100 and 500MBytes of data are collected 3M women with mammograms in UK alone, 400TBytes/year for UK Today often saved as film and diskettes. Costs savings: e.g. in storage space, and labor to keep it organized, automated digital library can scale to larger numbers of patients Better patient care: ability to share images in real-time with distributed experts, provide ability to mine information for better clues on treatments

When will it have an impact ESnet traffic doubling/year since 1990 SLAC capacity increasing by 90%/year since 1982 SLAC Internet traffic increased by factor 2.5 in last year International throughput increase by factor 10 in 4 years So traffic increases by factor 10 in 3.5 to 4 years, so in: 3.5 to 5 years 622 Mbps => 10Gbps 3-4 years 155 Mbps => 1Gbps 3.5-5 years 45Mbps => 622Mbps 2010-2012: 100s Gbits for high speed production net end connections 10Gbps will be mundane for R&E and business Home: doubling ~ every 2 years, 100Mbits/s by end of decade? Throughput from US Throughput Mbits/s

Impact Caught technical press attention Reported in places such as CNN, the BBC, Times of India, Wired, Nature Reported in English, Spanish, Portuguese, French, Dutch, Japanese Also on TechTV and ABC Radio

What’s next? Break 2.5Gbits/s limit Disk-to-disk throughput & useful applications Need faster cpus (extra 60% MHz/Mbits/s over TCP for disk to disk), understand how to use multi-processors Evaluate new stacks with real-world links, and other equipment Other NICs Response to congestion, pathologies Fairnesss Deploy for some major (e.g. HENP/Grid) customer applications Understand how to make 10GE NICs work well with 1500B MTUs Move from “hero” demonstrations to commonplace

More Information Not quite as short as: when Everest pioneer George Mallory was asked why he wanted to scale the imposing peak, Mallory's answer was simple: Because it is there. Internet2 Land Speed Record Publicity www-iepm.slac.stanford.edu/lsr/ www-iepm.slac.stanford.edu/lsr2/ 10GE tests www-iepm.slac.stanford.edu/monitoring/bulk/10ge/ sravot.home.cern.ch/sravot/Networking/10GbE/10GbE_test.html