The Design and Demonstration of the UltraLight Network Testbed Presented by Xun Su GridNets 2006, Oct.

Slides:



Advertisements
Similar presentations
Electronic Visualization Laboratory University of Illinois at Chicago EVL Optical Networking Research Oliver Yu Electronic Visualization Laboratory University.
Advertisements

Storage System Integration with High Performance Networks Jon Bakken and Don Petravick FNAL.
Current Testbed : 100 GE 2 sites (NERSC, ANL) with 3 nodes each. Each node with 4 x 10 GE NICs Measure various overheads from protocols and file sizes.
 Contributing >30% of throughput to ATLAS and CMS in Worldwide LHC Computing Grid  Reliant on production and advanced networking from ESNET, LHCNET and.
Shawn P. McKee University of Michigan University of Michigan UltraLight Meeting, NSF January 26, 2005 Network Working Group Report.
TeraPaths: End-to-End Network Path QoS Configuration Using Cross-Domain Reservation Negotiation Bruce Gibbard Dimitrios Katramatos Shawn McKee Dantong.
GNEW 2004 CERN, Geneva, Switzerland March 16th, 2004Shawn McKee The UltraLight Program UltraLight: An Overview for GNEW2004 Shawn McKee University of Michigan.
1 In VINI Veritas: Realistic and Controlled Network Experimentation Jennifer Rexford with Andy Bavier, Nick Feamster, Mark Huang, and Larry Peterson
Global Lambdas and Grids for Particle Physics in the LHC Era Harvey B. Newman Harvey B. Newman California Institute of Technology SC2005 Seattle, November.
Transport SDN: Key Drivers & Elements
UltraLight: Network & Applications Research at UF Dimitri Bourilkov University of Florida CISCO - UF Collaborative Team Meeting Gainesville, FL, September.
Questionaire answers D. Petravick P. Demar FNAL. 7/14/05 DLP -- GDB2 FNAL/T1 issues In interpreting the T0/T1 document how do the T1s foresee to connect.
TeraPaths : A QoS Collaborative Data Sharing Infrastructure for Petascale Computing Research USATLAS Tier 1 & Tier 2 Network Planning Meeting December.
Circuit Services - IPTV Christian Todorov Internet2 Fall Member Meeting October 9, 2007.
TeraPaths: A QoS Collaborative Data Sharing Infrastructure for Petascale Computing Research Bruce Gibbard & Dantong Yu High-Performance Network Research.
Large File Transfer on 20,000 km - Between Korea and Switzerland Yusung Kim, Daewon Kim, Joonbok Lee, Kilnam Chon
J. Bunn, D. Nae, H. Newman, S. Ravot, X. Su, Y. Xia California Institute of Technology High speed WAN data transfers for science Session Recent Results.
J. Bunn, D. Nae, H. Newman, S. Ravot, X. Su, Y. Xia California Institute of Technology State of the art in the use of long distance network International.
Shawn McKee / University of Michigan USATLAS Tier1 & Tier2 Network Planning Meeting December 14, BNL UltraLight Overview.
LHC Tier 2 Networking BOF Joe Metzger Joint Techs Vancouver 2005.
100G R&D at Fermilab Gabriele Garzoglio (for the High Throughput Data Program team) Grid and Cloud Computing Department Computing Sector, Fermilab Overview.
The Singapore Advanced Research & Education Network.
HOPI Update Rick Summerhill Director Network Research, Architecture, and Technologies Jerry Sobieski MAX GigaPoP and TSC Program Manager Mark Johnson MCNC.
High-quality Internet for higher education and research GigaPort  Overview SURFnet6 Niels den Otter SURFnet EVN-NREN Meeting Amsterdam October 12, 2005.
February 2006 Iosif Legrand 1 Iosif Legrand California Institute of Technology February 2006 February 2006 An Agent Based, Dynamic Service System to Monitor,
1 VINCI : Virtual Intelligent Networks for Computing Infrastructures An Integrated Network Services System to Control and Optimize Workflows in Distributed.
LambdaStation Monalisa DoE PI meeting September 30, 2005 Sylvain Ravot.
Thoughts on Future LHCOPN Some ideas Artur Barczyk, Vancouver, 31/08/09.
Shawn P. McKee / University of Michigan International ICFA Workshop on HEP Networking, Grid and Digital Divide Issues for Global e-Science May 25, 2005.
A Framework for Internetworking Heterogeneous High-Performance Networks via GMPLS and Web Services Xi Yang, Tom Lehman Information Sciences Institute (ISI)
Delivering Circuit Services to Researchers: The HOPI Testbed Rick Summerhill Director, Network Research, Architecture, and Technologies, Internet2 Joint.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
Copyright 2004 National LambdaRail, Inc N ational L ambda R ail Update 9/28/2004 Debbie Montano Director, Development & Operations
Shawn McKee University of Michigan University of Michigan UltraLight: A Managed Network Infrastructure for HEP CHEP06, Mumbai, India February 14, 2006.
ASCR/ESnet Network Requirements an Internet2 Perspective 2009 ASCR/ESnet Network Requirements Workshop April 15/16, 2009 Richard Carlson -- Internet2.
© 2006 National Institute of Informatics 1 Jun Matsukata National Institute of Informatics SINET3: The Next Generation SINET July 19, 2006.
Office of Science U.S. Department of Energy ESCC Meeting July 21-23, 2004 Network Research Program Update Thomas D. Ndousse Program Manager Mathematical,
1 Network Measurement Summary ESCC, Feb Joe Metzger ESnet Engineering Group Lawrence Berkeley National Laboratory.
1 Dynamic Service Provisioning in Converged Network Infrastructure Muckai Girish Atoga Systems.
Services provided by CERN’s IT Division Ethernet in controls applications European Organisation for Nuclear Research European Laboratory for Particle.
TeraPaths The TeraPaths Collaboration Presented by Presented by Dimitrios Katramatos, BNL Dimitrios Katramatos, BNL.
Online-Offsite Connectivity Experiments Catalin Meirosu *, Richard Hughes-Jones ** * CERN and Politehnica University of Bucuresti ** University of Manchester.
Dynamic Lightpath Services on the Internet2 Network Rick Summerhill Director, Network Research, Architecture, Technologies, Internet2 TERENA May.
1 Role of Ethernet in Optical Networks Debbie Montano Director R&E Alliances Internet2 Member Meeting, Apr 2006.
Rick Cavanaugh, University of Florida CHEP06 Mumbai, 13 February, 2006 An Ultrascale Information Facility for Data Intensive Research.
Terapaths: MPLS based Data Sharing Infrastructure for Peta Scale LHC Computing Bruce Gibbard and Dantong Yu USATLAS Computing Facility DOE Network Research.
TeraPaths: A QoS Enabled Collaborative Data Sharing Infrastructure for Petascale Computing Research The TeraPaths Project Team CHEP 06.
Advanced Networks: The Past and the Future – The Internet2 Perspective APAN 7 July 2004, Cairns, Australia Douglas Van Houweling, President & CEO Internet2.
. Large internetworks can consist of the following three distinct components:  Campus networks, which consist of locally connected users in a building.
Performance Engineering E2EpiPEs and FastTCP Internet2 member meeting - Indianapolis World Telecom Geneva October 15, 2003
TeraPaths: A QoS Enabled Collaborative Data Sharing Infrastructure for Petascale Computing Research The TeraPaths Project Team Usatlas Tier 2 workshop.
Dynamic Network Services In Internet2 John Vollbrecht /Dec. 4, 2006 Fall Members Meeting.
INDIANAUNIVERSITYINDIANAUNIVERSITY HOPI: Hybrid Packet and Optical Infrastructure Chris Robb and Jim Williams Indiana University 7 July 2004 Cairns, AU.
I Arlington, VA April 20th, 2004Shawn McKee The UltraLight Program UltraLight: An Overview for the Internet2 Spring 2004 Meeting Shawn McKee University.
S. Ravot, J. Bunn, H. Newman, Y. Xia, D. Nae California Institute of Technology CHEP 2004 Network Session September 1, 2004 Breaking the 1 GByte/sec Barrier?
A WAN-in-LAB for Protocol Development Netlab, Caltech George Lee, Lachlan Andrew, David Wei, Bartek Wydrowski, Cheng Jin, John Doyle, Steven Low, Harvey.
HENP SIG Austin, TX September 27th, 2004Shawn McKee The UltraLight Program UltraLight: An Overview and Update Shawn McKee University of Michigan.
J. Bunn, D. Nae, H. Newman, S. Ravot, R. Voicu, X. Su, Y. Xia California Institute of Technology US LHCNet US LHC Network Working Group Meeting October.
1 Network related topics Bartosz Belter, Wojbor Bogacki, Marcin Garstka, Maciej Głowiak, Radosław Krzywania, Roman Łapacz FABRIC meeting Poznań, 25 September.
Fall 2005 Internet2 Member Meeting International Task Force Julio Ibarra, PI Heidi Alvarez, Co-PI Chip Cox, Co-PI John Silvester, Co-PI September 19, 2005.
Recent experience with PCI-X 2.0 and PCI-E network interfaces and emerging server systems Yang Xia Caltech US LHC Network Working Group October 23, 2006.
100G R&D at Fermilab Gabriele Garzoglio (for the High Throughput Data Program team) Grid and Cloud Computing Department Computing Sector, Fermilab Overview.
J. Bunn, D. Nae, H. Newman, S. Ravot, X. Su, Y. Xia California Institute of Technology US LHCNet LHCNet WG September 12 th 2006.
HOPI Update Rick Summerhill Director Network Research, Architecture, and Technologies Internet2 Joint Techs 17 July 2006 University of Wisconsin, Madison,
Joint Genome Institute
California Institute of Technology
Establishing End-to-End Guaranteed Bandwidth Network Paths Across Multiple Administrative Domains The DOE-funded TeraPaths project at Brookhaven National.
Wide Area Networking at SLAC, Feb ‘03
Experiences from SLAC SC2004 Bandwidth Challenge
The UltraLight Program
Presentation transcript:

The Design and Demonstration of the UltraLight Network Testbed Presented by Xun Su GridNets 2006, Oct 2 nd, 2006

Long Term Trends in Network Traffic Volumes: X/10Yrs  SLAC Traffic ~400 Mbps; Growth in Steps (ESNet Limit): ~ 10X/4 Years.  Summer ‘05: 2x10 Gbps links: one for production, one for R&D  Projected: ~2 Terabits/s by ~2014 W. Johnston L. Cottrell Progress in Steps 10 Gbit/s TERABYTES Per Month ESnet Accepted Traffic 1990 – 2005 Exponential Growth: Avg. +82%/Year for the Last 15 Years

Motivation  Provide the network advances required to enable petabyte- scale analysis of globally distributed data.  Current Grid-based infrastructures provide massive computing and storage resources, but are currently limited by their treatment of the network as an external, passive, and largely unmanaged resource.  The mission of UltraLight is to:  Develop and deploy prototype global services which broaden existing Grid computing systems by promoting the network as an actively managed component.  Integrate and test UltraLight in Grid-based physics production and analysis systems currently under development in ATLAS and CMS.  Engineer and operate a trans- and intercontinental optical network testbed for broader community

UltraLight Backbone  The UltraLight testbed is a non-standard core network with dynamic links and varying bandwidth inter-connecting our nodes.  The core of UltraLight is dynamically evolving as function of available resources on other backbones such as NLR, HOPI, Abilene and ESnet.  The main resources for UltraLight:  US LHCnet (IP, L2VPN, CCC)  Abilene (IP, L2VPN)  ESnet (IP, L2VPN)  UltraScienceNet (L2)  Cisco Research Wave (10 Gb Ethernet over NLR)  NLR Layer 3 Service  HOPI NLR waves (Ethernet; provisioned on demand)  UltraLight nodes: Caltech, SLAC, FNAL, UF, UM, StarLight, CENIC PoP at LA, CERN, Seattle

UltraLight topology: point of presence

  GOAL: Determine an effective mix of bandwidth-management techniques for this application-space, particularly:  Best-effort and “scavenger” using “effective” protocols  MPLS with QOS-enabled packet switching  Dedicated paths provisioned with TL1 commands, GMPLS  PLAN: Develop, Test the most cost-effective integrated combination of network technologies on our unique testbed:  Exercise UltraLight applications on NLR, Abilene and campus networks, as well as LHCNet, and our international partners  Deploy and systematically study ultrascale protocol stacks (such as FAST) addressing issues of performance & fairness  Use MPLS/QoS and other forms of BW management, to optimize end-to-end performance among a set of virtualized disk servers  Address “end-to-end” issues, including monitoring and end- hosts UltraLight Network Engineering

UltraLight: Effective Protocols  The protocols used to reliably move data are a critical component of Physics “end-to-end” use of the network  TCP is the most widely used protocol for reliable data transport, but is becoming ever more ineffective for higher and higher bandwidth-delay networks.  UltraLight is exploring extensions to TCP (HSTCP, Westwood+, HTCP, FAST, MaxNet) designed to maintain fair-sharing of networks and, at the same time, to allow efficient, effective use of these networks.

FAST others Gigabit WAN  5x higher utilization  Small delay FAST: 95% Reno: 19% Random packet loss  10x higher throughput  Resilient to random loss FAST Protocol Comparisons

Optical Path Developments  Emerging “light path” technologies are arriving:  They can extend and augment existing grid computing infrastructures, currently focused on CPU/storage, to include the network as an integral Grid component.  Those technologies seem to be the most effective way to offer network resource provisioning on- demand between end-systems.  We are developing a multi-agent system for secure light path provisioning based on dynamic discovery of the topology in distributed networks (VINCI)  We are working to further develop this distributed agent system and to provide integrated network services capable of efficiently using and coordinating shared, hybrid networks, improving the performance and throughput for data intensive grid applications.  This includes services able to dynamically configure routers and to aggregate local traffic on dynamically created optical connections.

GMPLS Optical Path Provisioning  Collaboration efforts between UltraLight and Enlightened Computing.  Interconnecting Calient switches across the US for the purpose of unified GMPLS control plane.  Control Plane: IPv4 connectivity between site for control messages  Data Plane:  Cisco Research wave: between LA and Starlight  EnLIGHTened wave: between StarLight and MCNC Raleigh  LONI wave: between Starlight and LSU Baton Rouge over LONI DWDM.

GMPLS Optical Path Network Diagram

 Realtime end-to-end Network monitoring is essential for UltraLight.  We need to understand our network infrastructure and track its performance both historically and in real-time to enable the network as a managed robust component of our infrastructure. Caltech’s MonALISA: SLAC’s IEPM:  We have a new effort to push monitoring to the “ends” of the network: the hosts involved in providing services or user workstations. Monitoring for UltraLight

MonALISA UltraLight Repository The UL repository:

The Functionality of the VINCI System Layer 3 Layer 2 Layer 1 Site ASite BSite C MonALISA ML Agent MonALISA ML Agent MonALISA ML Agent ML proxy services Agent ROUTERS ETHERNET LAN-PHY or WAN-PHY DWDM FIBER Agent

SC|05 Global Lambdas for Particle Physics  We previewed the global-scale data analysis of the LHC Era Using a realistic mixture of streams:  Organized transfer of multi-TB event datasets; plus  Numerous smaller flows of physics data that absorb the remaining capacity  We used Twenty Two [*] 10 Gbps waves to carry bidirectional traffic between Fermilab, Caltech, SLAC, BNL, CERN and other partner Grid sites including: Michigan, Florida, Manchester, Rio de Janeiro (UERJ) and Sao Paulo (UNESP) in Brazil, Korea (KNU), and Japan (KEK)  The analysis software suites are based on the Grid-enabled UltraLight Analysis Environment (UAE) developed at Caltech and Florida, as well as the bbcp and Xrootd applications from SLAC, and dcache/SRM from FNAL  Monitored by Caltech’s MonALISA global monitoring and control system [*] 15 at the Caltech/CACR Booth and 7 at the FNAL/SLAC Booth

Switch and Server Interconnections at the Caltech Booth  15 10G Waves  64 10G Switch Ports: 2 Fully Populated Cisco 6509Es  43 Neterion 10 GbE NICs  70 nodes with 280 Cores  200 SATA Disks  40 Gbps (20 HBAs) to StorCloud  Thursday - Sunday

Monitoring NLR, Abilene/HOPI, LHCNet, USNet, TeraGrid, PWave, SCInet, Gloriad, JGN2, WHREN, other Int’l R&E Nets, and Grid Nodes at 250 Sites (250k Paramters) Simultaneously I. Legrand HEP at SC2005 Global Lambdas for Particle Physics

RESULTS  151 Gbps peak, 100+ Gbps of throughput sustained for hours: 475 Terabytes of physics data transported in < 24 hours  131 Gbps measured by SCInet BWC team on 17 of our waves  Sustained rate of 100+ Gbps translates to > 1 Petayte per day  Linux kernel optimized for TCP-based protocols, including Caltech’s FAST  Surpassing our previous SC2004 BWC Record of 101 Gbps Global Lambdas for Particle Physics Caltech/CACR and FNAL/SLAC Booths

Above 100 Gbps for Hours

475 TBytes Transported in < 24 Hours Sustained Peak Projects to > 1 Petabyte Per Day

It was the first time: a struggle for the equipment and the team We will stabilize, package and more widely deploy these methods and tools in 2006

SC05 BWC Lessons Learned  Take-aways from this Marathon exercise:  An optimized Linux kernel ( FAST-TCP + NFSv4) for data transport; after 7 full kernel-build cycles in 4 days  Scaling up SRM/gridftp to near 10 Gbps per wave, using Fermilab’s production clusters  A newly optimized application-level copy program, bbcp, that matches the performance of iperf under some conditions  Extensions of SLAC’s Xrootd, an optimized low-latency file access application for clusters, across the wide area  Understanding of the limits of 10 Gbps-capable computer systems, network switches and interfaces under stress

Thank You