Shawn McKee University of Michigan University of Michigan UltraLight: A Managed Network Infrastructure for HEP CHEP06, Mumbai, India February 14, 2006.

Slides:



Advertisements
Similar presentations
Electronic Visualization Laboratory University of Illinois at Chicago EVL Optical Networking Research Oliver Yu Electronic Visualization Laboratory University.
Advertisements

Shawn P. McKee University of Michigan University of Michigan UltraLight Meeting, NSF January 26, 2005 Network Working Group Report.
Network+ Guide to Networks, Fourth Edition
TeraPaths: End-to-End Network Path QoS Configuration Using Cross-Domain Reservation Negotiation Bruce Gibbard Dimitrios Katramatos Shawn McKee Dantong.
GNEW 2004 CERN, Geneva, Switzerland March 16th, 2004Shawn McKee The UltraLight Program UltraLight: An Overview for GNEW2004 Shawn McKee University of Michigan.
Copyright 2009 FUJITSU TECHNOLOGY SOLUTIONS PRIMERGY Servers and Windows Server® 2008 R2 Benefit from an efficient, high performance and flexible platform.
1 In VINI Veritas: Realistic and Controlled Network Experimentation Jennifer Rexford with Andy Bavier, Nick Feamster, Mark Huang, and Larry Peterson
Network+ Guide to Networks, Fourth Edition Chapter 1 An Introduction to Networking.
Transport SDN: Key Drivers & Elements
UltraLight: Network & Applications Research at UF Dimitri Bourilkov University of Florida CISCO - UF Collaborative Team Meeting Gainesville, FL, September.
TeraPaths : A QoS Collaborative Data Sharing Infrastructure for Petascale Computing Research USATLAS Tier 1 & Tier 2 Network Planning Meeting December.
Network+ Guide to Networks, Fourth Edition Chapter 1 An Introduction to Networking.
1 ESnet Network Measurements ESCC Feb Joe Metzger
Circuit Services - IPTV Christian Todorov Internet2 Fall Member Meeting October 9, 2007.
TeraPaths: A QoS Collaborative Data Sharing Infrastructure for Petascale Computing Research Bruce Gibbard & Dantong Yu High-Performance Network Research.
J. Bunn, D. Nae, H. Newman, S. Ravot, X. Su, Y. Xia California Institute of Technology High speed WAN data transfers for science Session Recent Results.
J. Bunn, D. Nae, H. Newman, S. Ravot, X. Su, Y. Xia California Institute of Technology State of the art in the use of long distance network International.
Dan Nae California Institute of Technology US LHCNet Update.
Dan Nae California Institute of Technology The US LHCNet Project ICFA Workshop, Krakow October 2006.
Shawn McKee / University of Michigan USATLAS Tier1 & Tier2 Network Planning Meeting December 14, BNL UltraLight Overview.
HOPI Update Rick Summerhill Director Network Research, Architecture, and Technologies Jerry Sobieski MAX GigaPoP and TSC Program Manager Mark Johnson MCNC.
February 2006 Iosif Legrand 1 Iosif Legrand California Institute of Technology February 2006 February 2006 An Agent Based, Dynamic Service System to Monitor,
HOPI: Making the Connection Chris Robb 23 June 2004 Broomfield, CO Quilt Meeting.
1 VINCI : Virtual Intelligent Networks for Computing Infrastructures An Integrated Network Services System to Control and Optimize Workflows in Distributed.
Shawn P. McKee / University of Michigan International ICFA Workshop on HEP Networking, Grid and Digital Divide Issues for Global e-Science May 25, 2005.
A Framework for Internetworking Heterogeneous High-Performance Networks via GMPLS and Web Services Xi Yang, Tom Lehman Information Sciences Institute (ISI)
Delivering Circuit Services to Researchers: The HOPI Testbed Rick Summerhill Director, Network Research, Architecture, and Technologies, Internet2 Joint.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
Copyright 2004 National LambdaRail, Inc N ational L ambda R ail Update 9/28/2004 Debbie Montano Director, Development & Operations
DYNES Storage Infrastructure Artur Barczyk California Institute of Technology LHCOPN Meeting Geneva, October 07, 2010.
John D. McCoy Principal Investigator Tom McKenna Project Manager UltraScienceNet Research Testbed Enabling Computational Genomics Project Overview.
The Internet2 HENP Working Group Internet2 Spring Meeting April 9, 2003.
Office of Science U.S. Department of Energy ESCC Meeting July 21-23, 2004 Network Research Program Update Thomas D. Ndousse Program Manager Mathematical,
Dynamic Lightpath Services on the Internet2 Network Rick Summerhill Director, Network Research, Architecture, Technologies, Internet2 TERENA May.
1 Role of Ethernet in Optical Networks Debbie Montano Director R&E Alliances Internet2 Member Meeting, Apr 2006.
ATLAS Great Lakes Tier-2 (AGL-Tier2) Shawn McKee (for the AGL Tier2) University of Michigan US ATLAS Tier-2 Meeting at Harvard Boston, MA, August 17 th,
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Rick Cavanaugh, University of Florida CHEP06 Mumbai, 13 February, 2006 An Ultrascale Information Facility for Data Intensive Research.
Terapaths: MPLS based Data Sharing Infrastructure for Peta Scale LHC Computing Bruce Gibbard and Dantong Yu USATLAS Computing Facility DOE Network Research.
TeraPaths: A QoS Enabled Collaborative Data Sharing Infrastructure for Petascale Computing Research The TeraPaths Project Team CHEP 06.
Advanced Networks: The Past and the Future – The Internet2 Perspective APAN 7 July 2004, Cairns, Australia Douglas Van Houweling, President & CEO Internet2.
The Design and Demonstration of the UltraLight Network Testbed Presented by Xun Su GridNets 2006, Oct.
Lambda Station Matt Crawford, Fermilab co-PI: Don Petravick, Fermilab co-PI: Harvey Newman, Caltech.
Performance Engineering E2EpiPEs and FastTCP Internet2 member meeting - Indianapolis World Telecom Geneva October 15, 2003
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
TeraPaths: A QoS Enabled Collaborative Data Sharing Infrastructure for Petascale Computing Research The TeraPaths Project Team Usatlas Tier 2 workshop.
TCP transfers over high latency/bandwidth networks & Grid DT Measurements session PFLDnet February 3- 4, 2003 CERN, Geneva, Switzerland Sylvain Ravot
INDIANAUNIVERSITYINDIANAUNIVERSITY HOPI: Hybrid Packet and Optical Infrastructure Chris Robb and Jim Williams Indiana University 7 July 2004 Cairns, AU.
9 th November 2005David Foster CERN IT-CS 1 LHCC Review WAN Status David Foster Networks and Communications Systems Group Leader.
I Arlington, VA April 20th, 2004Shawn McKee The UltraLight Program UltraLight: An Overview for the Internet2 Spring 2004 Meeting Shawn McKee University.
S. Ravot, J. Bunn, H. Newman, Y. Xia, D. Nae California Institute of Technology CHEP 2004 Network Session September 1, 2004 Breaking the 1 GByte/sec Barrier?
-1- ESnet On-Demand Secure Circuits and Advance Reservation System (OSCARS) David Robertson Internet2 Joint Techs Workshop July 18,
HENP SIG Austin, TX September 27th, 2004Shawn McKee The UltraLight Program UltraLight: An Overview and Update Shawn McKee University of Michigan.
Fall 2005 Internet2 Member Meeting International Task Force Julio Ibarra, PI Heidi Alvarez, Co-PI Chip Cox, Co-PI John Silvester, Co-PI September 19, 2005.
Run - II Networks Run-II Computing Review 9/13/04 Phil DeMar Networks Section Head.
US ATLAS Tier-2 Networking Shawn McKee University of Michigan US ATLAS Tier-2 Meeting San Diego, March 8 th, 2007.
Recent experience with PCI-X 2.0 and PCI-E network interfaces and emerging server systems Yang Xia Caltech US LHC Network Working Group October 23, 2006.
TeraPaths: A QoS Enabled Collaborative Data Sharing Infrastructure for Petascale Computing Research The TeraPaths Project Team Usatlas Tier 2 workshop.
J. Bunn, D. Nae, H. Newman, S. Ravot, X. Su, Y. Xia California Institute of Technology US LHCNet LHCNet WG September 12 th 2006.
HOPI Update Rick Summerhill Director Network Research, Architecture, and Technologies Internet2 Joint Techs 17 July 2006 University of Wisconsin, Madison,
Grid Optical Burst Switched Networks
University of Maryland College Park
California Institute of Technology
Establishing End-to-End Guaranteed Bandwidth Network Paths Across Multiple Administrative Domains The DOE-funded TeraPaths project at Brookhaven National.
UltraLight Status Report
The Internet2 HENP Working Group Internet2 Fall Meeting October 27, 2002 Hi Everyone, I am happy to have a chance to brief you on the High Energy and.
Wide Area Networking at SLAC, Feb ‘03
Network+ Guide to Networks, Fourth Edition
The UltraLight Program
Presentation transcript:

Shawn McKee University of Michigan University of Michigan UltraLight: A Managed Network Infrastructure for HEP CHEP06, Mumbai, India February 14, 2006

S. McKee, UM - CHEP06 2 Introduction The UltraLight Collaboration was formed to develop the network as a managed component of our infrastructure for High- Energy Physics (See Rick Cavanaugh’s overview from yesterday) Given a funding level at 50% of our request we had to reduce our scope while maintaining the core network infrastucture. This talk will focus on the UltraLight network and the activities of the UL network group to enable effective use of 10 Gbps networks.

S. McKee, UM - CHEP06 3 Overview Overview of UltraLight Network Effort Overview of UltraLight Network Effort Network Engineering Plans and Status Network Engineering Plans and Status Monitoring Monitoring Kernel/End-host Development Kernel/End-host Development Status and Summary Status and Summary

S. McKee, UM - CHEP06 4 S. McKee (UM, Team Leader) S. Ravot (LHCNet) D. Nae (LHCNet) R. Summerhill (Abilene/HOPI) D. Pokorney (FLR) S. Gerstenberger (MiLR) C. Griffin (UF) S. Bradley (BNL) J. Bigrow (BNL) The UltraLight Network Engineering Team J. Ibarra (WHREN, AW) C. Guok (ESnet) L. Cottrell (SLAC) C. Heerman (I2/HOPI) D. Petravick (FNAL) M. Crawford (FNAL) R. Hockett (UM) E. Rubi (FIU) Many individuals have worked deploying and maintaining the UltraLight Network

S. McKee, UM - CHEP06 5 UltraLight Backbone UltraLight has a non-standard core network with dynamic links and varying bandwidth inter-connecting our nodes.  Optical Hybrid Global Network The core of UltraLight is dynamically evolving as function of available resources on other backbones such as NLR, HOPI, Abilene and ESnet. The main resources for UltraLight: LHCnet (IP, L2VPN, CCC) LHCnet (IP, L2VPN, CCC) Abilene (IP, L2VPN) Abilene (IP, L2VPN) ESnet (IP, L2VPN) ESnet (IP, L2VPN) UltraScienceNet (L2) UltraScienceNet (L2) Cisco NLR wave (Ethernet) Cisco NLR wave (Ethernet) Cisco Research Service (L3) Cisco Research Service (L3) HOPI NLR waves (Ethernet; provisioned on demand) HOPI NLR waves (Ethernet; provisioned on demand) UltraLight nodes: Caltech, SLAC, FNAL, UF, UM, StarLight, CENIC PoP at LA, CERN, Seattle UltraLight nodes: Caltech, SLAC, FNAL, UF, UM, StarLight, CENIC PoP at LA, CERN, Seattle

S. McKee, UM - CHEP06 6 UltraLight Network Infrastructure Elements  Trans-US 10G  s Riding on NLR, Plus CENIC, FLR, MiLR  LA – CHI (4 Waves): HOPI (2 Waves), USN, and Cisco Research  CHI – JAX (Florida Lambda Rail/NLR)  Dark Fiber Caltech – L.A.: 2 X 10G Waves (One to WAN In Lab); 10G Wave L.A. to Sunnyvale for UltraScience Net Connection  Dark Fiber with 10G Waves (2 Waves): StarLight – Fermilab  Dedicated Wave StarLight (1 + 2 Waves) – Michigan Light Rail  SLAC: ESnet MAN to Provide Links (from July): One for Production, and One for Research  Partner with Advanced Research & Production Networks  LHCNet (Starlight- CERN), Abilene/HOPI, ESnet, NetherLight, GLIF, UKLight, CA*net4  Intercont’l extensions: Brazil (CHEPREO/WHREN), GLORIAD, Tokyo, AARNet, Taiwan, China

S. McKee, UM - CHEP06 7 UltraLight Network Infrastructure Elements  Trans-US 10G  s Riding on NLR, Plus CENIC, FLR, MiLR  LA – CHI (4 Waves): HOPI (2 Waves), USN, and Cisco Research  CHI – JAX (Florida Lambda Rail/NLR)  Dark Fiber Caltech – L.A.: 2 X 10G Waves (One to WAN In Lab); 10G Wave L.A. to Sunnyvale for UltraScience Net Connection  Dark Fiber with 10G Waves (2 Waves): StarLight – Fermilab  Dedicated Wave StarLight (1 + 2 Waves) – Michigan Light Rail  SLAC: ESnet MAN to Provide Links (from July): One for Production, and One for Research  Partner with Advanced Research & Production Networks  LHCNet (Starlight- CERN), Abilene/HOPI, ESnet, NetherLight, GLIF, UKLight, CA*net4  Intercont’l extensions: Brazil (CHEPREO/WHREN), GLORIAD, Tokyo, AARNet, Taiwan, China

S. McKee, UM - CHEP06 8 UltraLight Points-of-Presence StarLight (Chicago) HOPI (2 x 10GE), USNet (2 x 10GE), NLR (4 x 10GE) UM (3 x 10GE), TeraGrid, ESnet, Abilene FNAL, US-LHCNet (2 x 10GE), TIFR (Mumbai) MANLAN (New York) HOPI (2 x 10GE), US-LHCNet (2 x 10GE), BNL, Buffalo (2 x 10GE), Cornell, Nevis Seattle GLORIAD, JGN2, Pwave, NLR (2 x 10GE) CENIC (Los-Angeles) HOPI (2 x 10GE), NLR (4 x 10GE) Caltech (2 x 10GE), Pwave Level3 (Sunnyvale) USNet (2 x 10GE), NLR, SLAC

S. McKee, UM - CHEP06 9 International Partners One of the UltraLight program’s strengths is the large number of important international partners we have: UltraLight is well positioned to develop and coordinate global advances to networks for LHC Physics

S. McKee, UM - CHEP06 10 Global Services support management / co-scheduling of multiple resource types with a strategic end-to-end view. Global Services support management / co-scheduling of multiple resource types with a strategic end-to-end view.  Provide strategic recovery mechanisms from system failures  Schedule decisions based on CPU, I/O, Network capability and End-to-end task performance estimates, incl. loading effects  Constrained by local and global policies Global Services Consist of: Global Services Consist of:  Network and System Resource Monitoring  Network Path Discovery and Construction Services  Policy Based Job Planning Services  Task Execution Services These types of services are required to deliver a managed network. These types of services are required to deliver a managed network. UltraLight Global Services See “VINCI” talk on Thursday

S. McKee, UM - CHEP06 11  GOAL: Determine an effective mix of bandwidth- management techniques for this application-space, particularly: Best-effort and “scavenger” using “effective” protocols Best-effort and “scavenger” using “effective” protocols MPLS with QOS-enabled packet switching MPLS with QOS-enabled packet switching Dedicated paths arranged with TL1 commands, GMPLS Dedicated paths arranged with TL1 commands, GMPLS  PLAN: Develop, Test the most cost-effective integrated combination of network technologies on our unique testbed: 1. Exercise UltraLight applications on NLR, Abilene and campus networks, as well as LHCNet, and our international partners 2. Deploy and systematically study ultrascale protocol stacks (such as FAST) addressing issues of performance & fairness 3. Use MPLS/QoS and other forms of BW management, to optimize end-to-end performance among a set of virtualized disk servers 4. Address “end-to-end” issues, including monitoring and end-hosts UltraLight Network Engineering

S. McKee, UM - CHEP06 12 UltraLight: Effective Protocols The protocols used to reliably move data are a critical component of Physics “end-to-end” use of the network TCP is the most widely used protocol for reliable data transport, but is becoming ever more ineffective for higher and higher bandwidth-delay networks. UltraLight is exploring extensions to TCP (HSTCP, Westwood+, HTCP, FAST, MaxNet) designed to maintain fair-sharing of networks and, at the same time, to allow efficient, effective use of these networks. We identified a need to provide an “UltraLight” kernel to make protocol testing easy among the UltraLight sites. UltraLight plans to identify the most effective fair protocol and implement it in support of our “Best Effort” network components.

S. McKee, UM - CHEP06 13 FAST others Gigabit WAN 5x higher utilization Small delay FAST: 95% Reno: 19% Random packet loss 10x higher thruput Resilient to random loss FAST Protocol Comparisons

S. McKee, UM - CHEP06 14 MPLS/QoS for UltraLight UltraLight plans to explore the full range of end-to-end connections across the network, from best-effort, packet- switched through dedicated end-to-end light-paths. MPLS paths with QoS attributes fill a middle ground in this network space and allow fine-grained allocation of virtual pipes, sized to the needs of the application or user. UltraLight, in conjunction with the DoE/MICS funded TeraPaths and OSCARS effort, is working toward extensible solutions for implementing such capabilities in next generation networks TeraPaths SC|05 QoS Data Movement BNL → UM See “Terapaths” talk in previous session today

S. McKee, UM - CHEP06 15 Optical Path Developments Emerging “light path” technologies are arriving: They can extend and augment existing grid computing infrastructures, currently focused on CPU/storage, to include the network as an integral Grid component. They can extend and augment existing grid computing infrastructures, currently focused on CPU/storage, to include the network as an integral Grid component. Those technologies seem to be the most effective way to offer network resource provisioning on-demand between end- systems. Those technologies seem to be the most effective way to offer network resource provisioning on-demand between end- systems. We have developed a multi-agent system for secure light path provisioning based on dynamic discovery of the topology in distributed networks. [See VINCI talk on Thursday, Feb. 16] We are working to further develop this distributed agent system and to provide integrated network services capable to efficiently use and coordinate shared, hybrid networks and to improve the performance and throughput for data intensive grid applications. This includes services able to dynamically configure routers and to aggregate local traffic on dynamically created optical connections.

S. McKee, UM - CHEP06 16 Realtime end-to-end Network monitoring is essential for UltraLight. You can’t manage what you can’t see! We need to understand our network infrastructure and track its performance both historically and in real-time to enable the network as a managed robust component of our infrastructure. MonALISA IEPM We have a new effort to push monitoring to the “ends” of the network: the hosts involved in providing services or user workstations. Monitoring for UltraLight

S. McKee, UM - CHEP06 17 MonALISA UltraLight Repository The UL repository:

S. McKee, UM - CHEP06 18 Host Monitoring with UltraLight Many “network” problems are actually endhost problems: misconfigured or underpowered end-systems The LISA application (see Iosif’s talk later) was designed to monitor the endhost and its view of the network. For SC|05 we developed a Perl script to gather the relevant host details related to network performance and integrated the script with ApMon (an API for MonALISA) to allow us to “publish” this data to a MonALISA repository. Information on the system information, TCP configuration and network device setup was gathered and accessible from one site. Future plans are to coordinate this with LISA and deploy this as part of OSG. The Tier-2 centers are a primary target.

S. McKee, UM - CHEP06 19 Host Monitoring with UltraLight Many “network” problems are actually endhost problems: misconfigured or underpowered end-systems The LISA application (see Iosif’s talk later) was designed to monitor the endhost and its view of the network. For SC|05 we developed a Perl script to gather the relevant host details related to network performance and integrated the script with ApMon (an API for MonALISA) to allow us to “publish” this data to a MonALISA repository. Information on the system information, TCP configuration and network device setup was gathered and accessible from one site. Future plans are to coordinate this with LISA and deploy this as part of OSG. The Tier-2 centers are a primary target. Network Device Information TCP Settings Host/System Information

S. McKee, UM - CHEP06 20 UltraLight Kernel Kernels and the associated device drivers are critical to the achievable performance of hardware and software. The FAST protocol implementation for Linux requires a modified kernel to work.. We learned to deal with many pitfalls in the configuration and varying versions of linux kernels, particularly how they impact the performance of the system on the network. Currently we are working on a new version based upon Will include FAST, NFSv4, newest drivers for 10GE NICs and RAID cards. Because of the need to have FAST easily available, and the desire to achieve the best performance possible, we plan for all UltraLight “hosts” to install and utilize this Kernel.

S. McKee, UM - CHEP06 21 End-Systems performance Latest disk to disk over 10Gbps WAN: 4.3 Gbits/sec (536 MB/sec) - 8 TCP streams from CERN to Caltech; windows, 1TB file, 24 JBOD disks Quad Opteron AMD GHz processors with 3 AMD-8131 chipsets: bit/133MHz PCI-X slots. 3 Supermicro Marvell SATA disk controllers + 24 SATA 7200rpm SATA disks Local Disk IO – 9.6 Gbits/sec (1.2 GBytes/sec read/write, with <20% CPU utilization) Local Disk IO – 9.6 Gbits/sec (1.2 GBytes/sec read/write, with <20% CPU utilization) 10GE NIC 10 GE NIC – 9.3 Gbits/sec (memory-to-memory, with 52% CPU utilization, PCI-X 2.0 Caltech-Starlight) 10 GE NIC – 9.3 Gbits/sec (memory-to-memory, with 52% CPU utilization, PCI-X 2.0 Caltech-Starlight) 2*10 GE NIC (802.3ad link aggregation) – 11.1 Gbits/sec (mem-2-mem) 2*10 GE NIC (802.3ad link aggregation) – 11.1 Gbits/sec (mem-2-mem) Need PCI-Express?, TCP offload engines Need PCI-Express?, TCP offload engines Use 64 bit OS? Which architectures and hardware? Use 64 bit OS? Which architectures and hardware? GOAL: Single server to server at 1 GByte/sec Discussions are underway with 3Ware, Myricom to try to prototype viable servers capable of driving 10 GE networks in the WAN. Targeting Spring 2006

S. McKee, UM - CHEP06 22 UltraLight Network Summary  The network technical group has been hard at work on implementing UltraLight. Network and basic services are deployed and operating.  Significant progress has been made in monitoring, kernels, prototype services and disk-to-disk transfers in the WAN.  Our global collaborators are working with us on achieving the UltraLight vision will be busy, critical (for LHC) and productive year working on UltraLight!