LHC Open Network Environment an update Artur Barczyk California Institute of Technology Atlas TIM, Annecy, April 20 th, 2012 1.

Slides:



Advertisements
Similar presentations
NORDUnet Nordic infrastructure for Research & Education LHCONE “Point-to-Point Connection Service” Service Definition Jerry Sobieski.
Advertisements

Dynamically Provisioned Networks as a Substrate for Science David Foster CERN.
Trial of the Infinera PXM Guy Roberts, Mian Usman.
Transatlantic Connectivity in LHCONE Artur Barczyk California Institute of Technology LHCONE Meeting Washington DC, June 13 th,
ESnet On-demand Secure Circuits and Advance Reservation System (OSCARS) Chin Guok Network Engineering Group Thomas Ndousse Visit February Energy.
IRNC Special Projects: IRIS and DyGIR Eric Boyd, Internet2 October 5, 2011.
National Science Foundation Arlington, Virginia January 7-8, 2013 Tom Lehman University of Maryland Mid-Atlantic Crossroads.
NORDUnet NORDUnet The Fibre Generation Lars Fischer CTO NORDUnet.
| BoD over GÉANT (& NRENs) for FIRE and GENI users GENI-FIRE Workshop Washington DC, 17th-18th Sept 2015 Michael Enrico CTO (GÉANT Association)
ASGC T1 report HSIN-YEN CHEN APAN38 NanTou 13 Aug
LHC network in Asia HSIN-YEN CHEN APAN38 NanTou 13 Aug
HOPI Update Rick Summerhill Director Network Research, Architecture, and Technologies Jerry Sobieski MAX GigaPoP and TSC Program Manager Mark Johnson MCNC.
Connect communicate collaborate LHCONE L3VPN Status Update Mian Usman LHCONE Meeting Rome 28 th – 29 th Aprils 2014.
1 ESnet Update Joint Techs Meeting Minneapolis, MN Joe Burrescia ESnet General Manager 2/12/2007.
LHC Open Network Environment LHCONE Artur Barczyk California Institute of Technology LISHEP Workshop on the LHC Rio de Janeiro, July 9 th,
Thoughts on Future LHCOPN Some ideas Artur Barczyk, Vancouver, 31/08/09.
A Framework for Internetworking Heterogeneous High-Performance Networks via GMPLS and Web Services Xi Yang, Tom Lehman Information Sciences Institute (ISI)
Bandwidth-on-Demand evolution Gerben van Malenstein Fall 2011 Internet2 Member Meeting Raleigh, North Carolina, USA – October 3, 2011.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
LHC Open Network Environment LHCONE David Foster CERN IT LCG OB 30th September
Scientific Networking: The Cause of and Solution to All Problems April 14 th Workshop on High Performance Applications of Cloud and Grid Tools Jason.
Connect communicate collaborate GÉANT3 Services Connectivity and Monitoring Services by and for NRENs Ann Harding, SWITCH TNC 2010.
1 Networking in the WLCG Facilities Michael Ernst Brookhaven National Laboratory.
Global Networking for the LHC Artur Barczyk California Institute of Technology ECOC Conference Geneva, September 18 th,
TeraPaths TeraPaths: Establishing End-to-End QoS Paths through L2 and L3 WAN Connections Presented by Presented by Dimitrios Katramatos, BNL Dimitrios.
DYNES Storage Infrastructure Artur Barczyk California Institute of Technology LHCOPN Meeting Geneva, October 07, 2010.
Connect. Communicate. Collaborate perfSONAR MDM Service for LHC OPN Loukik Kudarimoti DANTE.
ANSE: Advanced Network Services for [LHC] Experiments Artur Barczyk, California Institute of Technology for the ANSE team LHCONE Point-to-Point Service.
Building Dynamic Lightpaths in GÉANT Tangui Coulouarn, DeIC E-Infrastructure Autumn Workshop, Chiinău 8 September 2014.
LHC OPEN NETWORK ENVIRONMENT STATUS UPDATE Artur Barczyk/Caltech Tokyo, May 2013 May 14, 2013
From the Transatlantic Networking Workshop to the DAM Jamboree to the LHCOPN Meeting (Geneva-Amsterdam-Barcelona) David Foster CERN-IT.
NORDUnet Nordic Infrastructure for Research & Education Workshop Introduction - Finding the Match Lars Fischer LHCONE Workshop CERN, December 2012.
LHC Open Network Environment Architecture Overview and Status Artur Barczyk/Caltech LHCONE meeting Amsterdam, September 26 th,
Connect communicate collaborate LHCONE moving forward Roberto Sabatino, Mian Usman DANTE LHCONE technical workshop SARA, 1-2 December 2011.
LHC network in Asia HSIN-YEN CHEN LHCOPN workshop INFN 28 Apr 2014.
OSCARS Roadmap Chin Guok Feb 6, 2009 Energy Sciences Network Lawrence Berkeley National Laboratory Networking for the Future of.
LHCONE Point-to-Point Circuit Experiment Authentication and Authorization Model Discussion LHCONE meeting, Rome April 28-29, 2014 W. Johnston, Senior Scientist.
Connect communicate collaborate LHCONE Diagnostic & Monitoring Infrastructure Richard Hughes-Jones DANTE Delivery of Advanced Network Technology to Europe.
TeraPaths: A QoS Enabled Collaborative Data Sharing Infrastructure for Petascale Computing Research The TeraPaths Project Team Usatlas Tier 2 workshop.
ANSE: Advanced Network Services for Experiments Institutes: –Caltech (PI: H. Newman, Co-PI: A. Barczyk) –University of Michigan (Co-PI: S. McKee) –Vanderbilt.
Dynamic Network Services In Internet2 John Vollbrecht /Dec. 4, 2006 Fall Members Meeting.
SDN/OPENFLOW IN LHCONE A discussion June 5, 2013
9 th November 2005David Foster CERN IT-CS 1 LHCC Review WAN Status David Foster Networks and Communications Systems Group Leader.
DICE: Authorizing Dynamic Networks for VOs Jeff W. Boote Senior Network Software Engineer, Internet2 Cándido Rodríguez Montes RedIRIS TNC2009 Malaga, Spain.
NORDUnet Nordic Infrastructure for Research & Education Report of the CERN LHCONE Workshop May 2013 Lars Fischer LHCONE Meeting Paris, June 2013.
Connect communicate collaborate LHCONE European design & implementation Roberto Sabatino, DANTE LHCONE Meeting, Washington, June
Strawman LHCONE Point to Point Experiment Plan LHCONE meeting Paris, June 17-18, 2013.
LHC OPN and LHC ONE ”LHC networks” Marco Marletta (GARR) – Stefano Zani (INFN CNAF) Workshop INFN CCR - GARR, Napoli, 2012.
The DYNES Architecture & LHC Data Movement Shawn McKee/University of Michigan For the DYNES collaboration Contributions from Artur Barczyk, Eric Boyd,
ESnet’s Use of OpenFlow To Facilitate Science Data Mobility Chin Guok Inder Monga, and Eric Pouyoul OGF 36 OpenFlow Workshop Chicago, Il Oct 8, 2012.
ASGC Activities Update Hsin-Yen Chen ASGC LHCONE/LHCOPN meeting Taipei 13 Mar
100GE Upgrades at FNAL Phil DeMar; Andrey Bobyshev CHEP 2015 April 14, 2015.
Network CERN Tony Cass, 22 nd March 2016.
1 Network Measurement Challenges LHC E2E Network Research Meeting October 25 th 2006 Joe Metzger Version 1.1.
© 2006 Open Grid Forum The Network Services Interface An Overview of the NSI Framework and the GLIF Automated GOLE dynamic network provisioning demonstration.
LHCOPN / LHCONE Status Update John Shade /CERN IT-CS Summary of the LHCOPN/LHCONE meeting in Amsterdam Grid Deployment Board, October 2011.
LHCOPN / LHCONE Status Update John Shade /CERN IT-CS Summary of the LHCOPN/LHCONE meeting in Berkeley Grid Deployment Board, January 2012.
Inter-Domain Network Provisioning Technology for LHC data transfer
LHCOPN/LHCONE status report pre-GDB on Networking CERN, Switzerland 10th January 2017
Multi-layer software defined networking in GÉANT
Openflow-based Multipath Switching in Wide Area Networks
Dynamic Network Services In Internet2
LHCONE Point- to-Point service WORKSHOP SUMMARY
Establishing End-to-End Guaranteed Bandwidth Network Paths Across Multiple Administrative Domains The DOE-funded TeraPaths project at Brookhaven National.
LHCONE L3VPN Status update Mian Usman LHCOPN-LHCONE meeting
Integration of Network Services Interface version 2 with the JUNOS Space SDK
LHC Open Network Project status and site involvement
LHCONE In Europe Mian Usman DANTE.
K. Schauerhammer, K. Ullmann (DFN)
OSCARS Roadmap Chin Guok
Presentation transcript:

LHC Open Network Environment an update Artur Barczyk California Institute of Technology Atlas TIM, Annecy, April 20 th,

In a nutshell, LHCONE was born (out the 2010 transatlantic workshop at CERN) to address two main issues: –To ensure that the services to the science community maintain their quality and reliability –To protect existing R&E infrastructures against the potential “threats” of very large data flows that look like ‘denial of service’ attacks LHCONE is expected to –Provide some guarantees of performance Large data flows across managed bandwidth that would provide better determinism than shared IP networks Segregation from competing traffic flows Manage capacity as # sites x Max flow/site x # Flows increases –Provide ways for better utilisation of resources Use all available resources, especially transatlantic Provide Traffic Engineering and flow management capability –Leverage investments being made in advanced networking LHCONE in a Nutshell 2

Some Background So far, T1-T2, T2-T2, and T3 data movements have been using General Purpose R&E Network infrastructure –Shared resources (with other science fields) –Mostly best effort service Increased reliance on network performance  need more than best effort Separate large LHC data flows from routed R&E GPN Collaboration on global scale, diverse environment, many parties –Solution has to be Open, Neutral and Diverse –Agility and Expandability Scalable in bandwidth, extent and scope Organic activity, growing over time according to needs LHCONE Services being constructed: –Multipoint, virtual network (logical traffic separation and TE possibility) –Static/dynamic point-to-point Layer 2 circuits (guaranteed bandwidth, high-throughput data movement) –Monitoring/diagnostic 3

LHCONE Initial Architecture, The 30’000 ft view 4 LHCOPN Meeting, Lyon, February 2011

2011 Activity During 2011, LHCONE consisted of two implementations, each successful in its own scope: –Transatlantic Layer 2 domain Aka vlan 3000, implemented by USLHCNet, SURFnet, Netherlight, Starlight; CERN, Caltech, UMICH, MIT, PIC, UNAM –European VPLS domain Mostly vlan 2000, implemented in RENATER, DFN, GARR, interconnected through GEANT backbone (DANTE) In addition, Internet2 deployed a VPLS based pilot in the US Problem: Connecting the VPLS domains at Layer 2 with other components of the LHCONE The new multipoint architecture therefore foresees inter-domain connections at Layer 3 5

New Timescales In the meantime, pressure lowered by increase in backbone capacities and increased GPN transatlantic capacity –True in particular in US and Europe, but this should not lead us to forget that LHCONE is a global framework The WLCG has encouraged us to look a at longer-term perspective rather than rush in implementation The large experiment data flows will continue and alternatives to managing such flows are needed LHC (short-term) time scale: –2012: LHC run will continue until November – : LHC shutdown, restart late 2014 –2015: LHC data taking at full nominal energy (14 TeV) 6

LHCONE activities With all the above in mind, the Amsterdam Architecture workshop (Dec. 2011) has defined 5 activities: 1.VRF-based multipoint service: a “quick-fix” to provide the multipoint LHCONE connectivity as needed in places 2.Layer 2 multipath: evaluate use of emerging standards like TRILL (IETF) or Shortest Path Bridging (SPB, IEEE 802.1aq) in WAN environment 3.Openflow: There was wide agreement at the workshop that Software Defined Networking is the probable candidate technology for the LHCONE in the long-term, however needs more investigations 4.Point-to-point dynamic circuits pilot: build on capabilities present or demonstrated in several R&E networks to create a service pilot in LHCONE 5.Diagnostic Infrastructure: each site to have the ability to perform end-to-end performance tests with all other LHCONE sites 7

MULTIPOINT SERVICE The VRF Implementation 8

VRF: Virtual Routing and Forwarding VRF: in basic form, implementation of multiple logical router instances inside a physical device Logical separation of network control plane between multiple clients VRF approach in LHCONE: regional networks implement VRF domains to logically separate LHCONE from other flows BGP peerings used inter-domain and to the end-sites Some potential for Traffic Engineering –although scalability is a concern BGP communities defined allowing sites to define path preferences 9

Multipoint Service: VRF Implementation 10 WIX MANLAN NETHERLIGHT STARLIGHT CERNLIGHT

Multipoint, with regionals 11

Exchange Point Example: MANLAN 12

Connecting Sites to the VRF 13

14 ESnet USA Chicago New York Amsterdam BNL-T1 Internet2 USA Harvard CANARIE Canada UVic SimFraU TRIUMF-T1 UAlb UTor McGilU Seattle TWAREN Taiwan NCU NTU ASGC Taiwan ASGC-T1 KERONET2 Korea KNU LHCONE VPN domain End sites – LHC Tier 2 or Tier 3 unless indicated as Tier 1 Regional R&E communication nexus Data communication links, 10, 20, and 30 Gb/s See for details. NTU Chicago LHCONE: A global infrastructure for the LHC Tier1 data center – Tier 2 analysis center connectivity NORDUnet Nordic NDGF-T1a NDGF-T1c DFN Germany DESY GSI DE-KIT-T1 GÉANT Europe GARR Italy INFN-Nap CNAF-T1 RedIRIS Spain PIC-T1 SARA Netherlands NIKHEF-T1 RENATER France GRIF-IN2P3 Washington CUDI Mexico UNAM CC-IN2P3-T1 Sub-IN2P3 CEA CERN Geneva CERN-T1 SLAC GLakes NE MidW SoW Geneva KISTI Korea TIFR India Korea FNAL-T1 MIT Caltech UFlorida UNeb PurU UCSD UWisc Bill Johnston ESNet

SOFTWARE DEFINED NETWORKING 15

Software Defined Networking SDN Paradigm - Network control by applications; provide an API to externally define network functionality 16 Packet Forwarding Hardware “Network Operating System” App Packet Forwarding Hardware Packet Forwarding Hardware Packet Forwarding Hardware... API SNMP, CLI, … OpenFlow

Standardized SDN protocol –Open Networking Foundation ( Let external controller access/modify flow tables Allows separation of control plane and data forwarding Simple protocol, large application space –Forwarding, access control, filtering, topology segmentation, load balancing, … Distributed or centralized Reactive or pro-active 17 OpenFlow Switch Controller (PC) OpenFlow Switch Controller (PC) OpenFlow Switch Controller (PC) OpenFlow Switch OpenFlow Switch Controller (PC) OpenFlow Switch Controller (PC) Flow Tables App OpenFlow Switch MAC src MAC dst … ACTION OpenFlow Protocol

DYNAMIC POINT-TO-POINT SERVICE PILOT 18

Dynamic Bandwidth Allocation Will be one of the services to be provided in LHCONE Allows to allocate network capacity on as-needed basis –Instantaneous (“Bandwidth on Demand”), or –Scheduled allocation Significant effort in R&E Networking community –Standardisation through OGF (OGF-NSI, OGF-NML) Dynamic Circuit Service is present in several advanced R&E networks –SURFnet (DRAC) –ESnet (OSCARS) –Internet2 (ION) –US LHCNet (OSCARS) Planned (or in experimental deployment) –E.g. JGN (Japan), GEANT (AutoBahn), RNP (OSCARS/DCN), … DYNES: NSF funded project to extend hybrid & dynamic network capabilities to campus & regional networks –In deployment phase; fully operational in

Dynamic Circuits: On-demand Point-to-Point Layer-2 Paths 20 Bandwidth requested by “User” Agent (application or GUI): -Scheduled -On-demand Bandwidth requested by “User” Agent (application or GUI): -Scheduled -On-demand 2009 Example: CERN-Caltech using the OSCARS reservation system developed by ESNet

Aiming at definition of a Connection Service in a technology agnostic way Network Service Agent (NSA) High-level protocol OGF NSI Framework 21 Jerry Sobieski, NORDUnet

GLIF Open Lightpath Exchanges 22 Exchange Points operated by the Research and Education Network community GOLE Example: Netherlight in Amsterdam Automated GOLE project: fabric of GOLEs for development, testing and demonstration of dynamic network services.

NSI + AutoGOLE demonstration 23 Jerry Sobieski, NORDUnet Software Implementations OpenNSA – NORDUnet (DK/SE/) OpenDRAC – SURFnet (NL) G-LAMBDA-A - AIST (JP) G-LAMBDA-K – KDDI Labs (JP) AutoBAHN – GEANT (EU) DynamicKL – KISTI (KR) OSCARS* – ESnet (US)

The Case for Dynamic Provisioning in LHC Data Processing Data models do not require full-rate all times On-demand data movement will augment and partially replace static pre-placement  Network utilisation will be more dynamic and less predictable Performance expectations will not decrease –More dependence on the network, for the whole data processing system to work well! Need to move large data sets fast between computing sites –On-demand: caching –Scheduled: pre-placement –Transfer latency important for workflow efficiency As data volumes grow, and experiments rely increasingly on the network performance - what will be needed in the future is –More efficient use of network resources –Systems approach including end-site resources and software stacks Note: Solutions for th e LHC community need global reach 24

Dynamic Circuits Related R&D Projects in HEP: StorNet, ESCPS StorNet – BNL, LBNL, UMICH –Integrated Dynamic Storage and Network Resource Provisioning and Management for Automated Data Transfers ESCPS – FNAL, BNL, Delaware –End Site Control Plane System 25 Building on previous developments in and experience from the TeraPaths and LambdaStation projects StorNet: Integration of TeraPaths and BeStMan

MULTIPATH The Layer 2 challenge 26

The Multipath Challenge High throughput favours operation at lower network layers We can do multipath at Layer 3 (ECMP, MEDs, …) We can do multipath at Layer 1 (SONET/VCAT) –But only point-to-point! We cannot (currently) do multipath at Layer 2 –Loop prevention mandates tree topology – inefficient and inflexible 27

Multipath in LHCONE For LHCONE, in practical terms: –How to use the many transatlantic paths at Layer 2? –USLHCNet, ACE, GEANT, SURFnet, NORDUnet, … Some approaches to Layer 2 multipath: –IETF: TRILL (TRansparent Interconnect of Lots of Links) –IEEE: 802.1aq (Shortest Path Bridging) None of those designed for WAN! –Some R&D needed – e.g. potential use case for OpenFlow? 28

MONITORING AND DIAGNOSTICS Briefly… 29

Monitoring Monitoring and diagnostic services will be crucial for distributed global operation of the LHCONE data services –For the network operators –For the experiments and their operations teams –For the end-site administrators Two threads: –Monitoring of LHCONE “onset”: verify and compare before and after –Provide infrastructure for performance monitoring and troubleshooting in LHCONE operation Both will be based on PerfSONAR-PS –US Atlas playing an important front-role See e.g. Shawn’s presentation at the recent WLCG GDB meeting: – slides&confId=155067http://indico.cern.ch/getFile.py/access?contribId=6&resId=0&materialId= slides&confId=

LHCONE Layer 1 connectivity (Bill Johnston, April 2012) 31

Summary LHCONE is currently putting in place the Multipoint and the Monitoring services –Some sites are already using LHCONE data paths –Be aware of the pre-production nature of the current infrastructure –Coordination from the Experiments’ side is welcome/necessary Point-to-point pilot is in preparation –Building on a solid foundation, leveraging R&D investment in major networks and collaborations –Future OGF NSI, NMC standards; current DICE IDCP de-facto standard –Participation of sites interested in advanced network services is crucial R&D activities defined and under way: –Multipath (TRILL/SPB) –OpenFlow Converge on 2 year time scale (by end of LS1) Next face-to-face meeting: Stockholm, May 3&4, 2012 – 32

THANK YOU! 33