Connect communicate collaborate LHCONE – Linking Tier 1 & Tier 2 Sites Background and Requirements Richard Hughes-Jones DANTE Delivery of Advanced Network.

Slides:



Advertisements
Similar presentations
An Improved TCP for transaction communications on Sensor Networks Tao Yu Tsinghua University 2/8/
Advertisements

Circuit Monitoring July 16 th 2011, OGF 32: NMC-WG Jason Zurawski, Internet2 Research Liaison.
0 - 0.
1 ALICE Grid Status David Evans The University of Birmingham GridPP 14 th Collaboration Meeting Birmingham 6-7 Sept 2005.
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
1 ALICE Grid Status David Evans The University of Birmingham GridPP 16 th Collaboration Meeting QMUL June 2006.
The ATLAS Computing Model Roger Jones Lancaster University CHEP06 Mumbai 13 Feb
1 DTI/EPSRC 7 th June 2005 Reacting to HCI Devices: Initial Work Using Resource Ontologies with RAVE Dr. Ian Grimstead Richard Potter BSc(Hons)
Bernd Panzer-Steindel, CERN/IT WAN RAW/ESD Data Distribution for LHC.
1 Jin Li Microsoft Research. Outline The Upcoming Video Tidal Wave Internet Infrastructure: Data Center/CDN/P2P P2P in Microsoft Locality aware P2P Conclusions.
The GATE-LAB system Sorina Camarasu-Pop, Pierre Gueth, Tristan Glatard, Rafael Silva, David Sarrut VIP Workshop December 2012.
MB - NG MB-NG Technical Meeting 03 May 02 R. Hughes-Jones Manchester 1 Task2 Traffic Generation and Measurement Definitions Pass-1.
© Paradigm Publishing Inc Chapter 10 Information Systems.
Week 1.
KIT – University of the State of Baden-Wuerttemberg and National Research Center of the Helmholtz Association Steinbuch Centre for Computing (SCC)
Advanced 5G Network Infrastructure for the Future Internet From IoT to U-HDTV, ubiquity Restless Pressure on bandwidth, spectrum crunch Complex traffic.
Trial of the Infinera PXM Guy Roberts, Mian Usman.
GridPP meeting Feb 03 R. Hughes-Jones Manchester WP7 Networking Richard Hughes-Jones.
Tier 2 Prague Institute of Physics AS CR Status and Outlook J. Chudoba, M. Elias, L. Fiala, J. Horky, T. Kouba, J. Kundrat, M. Lokajicek, J. Svec, P. Tylka.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Tier-2 Network Requirements Kors Bos LHC OPN Meeting CERN, October 7-8,
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
1 ESnet Network Measurements ESCC Feb Joe Metzger
Experience with the WLCG Computing Grid 10 June 2010 Ian Fisk.
Take on messages from Lecture 1 LHC Computing has been well sized to handle the production and analysis needs of LHC (very high data rates and throughputs)
Wenjing Wu Andrej Filipčič David Cameron Eric Lancon Claire Adam Bourdarios & others.
PPDG and ATLAS Particle Physics Data Grid Ed May - ANL ATLAS Software Week LBNL May 12, 2000.
Your university or experiment logo here Caitriana Nicholson University of Glasgow Dynamic Data Replication in LCG 2008.
LHC Open Network Environment LHCONE Artur Barczyk California Institute of Technology LISHEP Workshop on the LHC Rio de Janeiro, July 9 th,
LHC Open Network Environment LHCONE David Foster CERN IT LCG OB 30th September
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
ATLAS Grid Data Processing: system evolution and scalability D Golubkov, B Kersevan, A Klimentov, A Minaenko, P Nevski, A Vaniachine and R Walker for the.
Online-Offsite Connectivity Experiments Catalin Meirosu *, Richard Hughes-Jones ** * CERN and Politehnica University of Bucuresti ** University of Manchester.
From the Transatlantic Networking Workshop to the DAM Jamboree to the LHCOPN Meeting (Geneva-Amsterdam-Barcelona) David Foster CERN-IT.
NORDUnet Nordic Infrastructure for Research & Education Workshop Introduction - Finding the Match Lars Fischer LHCONE Workshop CERN, December 2012.
Content: India’s e-infrastructure an overview The Regional component of the Worldwide LHC Computing Grid (WLCG ) India-CMS and India-ALICE Tier-2 site.
Connect communicate collaborate LHCONE Diagnostic & Monitoring Infrastructure Richard Hughes-Jones DANTE Delivery of Advanced Network Technology to Europe.
Point-to-point Architecture topics for discussion Remote I/O as a data access scenario Remote I/O is a scenario that, for the first time, puts the WAN.
LHC Computing, CERN, & Federated Identities
The ATLAS Computing Model and USATLAS Tier-2/Tier-3 Meeting Shawn McKee University of Michigan Joint Techs, FNAL July 16 th, 2007.
The Network & ATLAS Workshop on transatlantic networking panel discussion CERN, June Kors Bos, CERN, Geneva & NIKHEF, Amsterdam ( ATLAS Computing.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Data Management Highlights in TSA3.3 Services for HEP Fernando Barreiro Megino,
Ian Bird WLCG Networking workshop CERN, 10 th February February 2014
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Connect communicate collaborate LHCONE European design & implementation Roberto Sabatino, DANTE LHCONE Meeting, Washington, June
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Strawman LHCONE Point to Point Experiment Plan LHCONE meeting Paris, June 17-18, 2013.
ATLAS Distributed Computing ATLAS session WLCG pre-CHEP Workshop New York May 19-20, 2012 Alexei Klimentov Stephane Jezequel Ikuo Ueda For ATLAS Distributed.
WLCG Status Report Ian Bird Austrian Tier 2 Workshop 22 nd June, 2010.
From the Transatlantic Networking Workshop to the DAM Jamboree David Foster CERN-IT.
1 June 11/Ian Fisk CMS Model and the Network Ian Fisk.
LHCONE Workshop Richard P Mount February 10, 2014 Concerns from Experiments ATLAS Richard P Mount SLAC National Accelerator Laboratory.
ATLAS Computing: Experience from first data processing and analysis Workshop TYL’10.
LHC collisions rate: Hz New PHYSICS rate: Hz Event selection: 1 in 10,000,000,000,000 Signal/Noise: Raw Data volumes produced.
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
Bob Jones EGEE Technical Director
Dynamic Extension of the INFN Tier-1 on external resources
Status of WLCG FCPPL project
WLCG Network Discussion
Ian Bird WLCG Workshop San Francisco, 8th October 2016
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
Computing models, facilities, distributed computing
Networking between China and Europe
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Project Status Report Computing Resource Review Board Ian Bird
K. Schauerhammer, K. Ullmann (DFN)
John Gordon, STFC GDB April 6th 2011
ExaO: Software Defined Data Distribution for Exascale Sciences
an overlay network with added resources
Development of LHCb Computing Model F Harris
Presentation transcript:

connect communicate collaborate LHCONE – Linking Tier 1 & Tier 2 Sites Background and Requirements Richard Hughes-Jones DANTE Delivery of Advanced Network Technology to Europe LHCONE Planning Meeting, RENATER Paris, 5 April 2011

connect communicate collaborate 2 Introduction: Describe some of the changes in the computing model of the LHC experiments. Demonstrate the importance and usage of the network. Show the relation between LHCONE and LHCOPN. Bring together and present the user requirements for future LHC physics analysis. Provide the information to facilitate the presentations on the Architecture and the Implementation of LHCONE.

connect communicate collaborate 3 A Little History Requirements paper from K. Bos (Atlas) and I. Fisk (CMS) in autumn Experiments had devised new compute and data models for LHC data evaluation basically assuming a high speed network connecting the T2s worldwide. Ideas & proposals were discussed at a workshop held at CERN in Jan Gave input from the networking community. An "LHCONE Architecture" doc finalised in Lyon in Feb Here K. Bos proposed to start with a prototype based on the commonly agreed architecture. K. Bos and I. Fisk produced a "Use Case" note with list of sites for the prototype. In Rome late Feb 2011 some NRENs & DANTE formed ideas for the " LHCONE prototype planning " doc.

connect communicate collaborate LHCOPN LHC: Changing Data Models (1) LHC computing model based on MONARC served well > 10 years ATLAS strictly hierarchal; CMS less so. The successful operation of the LHC accelerator & start of data analysis, brought a re-evaluation of the computing and data models. Flatter hierarchy: Any site might in the future pull data from any other site hosting it. LHCOPN 4 Artur Barczyk

connect communicate collaborate LHC: Changing Data Models (2) Data caching: A bit like web caching. Analysis sites will pull datasets from other sites on demand, including from Tier2s in other regions, then make it available for others. Possible strategic pre-placement of data sets Datasets put close to physicists studying that data / suitable CPU power. Use of continental replicas. Remote data access: jobs executing locally, using data cached at a remote site in quasi-real time. Traffic patterns are changing – more direct inter-country data transfers 5

connect communicate collaborate ATLAS Data Transfers Between all Tier levels Average: ~ 2.3 GB/s (daily average) Peak: ~ 7 GB/s (daily average) Data available on site within a few hours. 70 Gbit/s on LHCOPN ATLAS reprocessing Daniele Bonacorsi 6

connect communicate collaborate Data Flow EU – US ATLAS Tier 2s Example above is from US Tier 2 sites Example above is from US Tier 2 sites Exponential rise in April and May, after LHC start Changed data distribution model end of June – caching ESD and DESD Much slower rise since July, even as luminosity grows rapidly Kors Bos 7

connect communicate collaborate LHC: Evolving Traffic Patterns One example of data coming from the US 4 Gbit/s for ~ 1.5 days (11 Jan 11) Transatlantic link GÉANT Backbone NREN Access Link Not an isolated case Often made up of many data flows Users getting good at running gridftp 8

connect communicate collaborate Data Transfers over RENATER Peak rates a substantial fraction of 10 Gigabits, often for hours. Several LHC involved. Demand variable depending on user work. Francois-Xavier Andreu 9

connect communicate collaborate Data Transfers over DFN Peak rates saturate one of the 10 Gigabit links DFN-GÉANT. Demand variable depending on user work. Christian Grimm 10 Two different weeks from GÉANT to Aachen

connect communicate collaborate Data Transfers from GARR - CNAF T0-T1 + T1-T1 + T1-T2 Peak rates Gigabit/s. Traffic shows diurnal demand & is variable depending on user work. Sustained growth over last year Marco Marletta 11

connect communicate collaborate CMS Data Transfers Data Placement for Physics Analysis Once data is onto the WLCG, it must be made accessible to analysis applications. Largest fraction of analysis computing at LHC is at the Tier2s. New flexibility reduces latency for end users. Daniele Bonacorsi 12 T1 T2 dominates T2 T2 emerges

connect communicate collaborate Data Transfer Performance Site or Network? Test NorthGrid to GÉANT PoP London UDP throughput from SE 990 Mbit/s. 75% packet loss. Data transmitted by SE at 3.8 Gbit/s over 4 1 Gigabit interfaces. TCP transmits in bursts at 3.8 Gbit/s packet loss & re-tries mean low throughput 13 1 Gbit Bottleneck at receiver Classic packet loss from bottleneck Even more data with end-hosts fixed.

connect communicate collaborate LHCOPN linking Tier 0 to Tier 1s LHCONE for Tier 1s and Tier 2s 14 LHCONE Other regions T2s in a country LHCONE prototype in Europe. T1 are connected but not LHCOPN

connect communicate collaborate Requirements for LHCONE LHCOPN provides infrastructure to move data T0-T1 and T1-T1. New infrastructure required to improve transfers T1-T2 & T2-T2: Analysis is mainly done in Tier 2, so data is required from any T1 or any T2. T2-T2 is very important. Work done at a Tier 2: Simulations & Physics Analysis (50:50) Network BW needs of a T2 include: Re-processing efforts: 400 TByte refresh in a week = 5 Gbit/s Data bursts from user analysis : 25 Tbyte in a day = 2.5Gbit/s Feeding a 1000 core farm with LHC events: ~ 1Gbit/s Note this implies timely delivery of data not just average rates! Access link available bandwidth for Tier 2 sizes: Large 10 Gbit; Medium 5 Gbit; Small 1 Gbit 15

connect communicate collaborate Requirements for LHCONE Sites are free to choose the way they wish to connect. Flexibility & extensibility required: T2s change Analysis usage pattern is more chaotic – Dynamic Networks of interest World-wide connectivity required for LHC sites. There is concern about LHC traffic swamping other disciplines. Monitoring & fault-finding support should be built in. Cost effective solution required – may influence the Architecture. No isolation of sites must occur. No interruption of the data-taking or physics analysis A prototype is needed. 16

connect communicate collaborate Requirements Fitting in with LHC 2011 data taking 17 Machine development & Technical Stops provide pauses in the data taking. This does not mean there is plenty of time. LHCONE prototype might grow in phases.

connect communicate collaborate ANY QUESTIONS ? 18