Dr. Silvio Pardi INFN-Napoli Belle II Italia 21 December 2015

Slides:



Advertisements
Similar presentations
Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
Advertisements

Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
LCG Service Challenge Phase 4: Piano di attività e impatto sulla infrastruttura di rete 1 Service Challenge Phase 4: Piano di attività e impatto sulla.
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 25 th April 2012.
Grid Lab About the need of 3 Tier storage 5/22/121CHEP 2012, The need of 3 Tier storage Dmitri Ozerov Patrick Fuhrmann CHEP 2012, NYC, May 22, 2012 Grid.
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Belle II Computing Fabrizio Bianchi INFN and University of Torino Meeting Belle2 Italia 17/12/2014.
HEPiX IPv6 Working Group David Kelsey (STFC-RAL) GridPP33 Ambleside 22 Aug 2014.
Grid Operations in Germany T1-T2 workshop 2015 Torino, Italy Kilian Schwarz WooJin Park Christopher Jung.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
EGEE Data Management Services
Federating Data in the ALICE Experiment
a brief summary for users
Dynamic Extension of the INFN Tier-1 on external resources
WLCG IPv6 deployment strategy
on behalf of the Belle II Computing Group
DPM at ATLAS sites and testbeds in Italy
WLCG Network Discussion
LCG Service Challenge: Planning and Milestones
Virtualization and Clouds ATLAS position
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
Status Report on LHC_2 : ATLAS computing
StoRM: a SRM solution for disk based storage systems
Status of the SRM 2.2 MoU extension
Overview of the Belle II computing
Belle II Physics Analysis Center at TIFR
Computing activities in Napoli
BNL Tier1 Report Worker nodes Tier 1: added 88 Dell R430 nodes
James Casey, IT-GD, CERN CERN, 5th September 2005
Dynafed, DPM and EGI DPM workshop 2016 Speaker: Fabrizio Furano
Plans to support IPv6-only CPU on WLCG
Experiences with http/WebDAV protocols for data access in high throughput computing
Status and Plans on GRID related activities at KEK
Data Challenge with the Grid in ATLAS
Storage Interfaces and Access: Introduction
dCache – protocol developments and plans
INFN-GRID Workshop Bari, October, 26, 2004
StoRM Architecture and Daemons
Gfal/lcg-util -> Gfal2/gfal2-util
Future challenges for the BELLE II experiment
Bernd Panzer-Steindel, CERN/IT
Taming the protocol zoo
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
ATLAS Sites Jamboree, CERN January, 2017
Project Status Report Computing Resource Review Board Ian Bird
EGI UMD Storage Software Repository (Mostly former EMI Software)
Simulation use cases for T2 in ALICE
Ákos Frohner EGEE'08 September 2008
An introduction to the ATLAS Computing Model Alessandro De Salvo
The INFN Tier-1 Storage Implementation
WLCG and support for IPv6-only CPU
Data Management cluster summary
Grid Canada Testbed using HEP applications
R. Graciani for LHCb Mumbay, Feb 2006
LHC Data Analysis using a worldwide computing grid
 YongPyong-High Jan We appreciate that you give an opportunity to have this talk. Our Belle II computing group would like to report on.
Computing activities in Napoli
IPv6 update Duncan Rand Imperial College London
Dirk Duellmann ~~~ WLCG Management Board, 27th July 2010
The LHCb Computing Data Challenge DC06
Presentation transcript:

Dr. Silvio Pardi INFN-Napoli Belle II Italia 21 December 2015 Network & Storage Dr. Silvio Pardi INFN-Napoli Belle II Italia 21 December 2015

Countries Involved Currently there are 18 Countries involved in the main data movement activities and we have 23 Storage Endpoints. (Contributing Number of Countries per region) 6 3 3 1 4 1 Australia Austria Canada China Czech Rep. Germany India Italy Japan Korea Mexico Poland Russia Slovenia Taiwan Turkey Ukraine USA

Network Estimation 2016-2018 Belle II computing model has to manage several data flows, that can be classified in: RAW DATA MDST from Data Processing MDST from Data Reprocessing Monte Carlo Data User Analysis Each data flow has a different data distribution pattern and involves different sites.

Effect of the new RAW data distribution strategy RAW Data Strategy 2021-2024 2016-2020 Max 15Gbps Max of 8 Gbps (PNNL) Effect of the new RAW data distribution strategy

MDST Expected Data Flows Grey line 2021-2024 2016-2020 IN-BOUND OUT-BOUND

Monte Carlo Production (Data Challenge) Strategy under investigation: Each MC-Site will replicate the produced data, to 2 sites of different regions: the goal is to have 3 copies of the whole MC production. Site% Site% Site% Site% Site Fraction % KEK 22,31% PNNL 14,46% Korea 4,94% Germany 13,01% Italy 12,43% Slovenia 2,33% Australia 2,91% Canada Austria 1,74% China 2,62% Czech Rep. 1,16% India 3,20% Mexico Poland 2,03% Russia 8,14% Taiwan Turkey 0,58% Ukraine MC Site1 MC Site2 MC Site3 MC Site4 MC% MC% MC Site5 MC Site6 MC Site7 MC Site8 Site% Site% Site% Site%

Network usage during the MC5 Campaign

Distributed User Analysis Xrootd dash WLCG-FTS dash Average throughput since 2013 3GB/s vs 571MB/s Comparison between WLC-FTS dash and the xrootd dash of the CMS experiment 6GB/s Peak of 1.5GB/s Form a first and fast analysis seem that the whole xrootd activity (that we can mainly attribute at the distributed analysis) introduce an additional 15% of traffic to respect FTS data movement. In this work this value as been used as “rule of thumb” to estimate User Analysis incidence on the whole traffic. A more fine estimation will be complete after 2016

Peaks: 6.5 Gbps KEK, 3Gbps PNNL, 2Gbps Germany and Italy TOTAL IN-BOUND for 2016-2018 Peaks :8Gbps PNNL 6.5 Gbps KEK TOTAL OUT-BOUND for 2016-2018 Peaks: 6.5 Gbps KEK, 3Gbps PNNL, 2Gbps Germany and Italy

SINET4

SINET5 - 2016

Belle II worldwide network paths 20G 100G AMS 40G LON 10G 2x100G 100G PAR BOS GEN 100G NYC 100G MAD LAX WDC TOK 10G MUM 10G 10G SG Primary route Backup route

Data Challenge 2015: GRID Sites & FTS

Data Challenge 2015 Summary (I Phase) 2018 2024 2015 Data Challenge Results Comment KEK➔PNNL 4 Gbps > 12 Gbps average ~4Gbps Spikes at 9 Gbps Result from gsiftp Network tuning required KEK ➔ SIGNET 0.5Gbps 1 Gbps average 0.8 Gbps (high success rate) 2.4 Gbps (low success rate) KEK ➔ NAPOLI 1 Gbps * 5 Gbps/2* 3 Gbps 3 Gbps with appropriate load. Higher bandwidths may be possible KEK ➔ KIT 3.5 Gbps KEK ➔ DESY 3.0 Gbps * Equal site splitting in country

Data Challenge 2015 Phase II: Summary PNNL->CNAF First test ~700MB/s Test with CNAF Still ongoing V. Bansal - Belle-2 Network Meeting @ SC15 “Results and plans for Belle II Networking and Data Challenges”

Storage

Current Storage Elements 30 Storage currently working Backend Type: 5 DISET 2 xroot 10 DPM 6 dCache 6 StoRM 1 bestman2 Reserved disk space for BELLE: 1.9 PB ~30% used Capacity for storage Type 23 SRM

Ongoing certification procedure For each SE we are testing: The presence of BELLE Space Token The presence of the following directory structure and ACL settings /belle/ (Role=Null R-X, Role=production/lcgadmin RWX) /belle/DATA (Role=Null R-X, Role=production/lcgadmin RWX) /belle/TMP (Role=Null RWX, Role=production/lcgadmin RWX) We open a TEAM Ticket on GGUS for each site affected by misconfigurations

HTTP Deployment Task Force All the pieces are now in place for HEP exploitation of the HTTP protocol. Atlas and LHCb (at least) are developing their HTTP infrastructures. The majority of storage systems are able to offer an HTTP interface. ROOT support is available via davix, and studies have shown that this can allow HTTP to reach a performance comparable with that of the xroot protocol. At the moment there are many site misconfigurations and instabilities which could be tracked and fixed at the WLCG level for all experiments 07/12/2015 DPM Workshop 2015

GridFTP with redirection Client Head node Disk node DB FS GridFTP DMLite SRM GridFTP (control) GridFTP (data) (only if you really need to) Globus IPC RFIO It is fully supported by FTS3 and gfal2-util CLI It is seen as the most useful way of deploying GridFTP with DMLite-based storage It renders SRM (along with its overhead) unnecessary for most typical use cases GridFTP status update 7/12/2015

Http Testing infrastructure for Belle II Sta per iniziare una sperimentazione per l’utilizzo di http in belle II. Verificare che il software di belle II sia in grado di utilizzare http o individuare le modifiche necessarie. Selezionare un set di Storage Element configurati con Http che possano essere monitorati dalla Http-TaskForce. Gli storage saranno inseriti in una federazione http di test. A breve invierò una call for interest per collezionare I siti che sono interessati a partecipare alla prima sperimentazione.

Thank you

Backup

Belle II User analysis strategy in testing Options for Physics skims from bulk MDST Output Physics Object + MDST (μDST) Direct copy to WN using DIRAC Stream data using XRootD Output Index file which point to skimmed events Access bulk MDST directly via root on large cluster Stream from bulk MDST using XRootD Local Cluster/Workstation root The set of possible options for user analysis currently under investigation, includes xrootd-based strategies

Belle II analysis model with Index file Data is analyzed on remote WN. Only selected Events transferred. Ongoing tests in order to test the whole chain, the gbasf2 support and performances with xrootd. Storage considered for the ongoing tests Storage Type b2se.mel.coepp.org.au DPM coepp-dpm-01.ersa.edu.au belle-dpm-01.na.infn.it dcache.ijs.si dCache KEK StoRM

Network usage during the MC5 Campaign In the ongoing MC5 production we are managing about 50K transfer jobs for day, involving an increasing number of sites.

Network usage during the MC5 Campaign

perfSONAR dashboard

perfSONAR dashboard

FTS3 Dashboard

RAW Data Involved sites RAW data are produced at KEK and replicated at PNNL After 2020 the second RAW data copy will be distributed in additional sites: Italy, Germany, Canada, Korea, and India. 100% 100% KEK Until 2020 100% PNNL 30% AFTER 2020 ITALY GERMANY CANADA KOREA INDIA

MDST from Data Processing 2016-2018 MDST will be produced during data taking ad KEK and PNNL and then moved to other countries with the goal to have 3 full copy of processed data (Asia, USA, EU). (network transfer) After 2020: Italy, Germany, Canada, India and Korea will participate at the Raw Data processing. 100% 100% 40% 60% 40% 60% KEK 40% PNNL 60% ITALY GERMANY SLOVENIA CANADA AUSTRALIA

Reprocessed Data 2016-2018 Data reprocessing will be performed at PNNL (100%) and then replicated in other countries in order to have 3 full copy. 100% 100% KEK 100% PNNL ITALY GERMANY SLOVENIA CANADA AUSTRALIA

Monte Carlo RAW DATA 2016-2018 Two copies of RAW Monte Carlo Data, one in KEK and one in PNNL 100% 100% MC% MC% MC% MC% KEK MC% PNNL MC% MC% MC% MC Site1 MC Site2 MC Site3 MC Site4 MC Site5

the goal is to have 3 full copies of processed MC. Monte Carlo Analysis 2016-2018 MC% MC% MC% MC% MC Site1 MC Site2 MC Site3 MC Site4 100% MC% MC% MC% MC% MC% KEK MC% PNNL the goal is to have 3 full copies of processed MC. MC% ITALY GERMANY SLOVENIA CANADA AUSTRALIA