INFN Roma Grid Activities

Slides:



Advertisements
Similar presentations
Introduction to CMS computing CMS for summer students 7/7/09 Oliver Gutsche, Fermilab.
Advertisements

ATLAS Tier-3 in Geneva Szymon Gadomski, Uni GE at CSCS, November 2009 S. Gadomski, ”ATLAS T3 in Geneva", CSCS meeting, Nov 091 the Geneva ATLAS Tier-3.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
A tool to enable CMS Distributed Analysis
Les Les Robertson WLCG Project Leader WLCG – Worldwide LHC Computing Grid Where we are now & the Challenges of Real Data CHEP 2007 Victoria BC 3 September.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
A short introduction to the Worldwide LHC Computing Grid Maarten Litmaath (CERN)
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
ESFRI & e-Infrastructure Collaborations, EGEE’09 Krzysztof Wrona September 21 st, 2009 European XFEL.
SC4 Planning Planning for the Initial LCG Service September 2005.
The ATLAS Cloud Model Simone Campana. LCG sites and ATLAS sites LCG counts almost 200 sites. –Almost all of them support the ATLAS VO. –The ATLAS production.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks CRAB: the CMS tool to allow data analysis.
6 march Building the INFN Grid Proposal outline a.ghiselli,l.luminari,m.sgaravatto,c.vistoli INFN Grid meeting, milano.
INFSO-RI Enabling Grids for E-sciencE CRAB: a tool for CMS distributed analysis in grid environment Federica Fanzago INFN PADOVA.
INFSO-RI Enabling Grids for E-sciencE The EGEE Project Owen Appleton EGEE Dissemination Officer CERN, Switzerland Danish Grid Forum.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Distributed Analysis Tutorial Dietrich Liko. Overview  Three grid flavors in ATLAS EGEE OSG Nordugrid  Distributed Analysis Activities GANGA/LCG PANDA/OSG.
GDB, 07/06/06 CMS Centre Roles à CMS data hierarchy: n RAW (1.5/2MB) -> RECO (0.2/0.4MB) -> AOD (50kB)-> TAG à Tier-0 role: n First-pass.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
ATLAS Distributed Analysis DISTRIBUTED ANALYSIS JOBS WITH THE ATLAS PRODUCTION SYSTEM S. González D. Liko
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
1 S. JEZEQUEL- First chinese-french workshop 13 December 2006 Grid: An LHC user point of vue S. Jézéquel (LAPP-CNRS/Université de Savoie)
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
1-2 March 2006 P. Capiluppi INFN Tier1 for the LHC Experiments: ALICE, ATLAS, CMS, LHCb.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
Claudio Grandi INFN Bologna Workshop congiunto CCR e INFNGrid 13 maggio 2009 Le strategie per l’analisi nell’esperimento CMS Claudio Grandi (INFN Bologna)
Daniele Bonacorsi Andrea Sciabà
CernVM-FS vs Dataset Sharing
Workload Management Workpackage
Status: ATLAS Grid Computing
WLCG Tier-2 Asia Workshop TIFR, Mumbai 1-3 December 2006
Grid Computing: Running your Jobs around the World
(Prague, March 2009) Andrey Y Shevel
The EDG Testbed Deployment Details
U.S. ATLAS Tier 2 Computing Center
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
Moving the LHCb Monte Carlo production system to the GRID
Pasquale Migliozzi INFN Napoli
INFN Computing infrastructure - Workload management at the Tier-1
GWE Core Grid Wizard Enterprise (
Jan 12, 2005 Improving CMS data transfers among its distributed Computing Facilities N. Magini CERN IT-ES-VOS, Geneva, Switzerland J. Flix Port d'Informació.
Data Challenge with the Grid in ATLAS
INFN-GRID Workshop Bari, October, 26, 2004
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Readiness of ATLAS Computing - A personal view
ATLAS DC2 ISGC-2005 Taipei 27th April 2005
An introduction to the ATLAS Computing Model Alessandro De Salvo
LHC Data Analysis using a worldwide computing grid
CC and LQCD dimanche 13 janvier 2019dimanche 13 janvier 2019
Future EU Grid Projects
Gridifying the LHCb Monte Carlo production system
ATLAS DC2 & Continuous production
Short to middle term GRID deployment plan for LHCb
The LHCb Computing Data Challenge DC06
Presentation transcript:

INFN Roma Grid Activities Alessandro De Salvo Alessandro.DeSalvo@roma1.infn.it 7/5/2009 Outline Introduction Grid computing at INFN Roma The Tier2 Computing Center Activities and perspectives Conclusions

Alessandro De Salvo – Grid Activities in INFN Roma What is a grid? Relation to WWW? Uniform easy access to shared information Relation to distributed computing? Local clusters WAN (super)clusters Condor Relation to distributed file systems? NFS, AFS, GPFS, Lustre, PanFS… A grid gives selected user communities uniform access to distributed resources with independent administrations Computing, data storage, devices, … 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Alessandro De Salvo – Grid Activities in INFN Roma Why is it called grid? Analogy to power grid You do not need to know where your electricity comes from Just plug in your devices You should not need to know where your computing is done Just plug into the grid for your computing needs You should not need to know where your data is stored Just plug into the grid for your storage needs 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Where do we use the grids? Scientific collaborations High-energy physics Astrophysics Theoretical Physics Gravitational Waves Computational chemistry Biomed – biological and medical research Health-e-Child – linking pediatric centers WISDOM – “in silico” drug and vaccine discovery Earth sciences UNOSAT – satellite image analysis for the UN Can also serve in spreading know-how to developing countries Industry? Commerce? Research collaborations Intra-company grids Homes? Schools? E-learning 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Alessandro De Salvo – Grid Activities in INFN Roma Grids in the World The general INFN Roma computing infrastructure is part of the following Grids IGI (Italian Grid Infrastructure) WLCG (World LHC Computing Grid) EGEE (Enabling Grid for E-sciencE) The most important grid users in Roma are The ATLAS experiment at LHC The CMS experiment at LHC The VIRGO experiment The Theoretical Physics group 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Alessandro De Salvo – Grid Activities in INFN Roma Why Grid at LHC? 10-15 PetaByte/year Data analysis requires at least ~100k typical PC processors Scientists in tens of countries worldwide CERN can provide up to 20-30% of the storage and CPU 70-80% are provided by WLCG partners 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Alessandro De Salvo – Grid Activities in INFN Roma Why Grid for VIRGO? Gravitational wave detector Virgo has accumulated several months of data up to now (130 TB in total). The next scientific run, VSR2, will start in July 2009 and will last ~6 months, producing additional ~130 TB of data Some of the analysis to be carried on the data require large computational resources. In particular the ‘blind’ search for continuous gravitational signals from rotating neutron stars, an activity on which the Rome group has the leadership VIRGO uses the grid to transparently access the computing resources needed for the analysis. The minimal amount of computing power needed to analyze 6 months of data is ~1000 dedicated CPUs for 6 months. The bigger the available computing power, the larger is the portion of the parameter space that can be explored (computing bounded problem) 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Why Grid for Theoretical Physics? Study of the QCD properties, i.e. the theory of the quark-gluon interactions Mainly based on numerical simulations SU(3) Lattice QCD computations in 4 dimensions Computing requirements Different requirements on RAM, depending on the lattice volume: 124, 144, 184 <--> 250 MB, 500 MB, 1.1 GB respectively Trivial parallelism, using N copies of the same program with different initial random seeds (N=200 for the biggest lattice) Grid as a probability distribution generator Very small, optimized executable, written in C Using the SSE2 set of instructions, available in the most recent Intel CPUs Very limited amount of disk space needed for the output 2005-2006 (L.Giusti, S.Petrarca, B.Taglienti: POS Tucson 2006, PRD 76 094510 2007) 134800 configurations ≈ 800000 CPU hours ⇒ ~1 year @ 100 CPU dedicated farm Last 1.5 years 70 000 configurations (run) + 40000 configurations (test) ≈ 660000 CPU hours Only possible via Grid Computing! New lattice dimensions will require ≥ 2 GB of RAM and dedicated run sessions on more than 100 CPUs (or ~100 dedicated CPUs for 6 months) Challenge for the resources currently available in the grid Work to be presented at Lattice2009 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Organization of the Grid computing centers The Grid Sites are organized hierarchically Generally referred as “Tiers” The biggest sites, in terms of resources and support level, are at the top of the hierarchy (lower levels), while the smaller ones are at the bottom (higher levels) Example: the WLCG Tiered structure CERN (Tier-0) Regional (National) Centers (Tier-1s) Local Centers (Tier-2s) The sites located higher in the hierarchy are coordinating and supporting the sites below them Tier-2 sites WLCG Canada TRIUMF Tier-1 centers Taiwan ASGC NL SARA- NIKHEF USA BNL UK RAL Tier-0 CERN France CCIN2P3 Spain PIC Italy CNAF Nordic countries NDGF USA FNAL Germany FZK ROMA 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

The INFN Roma Tier2 center 7 cooling racks, expandable to 14 Self-conditioned, using water Remote Monitoring System Up to 40 units per rack With the current CPU form factor, it would be ~9000 CPU cores! Hosted experiments ATLAS (WLCG Tier2) 286 CPU cores LSF batch system ~200 TB storage space DPM Storage CMS (WLCG Tier2) 320 CPU cores 110 TB storage space dCache Storage VIRGO 416 CPU cores Torque/Maui batch system 10 TB storage space ATLAS 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Network connectivity for the Tier2 center GARR Roma1-Tier2 GARR Roma1 1 Gbps 1 Gbps INFN Roma Tier2 Center INFN Roma Network Internal Network (ATLAS) @ 10 Gbps 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Data processing using Grid tools ATLAS Data Transfer via the ATLAS DDM (DQ2) Throughput (MB/s) in ATLAS (04-05/2009) MonteCarlo Production PANDA (Production ANd Distributed Analysis) http://panda.cern.ch:25880/server/pandamon/query?dash=prod User Analysis PANDA GANGA (Gaudi / Athena aNd Grid Alliance) http://cern.ch/ganga Online Muon Calibration MC Production Jobs in ATLAS Italy (15-04-2009 / 04-05/2009) MC Production Walltime (seconds) in ATLAS Italy (15-04-2009 / 04-05/2009) 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Grid services in Roma: the ATLAS Software Installation System Core service for the ATLAS Software Installations in LCG/EGEE and OSG Grids Serving more than 200 sites in unattended or semi-automatic (assisted) mode Single instance running in Rome and serving all the ATLAS collaboration since 2003 Online view of the availability of the software deployment in all the LCG/EGEE (Europe) and OSG (Open Science Grid, US) sites On-demand installation, testing and management of the releases Based on the LJSFi (Light Job Submission Framework for Installation) engine, entirely developed in Rome for LCG in 2003, exanded to OSG n 2009 and planning to include NorduGrid (Nordic Countries) in the next months, unifying the Installation System for full WLCG ATLAS Collaboration Work to be done here! Designed for multi-Grid and multi-VO operations https://atlas-install.roma1.infn.it/atlas_install Presented at CHEP 2007 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

The ATLAS MDT calibration center Online Calibration of the Monitored Drift Tubes (MDT) of the ATLAS Muon Spectrometer Data coming from CERN via the ATLAS Distributed Data Management system (aka DQ2) must be processed in quasi-streaming mode to calculate the calibration constants in 24/48 hours 3 calibration centers (Michigan, Munich and Rome) share the calibration tasks in different regions of the detector The full calibration procedure is supervised in each center by an agent called Local Calibration Data Splitter (LCDS), fully integrated with the Grid and local tools needed The architecture design and the control tools are direct responsability of the INFN ATLAS Roma group ATLAS Tier 0 ATLAS DDM Michigan Tier-2 ( UltraLight Network) Munich Tier-2 Roma Tier-2 Local SE Local Calibration Data Splitter Calibration Farm Local Oracle CalibDB COOL DB Local Calibration Database (Oracle) LCDS engine LRMS interface LSF PBS SGE Condor (SQLite3) ATLAS DDM Storage interface File RFIO DCAP GSIDCAP GSIFTP SRM Calibration Farm Storage System Authorization System ByteStream decoder Web UI The LCDS Architecture Presented at CHEP 2009 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Data processing using Grid tools CMS MonteCarlo Production User Analysis CRAB (CMS Remote Analysis Builder) https://twiki.cern.ch/twiki/bin/view/CMS/SWGuideCrab Offline Calibration Data Management from PhedEx https://cmsweb.cern.ch/phedex/prod/Info::Main 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Alessandro De Salvo – Grid Activities in INFN Roma CMS ECAL calibration “on-line”: short term calibration (+laser monitoring correction) at CAF (CERN) “off-line”: long term calibration at Tier2 Copy of single and double electron dataset in RECO and ALCARECO formats Iterative calibration algorithms run every n pb-1 Crystal calibration constants sent back to CERN 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Data processing using the Grid tools VIRGO Mainly used for Data Analysis Computing resources accessible by VIRGO (higher priority) THEOPHYS ARGO Farm utilization in Roma since January 2009 43000 jobs 252800 h total CPU time Raw data are produced in Cascina and copied to Tier-1s Calibrated data are replicated in (some of the) Tier-2 (Roma, Pisa, RFKI) Cascina Tier-0 CNAF Lyon Tier-1 Tier-2 Roma Pisa Nikhef APC-Parigi RFKI-Budapest 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Data processing using the Grid tools Theoretical Physics 2005 - 2007 Plots from the Grid Monitoring Systems http://gridice4.cnaf.infn.it:50080/gridice/site/site.php https://dgas.cnaf.infn.it 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Summary of the Grid-related activities in Roma LHC experiments ATLAS and CMS calibration activities ATLAS Software Installation in the Grid sites European projects WLCG (World LHC Computing Grid) and EGEE (Enabling Grid for E-sciencE) High Energy Computing Clusters [ATLAS] High-level Job Scheduling system (Job Priority) LCG 3D Distributed Database System Mostly related to the Oracle installation for the ATLAS calibrations Active backup of the ATLAS LCG File Catalog at CNAF Work presented at CHEP 2009 Italian Grid Infrastructure (IGI) High-performance storage systems dCache [CMS] DPM [ATLAS, VIRGO] StoRM [ATLAS] Workload Management System interaction with HEP Clusters [ATLAS] Grid Computing in Roma ATLAS Alessandro De Salvo Daniela Anzellotti Danilo D’Angelo Franco Marzano Lamberto Luminari ATLAS CMS Cristina Bulfon Giovanni Organtini Leonardo Sala (Milano) Luciano M Barone Massimo Nuccetelli VIRGO Alberto Colla Cristiano Palomba Theoretical Physics Silvano Petrarca 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Alessandro De Salvo – Grid Activities in INFN Roma Conclusions Roma is fully integrated with the Italian Grid Infrastructure and the WLCG systems The ATLAS and CMS collaborations are actively using the resources located in Roma to produce Monte Carlo data, perform user analysis and computation of calibration constants The system is robust and powerful enough to cope with the incoming LHC data taking Other experiments or groups (VIRGO and the Theoretical Physics) are already actively using the grid services to cope with their computing requirements Several Grid-related activities are present in Rome Included in Italian, European and/or Worldwide projects Room for further development and improvements Always trying to improve the levels of reliability and stability of the systems All the users in Roma are invited to take advantage of the experience of the people involved in the Grid projects and join the efforts The Grid is waiting for you! 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma

Alessandro De Salvo – Grid Activities in INFN Roma More information http://igi.cnaf.infn.it http://www.cern.ch/lcg http://www.eu-egee.org http://www.gridcafe.org ATLAS http://atlas.web.cern.ch http://cms.web.cern.ch http://www.virgo.infn.it 7-5-2009 Alessandro De Salvo – Grid Activities in INFN Roma