1 The ALICE Tier-2’s in Italy Roberto Barbera (*) Univ. of Catania and INFN Workshop CCR INFN 2006 Otranto, 08.06.2006 (*) Many thanks to A. Dainese, D.

Slides:



Advertisements
Similar presentations
1 ALICE Grid Status David Evans The University of Birmingham GridPP 14 th Collaboration Meeting Birmingham 6-7 Sept 2005.
Advertisements

LCG Tiziana Ferrari - SC3: INFN installation status report 1 Service Challenge Phase 3: Status report Tiziana Ferrari on behalf of the INFN SC team INFN.
Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
ALICE Operations short summary and directions in 2012 WLCG workshop May 19-20, 2012.
1 Status of the ALICE CERN Analysis Facility Marco MEONI – CERN/ALICE Jan Fiete GROSSE-OETRINGHAUS - CERN /ALICE CHEP Prague.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Preparation of KIPT (Kharkov) computing facilities for CMS data analysis L. Levchuk Kharkov Institute of Physics and Technology (KIPT), Kharkov, Ukraine.
BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES P. Saiz (IT-ES) AliEn job agents.
LCG Service Challenge Phase 4: Piano di attività e impatto sulla infrastruttura di rete 1 Service Challenge Phase 4: Piano di attività e impatto sulla.
IST E-infrastructure shared between Europe and Latin America High Energy Physics Applications in EELA Raquel Pezoa Universidad.
Sejong STATUS Chang Yeong CHOI CERN, ALICE LHC Computing Grid Tier-2 Workshop in Asia, 1 th December 2006.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
The ALICE short-term use case DataGrid WP6 Meeting Milano, 11 Dec 2000Piergiorgio Cerello 1 Physics Performance Report (PPR) production starting in Feb2001.
1 PRAGUE site report. 2 Overview Supported HEP experiments and staff Hardware on Prague farms Statistics about running LHC experiment’s DC Experience.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Tier2 Centre in Prague Jiří Chudoba FZU AV ČR - Institute of Physics of the Academy of Sciences of the Czech Republic.
The ALICE Distributed Computing Federico Carminati ALICE workshop, Sibiu, Romania, 20/08/2008.
Status of PDC’07 and user analysis issues (from admin point of view) L. Betev August 28, 2007.
Fabric Monitoring at the INFN Tier1 Felice Rosso on behalf of INFN Tier1 Joint OSG & EGEE Operations WS, Culham (UK)
Tier 3 Status at Panjab V. Bhatnagar, S. Gautam India-CMS Meeting, July 20-21, 2007 BARC, Mumbai Centre of Advanced Study in Physics, Panjab University,
Site Report: Prague Jiří Chudoba Institute of Physics, Prague WLCG GridKa+T2s Workshop.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
Materials for Report about Computing Jiří Chudoba x.y.2006 Institute of Physics, Prague.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks The LCG interface Stefano BAGNASCO INFN Torino.
EGEE is a project funded by the European Union under contract INFSO-RI Grid accounting with GridICE Sergio Fantinel, INFN LNL/PD LCG Workshop November.
Status of AliEn2 Services ALICE offline week Latchezar Betev Geneva, June 01, 2005.
Service Challenge Report Federico Carminati GDB – January 11, 2006.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
ALICE Computing TDR Federico Carminati June 29, 2005.
ATLAS TIER3 in Valencia Santiago González de la Hoz IFIC – Instituto de Física Corpuscular (Valencia)
ALICE Physics Data Challenge ’05 and LCG Service Challenge 3 Latchezar Betev / ALICE Geneva, 6 April 2005 LCG Storage Management Workshop.
Gestion des jobs grille CMS and Alice Artem Trunov CMS and Alice support.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
Tier2 Centre in Prague Jiří Chudoba FZU AV ČR - Institute of Physics of the Academy of Sciences of the Czech Republic.
The ALICE Production Patricia Méndez Lorenzo (CERN, IT/PSS) On behalf of the ALICE Offline Project LCG-France Workshop Clermont, 14th March 2007.
LHCC meeting – Feb’06 1 SC3 - Experiments’ Experiences Nick Brook In chronological order: ALICE CMS LHCb ATLAS.
Pledged and delivered resources to ALICE Grid computing in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
Care and feeding of the alice grid Torino, Jan 15-16, 2009.
1-2 March 2006 P. Capiluppi INFN Tier1 for the LHC Experiments: ALICE, ATLAS, CMS, LHCb.
Kilian Schwarz ALICE Computing Meeting GSI, October 7, 2009
Dynamic Extension of the INFN Tier-1 on external resources
INFNGRID Technical Board, Feb
SuperB – INFN-Bari Giacinto DONVITO.
ARDA-ALICE activity in 2005 and tasks in 2006
ALICE and LCG Stefano Bagnasco I.N.F.N. Torino
Eleonora Luppi INFN and University of Ferrara - Italy
LCG Service Challenge: Planning and Milestones
StoRM: a SRM solution for disk based storage systems
INFN-GRID Workshop Bari, October, 26, 2004
Grid Computing for the ILC
GSIAF & Anar Manafov, Victor Penso, Carsten Preuss, and Kilian Schwarz, GSI Darmstadt, ALICE Offline week, v. 0.8.
ALICE Physics Data Challenge 3
Update on Plan for KISTI-GSDC
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
The INFN TIER1 Regional Centre
MC data production, reconstruction and analysis - lessons from PDC’04
Simulation use cases for T2 in ALICE
ALICE Computing Upgrade Predrag Buncic
R. Graciani for LHCb Mumbay, Feb 2006
LHC Data Analysis using a worldwide computing grid
The LHCb Computing Data Challenge DC06
Presentation transcript:

1 The ALICE Tier-2’s in Italy Roberto Barbera (*) Univ. of Catania and INFN Workshop CCR INFN 2006 Otranto, (*) Many thanks to A. Dainese, D. Di Bari, S. Lusso, and M. Masera for providing slides and information for this presentation.

Workshop CCR INFN 2006, Otranto, Outline The ALICE computing model and its parameters ALICE and the Grid(s) Layout Implementation Recent results ALICE Tier-2’s in Italy Catania Torino Bari LNL-PD Summary and conclusions

Workshop CCR INFN 2006, Otranto, The ALICE computing model (1/2) pp Quasi-online data distribution and first reconstruction at T0 Further reconstructions at T1’s AA Calibration, alignment and pilot reconstructions during data taking Data distribution and first reconstruction at T0 during four months after AA Further reconstructions at T1’s One copy of RAW at T0 and one distributed at T1’s

Workshop CCR INFN 2006, Otranto, The ALICE computing model (2/2) T0 First pass reconstruction, storage of one copy of RAW, calibration data and first-pass ESD’s T1 Reconstructions and scheduled analysis, storage of the second collective copy of RAW and one copy of all data to be kept, disk replicas of ESD’s and AOD’s T2 Simulation and end-user analysis, disk replicas of ESD’s and AOD’s

Workshop CCR INFN 2006, Otranto, Parameters of the ALICE computing model UnitppPbPb T1#7 T2#23 Size rawMB0.2x512.5 Recording rateHz100 ESDMB AODkB4250 Event CataloguekB10 Running time s Events / y# Reconstruction passes (av)#3 RAW duplication#2 AOD/ESD duplication#2 Scheduled analysis passes / rec ev / y (av)#3 Chaotic analysis passes / rec ev / y (av)#20

Workshop CCR INFN 2006, Otranto, Legenda: TQ= Task Queue  Central job DB CAT= Central Catalogue ALICE & the Grid(s) ALICE Agents & Daemons ROOT ALIROOT Computing framework Resources NU Grid Resources ALICE TQ ALICE Agents & Daemons OSG Resources ALICE user ALICE CAT

Workshop CCR INFN 2006, Otranto, Implementation: the “VO-Box” LCG Site LCG CE WN JobAgent LCG SE LCG RB TQ VO-Box SCA SA Job request LFC SURL Registration File Catalogue LFN Registration PackMan Request configuration

Workshop CCR INFN 2006, Otranto, Who does what ? Configure, submit and track jobs User interface with massive production support Job DB (Production and user) User and role management Install software on sites Package Managers Distribute and execute jobs Workload Management System (Broker, L&B,…) Computing Element software Information Services Interactive analysis jobs Store and catalogue data Data catalogues (file, replica, metadata, local,…) Storage Element software Move data around File Trasfer services and schedulers Access data files I/O services File management (SRM) Monitor all that stuff Transport infrastructure Sensors Web presentation..on top of that: Enforce security! MIXED PROOF Mon ALISA MIXED Xrootd

Workshop CCR INFN 2006, Otranto, Some statistics and results for SC3/PDC05 In the last two months of 2005: 22,500 jobs (Pb+Pb and p+p) Average CPU time: 8 hours Data volume produced: 20 TB (90% CASTOR2 at CERN, 10% remote sites) Resource Centres participating (22 in total) 4 T1: CERN, CNAF, GridKa, CCIN2P3 18 T2: Bari, Clermont (FR), GSI (D), Houston (USA), ITEP (RUS), JINR (RUS), KNU (UKR), Muenster (D), NIHAM (RO), OSC (USA), PNPI (RUS), SPbSU (RUS), Prague (CZ), RMKI (HU), SARA (NL), Sejong (SK), Torino, UiB (NO) Job share per site: T1: CERN 19%, CNAF 17% (CPU 20%), GridKa 31%, CCIN2P3 22% T2: total of 11% Failure rate di AliRoot: 2.5%

Workshop CCR INFN 2006, Otranto, Job execution profile during SC jobs (25% more than entire lxbatch capacity at Cern) Negative slope: AliEn problem during output retrieval. Fixed in the further release!

Workshop CCR INFN 2006, Otranto, Without INFN-T1 ~ job ~ job Memento: VO= Virtual Organization (esperimento) ALICE: 8% of the total number of jobs on the national grid Use of INFN Grid by LHC Exps.: JOB/VO (Sep Dec 2005)

Workshop CCR INFN 2006, Otranto, ~ 98 years, 2 month, 18 days ~ 358 years, 7 months, 11 days Without INFN-T1 ALICE: 14% of CPU time outside T1 Use of INFN Grid by LHC Exps.: CPU/VO (Sep Dec 2005)

Workshop CCR INFN 2006, Otranto, ALICE JOBS PER SITE. Warning: Job agents and real jobs are accounted in the same way

Workshop CCR INFN 2006, Otranto, ALICE Tier-2’s in Italy Four candidates: Bari, Catania, LNL-PD, and Torino (T2 projects available at the URL: The team of ALICE referees with representatives of the INFN Management Board visited all Tier-2 candidates between 10/2005 and 02/2006. Referees’ decision communicated at a meeting in Rome on 10/03/2006: Catania and Torino approved; Bari and LNL-PD “incubated” (kept in “life support” until real ALICE needs are proved by real test of the computing model in production mode).

Workshop CCR INFN 2006, Otranto, Network connectivity of T2-s ALICE Tier-2’s

Workshop CCR INFN 2006, Otranto, Catania (1/5) – Comp. room Present installation Future expansion Space available for installations: ~160 m 2

Workshop CCR INFN 2006, Otranto, Catania (2/5) - Infrastructure Traditional System High Density System

Workshop CCR INFN 2006, Otranto, Catania (3/5) - CPU 150 kSI2k SuperMicro dual AMD dual-core 275 with 4 GB RAM in 1U configuration IBM LS20 “blades” with dual AMD dual-core 280 with 4 GB RAM (within june) LSF 6.1 as LRMS

Workshop CCR INFN 2006, Otranto, Catania (4/5) - Storage 21+ TB with GPFS FC-2-SATA systems plus more traditional DAS with EIDE-2-SCSI controllers Filesystem: GPFS

Workshop CCR INFN 2006, Otranto, Catania (5/5) - Statistics Last month activity

Workshop CCR INFN 2006, Otranto, Torino (1/5) – Computing Room

Workshop CCR INFN 2006, Otranto, Torino (2/5) - Present installation Present solutions: blade servers (IBM) and 1U biprocessors Guidelines for the future: Minimize space Minimize power consumption

Workshop CCR INFN 2006, Otranto, Torino (3/5) - Resources CPU 38 Intel(R) Xeon(TM) CPU 2.40GHz; 12 Intel(R) Xeon(TM) CPU 3.06GHz. 45 Intel Biprocessors (<=4 years – 14 Blades) DISK ~6TB dedicated to ALICE 2TB shared among various VO’s (Classic-SE); 1 dCache SE with an internal disk of ~80GB for tests; ~15TB of disk space for ALICE is going to be commissioned soon. It is a FLX210 with 3 FLC200 expansions from di StorageTek Filesystem Ext3 for the ClassicSE; not yet defined for the new storage system; Tests with xrootd for local and remote access (through proxy) are scheduled. LRMS Torque-Maui; the default one coming with the INFN Grid release Open to all VO’s Dedicated to ALICE (at the moment)

Workshop CCR INFN 2006, Otranto, Torino (4/5) - Resources Future evolution Many nodes (~20 – the most recent) are being migrated from the ALICE farm to the LCG farm exploiting the forthcoming upgrade to gLite 3.0; New WN’s (80 cores – 130 KSI2K), recently bought, will be installed and configured very soon. Networking: All WN’s are in a hidden LAN (only outbound connectivity is allowed) and the NATting is done by an Extreme Networks switch. Almost all connection are Gigabit Ethernet. Monitoring: MRTG and NAGIOS for the local control of the farm.

Workshop CCR INFN 2006, Otranto, Torino (5/5) - Usage Scheduler locale. # di job LCG. Numero di Job Monitoring centrale ALICE

Workshop CCR INFN 2006, Otranto, Bari (1/2) Bari is a Tier-2 candidate both for ALICE and CMS. Bari supports also other VO’s. Priorities are given to the various VO’s proportionally to the different budgets for acquiring resources. In the last two years Bari has provided resources for ALICE both for PDC04 and SC3 and will provide for SC4.

Workshop CCR INFN 2006, Otranto, Bari (2/2) One 2 cpu 700 MHz PIII aligrid1.ba.infn.it - HD 40 GB One 2 cpu 1 GHz PIII alicegrid2.ba.infn.it - HD 160 GB Three 2 cpu Intel Xeon 1.8 GHz alicegrid4 - alicegrid6 (VOBOX) - 3 HD da 80GB One 2 cpu Intel Xeon 1.8 GHz alicegrid3.ba.infn.it - (SE for PDC04) with 0.7 TB of data One 2 cpu Intel Xeon 2.4 GHz alicegrid5.ba.infn.it - (SE for Finuda) with 1.5 TB disk space Three 2 cpu Intel Xeon 2.4 GHz - HD 80 GB One 2 cpu Intel Xeon 2.4 GHz alicegrid7.ba.infn.it - HD 80 GB - software repository + Quattor installation server One Opteron 2 dual core HD 120 GB Three 2 cpu Intel Xeon 2.8 GHz - HD 80 GB One 2 cpu Intel Xeon 3.0 GHz EM64T - HD 2 array x 2.5 TB (TOT 5 TB) (to be configured with xrootd for SC4)

Workshop CCR INFN 2006, Otranto, ALICE jobs at Bari (monitored by MonaLisa)

Workshop CCR INFN 2006, Otranto, LNL-PD Background: LNL-PD is an approved Tier-2 for CMS; Many-years experience in running a T2 prototype for CMS. Size of the existing Tier-2 for CMS: CPU: ~200 KSI2K (almost all “blades” dual core) Storage: EIDE-2-SCSI DAS with 3Ware + Storage Area Network LRMS: LSF Monitoring: Ganglia (local) + GridIce

Workshop CCR INFN 2006, Otranto, ALICE at LNL-PD ALICE activities already done: ALICE VO-box installed in 02/2006 Site testing with small productions OK Big ALICE production in April-May via LCG Future activities foreseen for the rest of 2006: Participation to PDC06 (~10 kSI2k dedicated resources + the possibility to use CMS resources, if/when available) Installation of an ALICE storage system with xrootd (~1 TB at the beginning)

Workshop CCR INFN 2006, Otranto, ALICE jobs at LNL-PD (monitored by GridIce) ALICE 15 April 2006 – 15 May 2006

Workshop CCR INFN 2006, Otranto, Common issues Need for a common solution for the infrastructure (to improve the economy of scale). Need for an affordable, reliable, and scalable solution for the storage. Need for a better organization of distributed support for Tier-2’s. Although new technologies (“blades” with low-power CPU’s) help a bit, power consumption at Tier-2 sites is becoming increasingly important from an economic point of view. Strict guidelines and a dedicated budget should be centrally created by INFN Management.

Workshop CCR INFN 2006, Otranto, The future: PDC06 (June 2006) Check of the distributed computing model: From raw-data to ESD Data tranfers among sites Calibration and alignment Analysis SC3 experience has helped a lot to improve AliEn (current version 2.10) Intense development of AliRoot to include calibration and alignment code for all sub- detectors and reduce the percentage of run time failures. Huge effort of the Italian groups in many sites.

Workshop CCR INFN 2006, Otranto, Resources ramp up at INFN Tier-2’s

Workshop CCR INFN 2006, Otranto, Summary and conclusions The ALICE computing model has been finalized and now it is ready to face the forthcoming data from LHC. INFN has identified the first official Tier-2’s for ALICE. Both for the design and the day-by-day operation of a LHC Tier-2 a strong collaboration between the Experiments, the INFN Grid Project, the INFN CCR, and the Computing&Network Services at the various INFN Departments is of vital importance.