CMS Data Challenge Experience on LCG-2

Slides:



Advertisements
Similar presentations
CMS Applications – Status and Near Future Plans
Advertisements

CMS Grid Batch Analysis Framework
HEPiX GFAL and LCG data management Jean-Philippe Baud CERN/IT/GD.
Réunion DataGrid France, Lyon, fév CMS test of EDG Testbed Production MC CMS Objectifs Résultats Conclusions et perspectives C. Charlot / LLR-École.
CMS-ARDA Workshop 15/09/2003 CMS/LCG-0 architecture Many authors…
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
A tool to enable CMS Distributed Analysis
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Stefano Belforte INFN Trieste 1 CMS SC4 etc. July 5, 2006 CMS Service Challenge 4 and beyond.
Dave Newbold, University of Bristol24/6/2003 CMS MC production tools A lot of work in this area recently! Context: PCP03 (100TB+) just started Short-term.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
Claudio Grandi INFN Bologna ACAT'03 - KEK 3-Dec-2003 CMS Distributed Data Analysis Challenges Claudio Grandi on behalf of the CMS Collaboration.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES P. Saiz (IT-ES) AliEn job agents.
Claudio Grandi INFN Bologna CHEP'03 Conference, San Diego March 27th 2003 Plans for the integration of grid tools in the CMS computing environment Claudio.
Interactive Job Monitor: CafMon kill CafMon tail CafMon dir CafMon log CafMon top CafMon ps LcgCAF: CDF submission portal to LCG resources Francesco Delli.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
CMS Stress Test Report Marco Verlato (INFN-Padova) INFN-GRID Testbed Meeting 17 Gennaio 2003.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
29 Sept 2004 CHEP04 A. Fanfani INFN Bologna 1 A. Fanfani Dept. of Physics and INFN, Bologna on behalf of the CMS Collaboration Distributed Computing Grid.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
6 march Building the INFN Grid Proposal outline a.ghiselli,l.luminari,m.sgaravatto,c.vistoli INFN Grid meeting, milano.
INFSO-RI Enabling Grids for E-sciencE CRAB: a tool for CMS distributed analysis in grid environment Federica Fanzago INFN PADOVA.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Enabling Grids for E-sciencE CMS/ARDA activity within the CMS distributed system Julia Andreeva, CERN On behalf of ARDA group CHEP06.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
CMS Production Management Software Julia Andreeva CERN CHEP conference 2004.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
June 22 L. Silvestris CMS/ARDA CMS/ARDA Lucia Silvestris INFN-Bari.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
EDG Project Conference – Barcelona 13 May 2003 – n° 1 A.Fanfani INFN Bologna – CMS WP8 – Grid Planning in CMS Outline  CMS Data Challenges  CMS Production.
Claudio Grandi INFN Bologna Workshop congiunto CCR e INFNGrid 13 maggio 2009 Le strategie per l’analisi nell’esperimento CMS Claudio Grandi (INFN Bologna)
Federating Data in the ALICE Experiment
Jean-Philippe Baud, IT-GD, CERN November 2007
Workload Management Workpackage
BaBar-Grid Status and Prospects
Real Time Fake Analysis at PIC
LCG Service Challenge: Planning and Milestones
ATLAS Use and Experience of FTS
U.S. ATLAS Grid Production Experience
INFN GRID Workshop Bari, 26th October 2004
Data Challenge with the Grid in ATLAS
Comparison of LCG-2 and gLite v1.0
BOSS: the CMS interface for job summission, monitoring and bookkeeping
BOSS: the CMS interface for job summission, monitoring and bookkeeping
INFN-GRID Workshop Bari, October, 26, 2004
CMS — Service Challenge 3 Requirements and Objectives
Sergio Fantinel, INFN LNL/PD
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
CMS Data Challenge 2004 Claudio Grandi CMS Grid Coordinator
BOSS: the CMS interface for job summission, monitoring and bookkeeping
Scalability Tests With CMS, Boss and R-GMA
LCG middleware and LHC experiments ARDA project
N. De Filippis - LLR-Ecole Polytechnique
R. Graciani for LHCb Mumbay, Feb 2006
LHC Data Analysis using a worldwide computing grid
INFNGRID Workshop – Bari, Italy, October 2004
The LHCb Computing Data Challenge DC06
Presentation transcript:

CMS Data Challenge Experience on LCG-2 Claudio Grandi INFN - Bologna INFNGrid Workshop - Bari October 26th 2004

INFNGrid Workshop - Bari Outline Definition of Data Challenge 2004 (DC04) Pre-Challenge Production (PCP) DC04 experience What’s next Conclusions INFNGrid Workshop - Bari October 26th 2004

INFNGrid Workshop - Bari CMS Data Challenge 2004 Aim of DC04: Reconstruct data at a sustained rate of 25Hz in the Tier-0 farm 25% of the target conditions for LHC startup register data and metadata to a catalogue transfer the reconstructed data to all Tier-1 centers analyze the reconstructed data at the Tier-1’s as they arrive publicize to the community the data produced at Tier-1’s monitor and archive of performance criteria of the ensemble of activities for debugging and post-mortem analysis Not a CPU challenge, but a full chain demonstration! Pre-challenge production in 2003/04 90M Monte Carlo events (40M Geant4) simulated and 85M events digitized to date 15M digitized in time for DC04, another 10M during DC04 Classic and grid (CMS/LCG-0, LCG-1, Grid3) productions Now continuous production INFNGrid Workshop - Bari October 26th 2004

Pre-Challenge Production layout Dataset metadata Phys.Group asks for a new dataset JDL Grid (LCG) Scheduler LCG-0/1 RLS Job metadata DAG job DAGMan (MOP) Chimera VDL Virtual Data Catalogue Planner Grid3 Production Manager defines assignments RefDB Computer farm shell scripts Data-level query Local Batch Manager Job level query BOSS DB McRunjob + plug-in CMSProd Site Manager starts an assignment Push data or info Pull info INFNGrid Workshop - Bari October 26th 2004

INFNGrid Workshop - Bari PCP statistics 850K jobs 4500 KSI2000 months 850K files 100 TB of data DC04 Simulation DC04 Digitization Now running continuous productions INFNGrid Workshop - Bari October 26th 2004

INFNGrid Workshop - Bari PCP on grid: CMS-LCG CMS-LCG Regional Center - 0.5 Mevts “heavy” pythia: ~2000 jobs ~10 KSI2000 months - 2.1 Mevts cmsim+oscar: ~8500 jobs ~130 KSI2000 months ~2 TB data DC04 Gen+Sim on LCG CMS/LCG-0 LCG-1 RefDB Eff:70%-90% Eff: ~60% Dataset metadata RLS CMS/LCG-0 Joint project CMS-LCG-EDT Based on LCG pilot distribution Including GLUE, VOMS, GridICE, RLS About 170 CPU’s and 4 TB disk Sites: Bari Bologna Bristol Brunel CERN CNAF Ecole Polytechnique Imperial College ISLAMABAD-NCP Legnaro Milano NCU-Taiwan Padova U.Iowa UI CE SE McRunjob + ImpalaLite JDL RB CE CE SE SE bdII WN BOSS CE Job metadata SE Push data or info Pull info INFNGrid Workshop - Bari October 26th 2004

PCP (and beyond) on Grid: Grid3 USMOP Regional Center - 7.7 Mevts pythia: ~30000 jobs ~0.7 KSI2000 monhs - 19 Mevts cmsim+oscar: ~19000 jobs ~1000 KSI2000 months ~13 TB data ~ 70% overall job efficiency Mainly non-grid related DC04 Simulation on Grid3 MOP System Master Site Remote Site 1 MCRunJob mop_submitter DAGMan Condor-G GridFTP Batch Queue Remote Site N Grid3 US grid projects + US LHC expt.’s Over 2000 CPU’s in 25 sites MOP Dagman and Condor-G for specification and submission Condor-based match-making process selects resources INFNGrid Workshop - Bari October 26th 2004

INFNGrid Workshop - Bari Data Challenge 2004 layout Tier-2 Physicist T2 storage ORCA Local Job Tier-2 Physicist T2 storage ORCA Local Job Tier-0 Castor IB fake on-line process RefDB POOL RLS catalogue TMDB ORCA RECO Job GDB data distribution agents EB Tier-2 Physicist T2 storage ORCA Local Job LCG-2 Services Tier-1 agent T1 storage ORCA Analysis Job MSS Grid Job Tier-1 agent T1 storage ORCA Analysis Job MSS Grid Job Tier-1 agent T1 storage ORCA Analysis Job MSS Grid Job INFNGrid Workshop - Bari October 26th 2004

INFNGrid Workshop - Bari DC04 setup CERN RB bdII RLS RefDB MonaLisa Tier-0 LCG tools CMS tools SE Castor VOMS SE TMDB SRM SRB other tools LCG-2 RM chain SRB chain SRM chain Tier-1’s CNAF (Italy) GridICE RB PIC (Spain) RB FNAL (USA) RAL (UK) MCAT GridKA (Germany) IN2P3 (France) SE Castor SE Castor SRM Enstore SRB Castor SRB Tivoli SRB HPSS UI CE SE UI CE SE Local farm Local farm Local farm Local farm Tier-2’s Legnaro CIEMAT UFL Caltech CE SE UI CE SE SRM Local farm SRM Local farm INFNGrid Workshop - Bari October 26th 2004

INFNGrid Workshop - Bari DC04 Processing Rate Processed more than 25M events Generally kept up with data transfers (CNAF, PIC, FNAL) Got above 25Hz on many short occasions But only one full day above 25Hz with full system Overloaded RLS, Castor@T0, MSS@T1, … INFNGrid Workshop - Bari October 26th 2004

INFNGrid Workshop - Bari DC04: RLS RLS used as a global POOL catalogue, with full file meta data LRC: Global file catalogue (GUID  PFNs) 570K LFNs registered in RLS, each with  5-10 PFN’s RMC: Global metadata catalogue (GUID  metadata) 9 metadata attributes per file (up to ~1 KB metadata per file) Performances: LRC fast enough C++ API programs (0.1-0.2sec/file) POOL CLI with GUID (secs/file) Need optimization for bulk queries RMC too slow secs/filewhen everything goes well hours to get info for a collection 25 Hz Time to register 16 files (1 reco job output) INFNGrid Workshop - Bari October 26th 2004

DC04 LCG-2 data transfer chain (1/2) Tier-0: Export Buffer: classic-SE 3 SE machines with 1 TB disk each Tier-1’s: Castor-SE receiving data but different underlying MSS hardware solution Performances: CNAF: Replica Manager CLI copy a file and inherently register it to the RLS, with file-size info stored in the LRC over-head introduced by CLI java processes safer against failed replicas PIC: globus-url-copy + LRC C++ API copy a file and later register to the RLS, no file-size check faster! no quality-check of replica operations TMDB RLS RM data distribution agent CERN Castor Disk SE EB Tier-1 Tier-1 agent Castor CASTOR SE Tier-2 Disk SE INFNGrid Workshop - Bari October 26th 2004

DC04 LCG-2 data transfer chain (2/2) Both CNAF and PIC approaches achieved good performances T1 agents robust, kept the pace with data available at EB Dealing with too many small files (affected all distribution chains) “bad” for: → efficient use of bandwidth → scalability of MSS systems network ‘stress-test’ at the end of DC04 with ‘big’ files: typical transfer rates >30 MB/s, CNAF sustained >42 MB/s for some hours CERN Tier-0: SE-EB of the LCG chain CNAF T1 network monitoring eth I/O eth I/O ~340 Mbps CNAF Tier-1: Castor-SE >3k files >750 GB CNAF Tier-1: Classic disk-SE INFNGrid Workshop - Bari October 26th 2004

DC04 Real-time Data Analysis Submit analysis jobs automatically as new data are available on Tier-1 and Tier-2 SE’s Job submission via the LCG-2 Resource Broker Job sent close to the data File access on SE via rfio Output data registered in RLS Job monitoring with BOSS Kept up with the rate of data coming from CERN > 17500 analysis jobs in ~2 weeks ~ 40 Hz maximum event frequency 90-95% grid efficiency ~20 minutes delay from data at Tier-0 to analysis at Tier-1 Real-time analysis was done only in LCG environment (Italy & Spain) INFNGrid Workshop - Bari October 26th 2004

What’s next: End-User Analysis Now focus in providing an End-to-End analysis system PhySh (Physics Shell) is an “end-user” environment Based on Clarens Web services A “glue” interface among different services User’s interface modeled as a virtual file system PhySh Virtual File System Interface XML RPC SOAP or Python Software Env. SCRAM(local) Job submission (local) Clarens Web Services Other XML RPC or SOAP client Shell client Web client Pyhton client Dataset Catalogue (PubDB/RefDB) Data transfer service (PheDex) Software Environment (SCRAM) Job Submission INFNGrid Workshop - Bari October 26th 2004

What’s next: Data Access Data Location service CMS specific DataSet catalogue: RefDB + PubDB Allow Dataset discovery and location of Dataset catalogues Data transfer service PheDEX (Physics Experiment Data Export) Evolution of TMDB system used in DC04 Allows request management and file transfer based on data subscriptions Dataset discovery RefDB Publication discovery (PubDB-URLs) PubDB (INFN) PubDB (CERN) Local catalogues Local catalogues INFNGrid Workshop - Bari October 26th 2004

What’s next: analysis job submission Data Location Interface File-based data location UI PhySh Resource Broker (RB) node RLS Logical File Name location (URL) Workload Management System dataset-based data location The end-user inputs: Datasets (runs,#event,…) + private code Logical Dataset DataSet Catalogue (PubDB/RefDB) PhySh location (URL) Several tools for job preparation, splitting, submission are under development integrated with: LCG (GROSS,GRAPE) Grid3 (Runjob) EGEE (gLite prototype) Monitoring services and jobs MonaLisa, GridICE, BOSS Storage Element Inform. Service Computing Element INFNGrid Workshop - Bari October 26th 2004

INFNGrid Workshop - Bari Conclusions CMS distributed production based on grid middleware used in the official CMS production system Grid3: reliable and scalable system for massive production LCG: large scale productions proved distributed sites consistent configuration and control is very important LCG-2 used in CMS Data Challenge LCG environment provides the functionalities for distributed computing Full chain (but the Tier-0 reconstruction) done on the LCG-2 system The catalogues are an issue! Grid point-to-point file transfer tools Infrastructure for data analysis LCG data distribution and data analysis chain successfully met the data challenge goals of large scale scheduled distribution to a set of Tier-1/2 and subsequent analysis New end-to end analysis system in development Components by several grid projects LCG, EGEE (GLite), GRID-3 … PhySh – virtual file system is the “glue” different services A consistent interface to the physicist A flexible application framework A set of back-end services INFNGrid Workshop - Bari October 26th 2004

INFNGrid Workshop - Bari Backup slides INFNGrid Workshop - Bari October 26th 2004

SRM Interaction Diagram INFNGrid Workshop - Bari October 26th 2004

DC04 SRM data transfer chain (1/2) Tier-0: SRM/dCache based DRM serving as an EB files are staged out of Castor to the dCache pool disk and pinned until transferred Tier-1: SRM/dCache/Enstore based HRM acting as Import Buffer with SRM interface providing access to Enstore via dCache SRM transactions to receive TURLs from EB, transfers via gridftp SRM client (T1 agent machine) SRM-COPY CERN T0 FNAL T1 SRM-GET (one file at a time, return TURL) SRM (performs stage, pin file) SRM (performs space reservation, write) GridFTP GET (pull mode) Network transfer Enstore EB dCache Pool dCache Pool INFNGrid Workshop - Bari October 26th 2004

DC04 SRM data transfer chain (2/2) In general quite robust tools e.g. SRM in error checking/retrying, dCache for authomatic migration to tape, .. Stressed a few software/hardware components to the breaking point e.g. monitoring not implemented to catch service failures, forcing manual interventions Again: problems of high-nb/small-size of DC files use of multiple streams with multiple files in each stream reduced the overhead of the authentication process MSS optimization necessary to handle the challenge load Inefficient use of tapes forced more tapes allocations + deployment of larger namespace service relevant improvements during ongoing DC operations e.g. reduction on delegated proxy’s modulus size in SRM yielding a speed-up of the interaction between SRM client and server of a factor 3.5 INFNGrid Workshop - Bari October 26th 2004

INFNGrid Workshop - Bari PhySh Data are organized as files in a directory structure Actions are expressed in terms of copying, linking, creating files Multiple clients satisfy different needs INFNGrid Workshop - Bari October 26th 2004

INFNGrid Workshop - Bari PhySh cp /Applications/ORCA/ORCA_8_4_0 /Tiers/FNAL/ cp /RefDB/eg03_foo/DST /Tiers/PIC Installs ORCA_8_4_0 at FNAL Copies DST for dataset eg03_bar to PIC INFNGrid Workshop - Bari October 26th 2004