Claudio Grandi INFN Bologna Workshop congiunto CCR e INFNGrid 13 maggio 2009 Le strategie per l’analisi nell’esperimento CMS Claudio Grandi (INFN Bologna)

Slides:



Advertisements
Similar presentations
1 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna A. Fanfani University of Bologna MC Production System & DM catalogue.
Advertisements

1 User Analysis Workgroup Update  All four experiments gave input by mid December  ALICE by document and links  Very independent.
CMS HLT production using Grid tools Flavia Donno (INFN Pisa) Claudio Grandi (INFN Bologna) Ivano Lippi (INFN Padova) Francesco Prelz (INFN Milano) Andrea.
A tool to enable CMS Distributed Analysis
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Client/Server Grid applications to manage complex workflows Filippo Spiga* on behalf of CRAB development team * INFN Milano Bicocca (IT)
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
Physicists's experience of the EGEE/LCG infrastructure usage for CMS jobs submission Natalia Ilina (ITEP Moscow) NEC’2007.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
Grid Resource Allocation and Management (GRAM) Execution management Execution management –Deployment, scheduling and monitoring Community Scheduler Framework.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Giuseppe Codispoti INFN - Bologna Egee User ForumMarch 2th BOSS: the CMS interface for job summission, monitoring and bookkeeping W. Bacchi, P.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
Claudio Grandi INFN Bologna CMS Computing Model Evolution Claudio Grandi INFN Bologna On behalf of the CMS Collaboration.
Jan 2010 OSG Update Grid Deployment Board, Feb 10 th 2010 Now having daily attendance at the WLCG daily operations meeting. Helping in ensuring tickets.
Claudio Grandi INFN Bologna CERN - WLCG Workshop 13 November 2008 CMS - Plan for shutdown and data-taking preparation Claudio Grandi Outline: Global Runs.
Use of the gLite-WMS in CMS for production and analysis Giuseppe Codispoti On behalf of the CMS Offline and Computing.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks CRAB: the CMS tool to allow data analysis.
INFSO-RI Enabling Grids for E-sciencE CRAB: a tool for CMS distributed analysis in grid environment Federica Fanzago INFN PADOVA.
DIRAC Pilot Jobs A. Casajus, R. Graciani, A. Tsaregorodtsev for the LHCb DIRAC team Pilot Framework and the DIRAC WMS DIRAC Workload Management System.
Daniele Spiga PerugiaCMS Italia 14 Feb ’07 Napoli1 CRAB status and next evolution Daniele Spiga University & INFN Perugia On behalf of CRAB Team.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Distributed Analysis Tutorial Dietrich Liko. Overview  Three grid flavors in ATLAS EGEE OSG Nordugrid  Distributed Analysis Activities GANGA/LCG PANDA/OSG.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
Monitoring the Readiness and Utilization of the Distributed CMS Computing Facilities XVIII International Conference on Computing in High Energy and Nuclear.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
1-2 March 2006 P. Capiluppi INFN Tier1 for the LHC Experiments: ALICE, ATLAS, CMS, LHCb.
EDG Project Conference – Barcelona 13 May 2003 – n° 1 A.Fanfani INFN Bologna – CMS WP8 – Grid Planning in CMS Outline  CMS Data Challenges  CMS Production.
Daniele Bonacorsi Andrea Sciabà
Dynamic Extension of the INFN Tier-1 on external resources
WLCG IPv6 deployment strategy
Workload Management Workpackage
Kevin Thaddeus Flood University of Wisconsin
Job monitoring and accounting data visualization
Regional Operations Centres Core infrastructure Centres
BaBar-Grid Status and Prospects
Xiaomei Zhang CMS IHEP Group Meeting December
LCG Service Challenge: Planning and Milestones
StoRM: a SRM solution for disk based storage systems
Moving the LHCb Monte Carlo production system to the GRID
Distributed Computing for HEP - present and future
POW MND section.
CREAM Status and Plans Massimo Sgaravatto – INFN Padova
BOSS: the CMS interface for job summission, monitoring and bookkeeping
BOSS: the CMS interface for job summission, monitoring and bookkeeping
INFN-GRID Workshop Bari, October, 26, 2004
Accounting at the T1/T2 Sites of the Italian Grid
CRAB and local batch submission
BOSS: the CMS interface for job summission, monitoring and bookkeeping
CRAB Server CRAB (CMS Remote Analysis Builder)
Compute Area Marco Cecchi Massimo Sgaravatto
Evolution of the distributed computing model The case of CMS
Artem Trunov and EKP team EPK – Uni Karlsruhe
Simulation use cases for T2 in ALICE
#01 Client/Server Computing
N. De Filippis - LLR-Ecole Polytechnique
DGAS Today and tomorrow
MonteCarlo production for the BaBar experiment on the Italian grid
#01 Client/Server Computing
The LHCb Computing Data Challenge DC06
Presentation transcript:

Claudio Grandi INFN Bologna Workshop congiunto CCR e INFNGrid 13 maggio 2009 Le strategie per l’analisi nell’esperimento CMS Claudio Grandi (INFN Bologna)

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 2 Analysis in CMS In the CMS Computing Model the analysis is done: –at Tier-2 centres User analysis, Physics Groups (PG) analysis, etc... Coexists with the MC productions done by the Data Operation team Shares based on VOMS group/role have been deployed –at Tier-1 centres for special and controlled analysis tasks e.g. Skimming, Data Quality Monitor (DQM), Calibration,... Most of the activities are carried out by the Data Operation team (identified by the VOMS role /cms/Role=production) A special VOMS role (/cms/Role=t1access) is granted to selected individuals and provides access to a limited amount of resources Analysis is done at the site hosting the data –Data placement controlled centrally and by the PGs –Results are staged out to sites associated with the user –CMS tools provide the environment for the preparation and the remote execution of the jobs

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 3 CMS Remote Analysis Builder CMS developed a tool (CRAB) for the transparent usage of the distributed system –It provides the user with a simple interface and a lightweight client –It provides a service platform to automate the user analysis workflow Includes: –interface to data discovery and location systems (DBS) locate sites with the desired data and get information on datasets –interface to CMS sites information database (SiteDB) find site configuration parameters related to user support –interface to the grid Information system find the CEs with the correct CMSSW release, etc... –interface to local and grid job management tools: BossLite Among others, BossLite provides interface to gLite-WMS, Condor- G, glidein-WMS and several local batch systems D.Spiga, CHEP09

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 4 Analysis with CRAB CRAB client Distributed Infrastructure Local Tier-2 SE Distributed Tier-2 SE Distributed Tier-2 SE Distributed Tier-2 SE Distributed Tier-2 SE Distributed Tier-2 SE /store/user /store/data /store/mc job submission & control job query small products access to official data access to user data remote stage-out CRAB Job tracking DB DBS data location

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 5 Analysis with CRAB Server CRAB client Distributed Infrastructure Local Tier-2 SE Distributed Tier-2 SE Distributed Tier-2 SE Distributed Tier-2 SE Distributed Tier-2 SE Distributed Tier-2 SE /store/user /store/data /store/mc job ops small products access to official data access to user data remote stage-out DBS data location task ops resubmissions CRAB Job tracking DB

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 6 Job distribution per activities From May 2008 to March 2009: 23 M jobs submitted 58% Success 25% application failures 12% grid failures 5% cancelled about 78% of the total analysis jobs are sumitted with the gLite WMS (the rest mainly CondorG) since years! ~600 distinct real users in the last 3 months 81% Success ~ 9% application failures 10% grid failures 8.8M Analysis Jobs 5.3M MC Production Jobs 87% Success 4% application failures 7% grid failures 6.6 M JobRobot 2% cancelled + 2,3M jobs other test activities G.Codispoti, CHEP09

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 7 Resources for user support Storage at Tier-2 centers is broken into 6 pieces –Transient and unmanaged to more persistent and centrally managed I.Fisk

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 8 Tier-2 storage control All numbers are for a nominal Tier-2 Central Space 30TB –Intended for RECO samples of Primary Datasets In 2008 we had expected to be able to store 2 copies of MC and data sample using the identified T2 space Physics Group Space 60-90TB –Assigned to 1-3 physics groups. Space allocated by physics data manager. The site data manager still approves the request, but only to ensure the group is below quota Local Storage Space 30TB-60TB –Controlled by the local storage manager. Intended to benefit the geographically associated community I.Fisk

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 9 Tier-2 – Physics Group association

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 10 Italian Tier-2s Four Italian Tier-2s are integrated into the CMS infrastructure but in some cases they suffer from lack of resources –No new resources allocated in 2009 as a consequence of the LHC incident Association with CMS physics groups defined Association with INFN institutes for local user support ongoing

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 11 Processing at the INFN Tier-2s Bari Legnaro Pisa Roma Analysis jobs regularly executed at INFN Tier-2s Monitored by the dashboard

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 12 Data at the INFN Tier-2s The Physics Group already started populating the Tier-2 storage Association to the groups is controlled by the subscribers (not rigorous at the time being) Monitored by PhEDEx

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 13 INFN Tier-2 monitoring Farm monitoring tools in use at sites –E.g. Pisa Grid Job Monitoring Authentication Job List Job Selection Configuration Job Detail S.Sarkar, CHEP09

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 14 Main issues to be addressed 1/3 Tools 1. The CRAB development team (and in general of CMS software tools) is understaffed –CRAB and BossLite are key components of CMS computing and are under INFN responsibility 2. User support for CRAB is very time consuming –The introduction of the CRAB server should simplify the support since most of the possible middleware and DB configuration problems is removed from the user domain 3. The CRAB server still has some instabilities –Would need to improve the stability of the processes on the server and the status monitoring. Depends on The CRAB server needs to be reviewed for security –Will probably be done in the framework of WLCG

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 15 Main issues to be addressed 2/3 Infrastructure (INFN) 5. The Tier-2s are understaffed –Unavailability of a single person at a site may have dramatic effects 6. Optimize the number of servers vs their size –More institutes are installing servers CMS-wide, but they need to be properly supported by expert people 7. The amount of resources is small w.r.t. other CMS sites and in most cases INFN Tier-2s are under the nominal Tier-2 size –Need to recover in 2010 with the new allocations 8. No resources for interactive operations –Not defined who provides interactive login to users at institutes not hosting a Tier-2. Now relying on local funds.

Claudio Grandi INFN Bologna 13 maggio 2009 Workshop congiunto CCR e INFNGrid 16 Main issues to be addressed 3/3 Procedures 8. At INFN not yet completed the association of local institutes with the supporting Tier-2 ( /store/user ) –Needs to speed up Remote stage out is fragile –Try to use asynchronous transfers (Phedex). Development may be needed