1 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna A. Fanfani University of Bologna MC Production System & DM catalogue.

Slides:



Advertisements
Similar presentations
Metadata Progress GridPP18 20 March 2007 Mike Kenyon.
Advertisements

Data Management Expert Panel - WP2. WP2 Overview.
INFSO-RI Enabling Grids for E-sciencE Workload Management System and Job Description Language.
1 CRAB Tutorial 19/02/2009 CERN F.Fanzago CRAB tutorial 19/02/2009 Marco Calloni CERN – Milano Bicocca Federica Fanzago INFN Padova.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL March 25, 2003 CHEP 2003 Data Analysis Environment and Visualization.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Workload Management Massimo Sgaravatto INFN Padova.
A tool to enable CMS Distributed Analysis
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Stuart Wakefield Imperial College London1 How (and why) HEP uses the Grid.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
GRACE Project IST EGAAP meeting – Den Haag, 25/11/2004 Giuseppe Sisto – Telecom Italia Lab.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
INFSO-RI Enabling Grids for E-sciencE Project Gridification: the UNOSAT experience Patricia Méndez Lorenzo CERN (IT-PSS/ED) CERN,
Nicola De Filippis CMS Italia, Napoli, Feb p. 1 Produzioni MC ai Tiers CMS nel 2007: prospettive CMS-wide e contributo italiano Università,
The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within.
1 M. Paganoni, HCP2007 Computing tools and analysis architectures: the CMS computing strategy M. Paganoni HCP2007 La Biodola, 23/5/2007.
Stuart Wakefield Imperial College London Evolution of BOSS, a tool for job submission and tracking W. Bacchi, G. Codispoti, C. Grandi, INFN Bologna D.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
Giuseppe Codispoti INFN - Bologna Egee User ForumMarch 2th BOSS: the CMS interface for job summission, monitoring and bookkeeping W. Bacchi, P.
June 24-25, 2008 Regional Grid Training, University of Belgrade, Serbia Introduction to gLite gLite Basic Services Antun Balaž SCL, Institute of Physics.
Job Life Cycle Management Libraries for CMS Workflow Management Projects Stuart Wakefield on behalf of CMS DMWM group Thanks to Frank van Lingen for the.
Getting started DIRAC Project. Outline  DIRAC information system  Documentation sources  DIRAC users and groups  Registration with DIRAC  Getting.
Stefano Belforte INFN Trieste 1 Middleware February 14, 2007 Resource Broker, gLite etc. CMS vs. middleware.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
DBS/DLS Data Management and Discovery Lee Lueking 3 December, 2006 Asia and EU-Grid Workshop 1-4 December, 2006.
Glite. Architecture Applications have access both to Higher-level Grid Services and to Foundation Grid Middleware Higher-Level Grid Services are supposed.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Use of the gLite-WMS in CMS for production and analysis Giuseppe Codispoti On behalf of the CMS Offline and Computing.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks CRAB: the CMS tool to allow data analysis.
The GridPP DIRAC project DIRAC for non-LHC communities.
INFSO-RI Enabling Grids for E-sciencE CRAB: a tool for CMS distributed analysis in grid environment Federica Fanzago INFN PADOVA.
Daniele Spiga PerugiaCMS Italia 14 Feb ’07 Napoli1 CRAB status and next evolution Daniele Spiga University & INFN Perugia On behalf of CRAB Team.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Enabling Grids for E-sciencE CMS/ARDA activity within the CMS distributed system Julia Andreeva, CERN On behalf of ARDA group CHEP06.
VOX Project Status T. Levshina. 5/7/2003LCG SEC meetings2 Goals, team and collaborators Purpose: To facilitate the remote participation of US based physicists.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
The GridPP DIRAC project DIRAC for non-LHC communities.
WMS baseline issues in Atlas Miguel Branco Alessandro De Salvo Outline  The Atlas Production System  WMS baseline issues in Atlas.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES The Common Solutions Strategy of the Experiment Support group.
Gestion des jobs grille CMS and Alice Artem Trunov CMS and Alice support.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Services for Distributed e-Infrastructure Access Tiziana Ferrari on behalf.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Job Priorities and Resource sharing in CMS A. Sciabà ECGI meeting on job priorities 15 May 2006.
Claudio Grandi INFN Bologna Workshop congiunto CCR e INFNGrid 13 maggio 2009 Le strategie per l’analisi nell’esperimento CMS Claudio Grandi (INFN Bologna)
Workload Management Workpackage
Overview of the Belle II computing
Practical: The Information Systems
POW MND section.
BOSS: the CMS interface for job summission, monitoring and bookkeeping
BOSS: the CMS interface for job summission, monitoring and bookkeeping
BOSS: the CMS interface for job summission, monitoring and bookkeeping
Short update on the latest gLite status
N. De Filippis - LLR-Ecole Polytechnique
EGEE Middleware: gLite Information Systems (IS)
Status and plans for bookkeeping system and production tools
Presentation transcript:

1 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna A. Fanfani University of Bologna MC Production System & DM catalogue

2 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna Production System Overview ProdRequest ProdMgr ProdAgent LCG/EGEE Resource OSG Resource User Request Get Work Jobs Report Progress User Interface to create requests Manage request Allocate work to PA when PA request it Tracks the global completion of the task Ask for work Convert work into processing jobs Create, submit, track jobs Manage the merge, failures, resubmit, local cataloguing, etc.. ProdMgr ProdRequest Under High development Basic chain PR PM PA works ProdAgent In production since the summer Aim at automating as much as possible, easy maintenance Various Grid/batch Middleware to support

3 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna ProdAgent Processing Workflow ProdAgent Grid WMS Tier-2 Tier-1 Tier-2 Processing Processing jobs sent to sites Output data left in local SE Report back to ProdAgent Data management cataloguing (registration in local DBS/DLS) Failed jobs handled automatically Small output file from Processing job SE Local DBS/DLS

4 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna ProdAgent Merge Workflow ProdAgent Grid WMS Tier-2 Tier-1 Tier-2 Merging Large output file from Merge job SE PhEDEx PhEDEx transfer invoked by PA Merge data at site Watch DBS/DLS for produced unmerged data send merge job at sites hosting data Transfer data PhEDEx injection Local DBS/DLS

5 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna ProdAgent Architecture Core MySQL DataBase Python API Core Services Work split into atomic Python Components Asynchronous Publish/Subscribe model for inter-component communications Simple API to communicate between components easy to add new functionality and build on existing features Persistent state recorded in DB ProdAgent core

6 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna Production Agent components ProdAgent Core DBS/DLS interface PhEDEx interface Job Tracking Job Creator Local DBS Local DLS PHEDEX BOSS DB BOSS submit Merge Sensor watchs LCG RB/gLite WMS ProdMgr interface Job Submitter Job Cleanup Error handler Retrieve work Job Queue Resource Monitor workflow Merge Accountant

7 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna Monitoring status of each component Status of each component overview of the current job status Overview of current job status mc-physval-120-ZToMuMu-StartUpLumiPU PA level monitoring for operators (developed by Bari team + Carlos)

8 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna Prod system Status & Plan ProdAgent implemented and deployed operationally since summer CSA06 pre-production of 66Mevents + organized skimming run at Tier1s PhysVal+HLT samples with CMSSW12x, see Nicolas talk Focus is now on automation to reduce manual work for operators i.e. automatic block management and PhEDEX injections and on performance to make it more scalable, more robust Bulk creation & bulk submission with gLite (with LCG RB 2000jobs/day per PA) Deployment of ProdRequest/ProdMgr/ProdAgent system The production teams will no longer have to inject workflows taking them from Twiki pages The production coordinator will assign work to teams with given priority via ProdManager DBS-2 integration Alpgen integration True collaborative development effort: Dave Evans, Frank Van Lingen, Giulio Eulisse (US) Carlos Kavka, Alessandra Fanfani, William Bacchi,Giuseppe Codispoti, contribution from Bari team (IT)

9 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna Data Management catalogues: DBS The Dataset Bookkeeping System (DBS) provides the means to define, discover and use CMS event data First version deployed for CSA06, including data discovery browser Development for 2nd generation (DBS-2) Prepare the system for describing real data Added info like run, luminosity sections, primary dataset description Preliminary support for Analysis dataset A subset of a Processed Dataset representing a coherent sample for physics analysis More functionalities for browsing data discovery Deployable with Oracle at CERN for Global DBS Deployable with MySQL too to be used as local scope DBS Under integration with CRAB, ProdAgent, PhEDEx, MTCC data

10 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna Data Management catalogues: DLS The Data Location Service (DLS) provides the means to locate replicas of data in the distributed computing system The DBS knows how datasets are organized in term of file-blocks The Data Location Service (DLS) maps file-blocks to storage elements (SEs) DLS based on LCG LFC used for CSA06: some drawbacks (performance issues for reverse lookup, i.e. data discovery), some advantages (production service mantained by LCG, VOMS authentication-authorization, DLI) No server-side work needed by CMS Serving us right so far Evaluating to have DLS on the same server as DBS Decide based on CMS use case Still keeping DLS API functionalities Add support for the Resource Broker to talk to directly (via DLI)