N. De Filippis - LLR-Ecole Polytechnique

Slides:



Advertisements
Similar presentations
1 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna A. Fanfani University of Bologna MC Production System & DM catalogue.
Advertisements

31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
1 CRAB Tutorial 19/02/2009 CERN F.Fanzago CRAB tutorial 19/02/2009 Marco Calloni CERN – Milano Bicocca Federica Fanzago INFN Padova.
Introduction to CMS computing CMS for summer students 7/7/09 Oliver Gutsche, Fermilab.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
A tool to enable CMS Distributed Analysis
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
Operational Experience with CMS Tier-2 Sites I. González Caballero (Universidad de Oviedo) for the CMS Collaboration.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
June 24-25, 2008 Regional Grid Training, University of Belgrade, Serbia Introduction to gLite gLite Basic Services Antun Balaž SCL, Institute of Physics.
Stefano Belforte INFN Trieste 1 CMS Simulation at Tier2 June 12, 2006 Simulation (Monte Carlo) Production for CMS Stefano Belforte WLCG-Tier2 workshop.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
DBS/DLS Data Management and Discovery Lee Lueking 3 December, 2006 Asia and EU-Grid Workshop 1-4 December, 2006.
Jean-Roch Vlimant, CERN Physics Performance and Dataset Project Physics Data & MC Validation Group McM : The Evolution of PREP. The CMS tool for Monte-Carlo.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks CRAB: the CMS tool to allow data analysis.
INFSO-RI Enabling Grids for E-sciencE CRAB: a tool for CMS distributed analysis in grid environment Federica Fanzago INFN PADOVA.
CMS Computing Model summary UKI Monthly Operations Meeting Olivier van der Aa.
1 Andrea Sciabà CERN The commissioning of CMS computing centres in the WLCG Grid ACAT November 2008 Erice, Italy Andrea Sciabà S. Belforte, A.
EGEE is a project funded by the European Union under contract IST Enabling bioinformatics applications to.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Finding Data in ATLAS. May 22, 2009Jack Cranshaw (ANL)2 Starting Point Questions What is the latest reprocessing of cosmics? Are there are any AOD produced.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
Operational Experience with CMS Tier-2 Sites I. González Caballero (Universidad de Oviedo) for the CMS Collaboration.
1 Tutorial:Initiation a l’Utilisation de la Grille EGEE/LCG, June 5-6 N. De Filippis CMS tools for distributed analysis N. De Filippis - LLR-Ecole Polytechnique.
Gestion des jobs grille CMS and Alice Artem Trunov CMS and Alice support.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
G. Russo, D. Del Prete, S. Pardi Kick Off Meeting - Isola d'Elba, 2011 May 29th–June 01th A proposal for distributed computing monitoring for SuperB G.
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Job Priorities and Resource sharing in CMS A. Sciabà ECGI meeting on job priorities 15 May 2006.
CMS data access Artem Trunov. CMS site roles Tier0 –Initial reconstruction –Archive RAW + REC from first reconstruction –Analysis, detector studies, etc.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
Claudio Grandi INFN Bologna Workshop congiunto CCR e INFNGrid 13 maggio 2009 Le strategie per l’analisi nell’esperimento CMS Claudio Grandi (INFN Bologna)
THE ATLAS COMPUTING MODEL Sahal Yacoob UKZN On behalf of the ATLAS collaboration.
Real Time Fake Analysis at PIC
Overview of the Belle II computing
Jan 12, 2005 Improving CMS data transfers among its distributed Computing Facilities N. Magini CERN IT-ES-VOS, Geneva, Switzerland J. Flix Port d'Informació.
BOSS: the CMS interface for job summission, monitoring and bookkeeping
INFN-GRID Workshop Bari, October, 26, 2004
CRAB and local batch submission
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Job workflow Pre production operations:
Artem Trunov and EKP team EPK – Uni Karlsruhe
Simulation use cases for T2 in ALICE
PROCESS - H2020 Project Work Package WP6 JRA3
R. Graciani for LHCb Mumbay, Feb 2006
ExaO: Software Defined Data Distribution for Exascale Sciences
The CMS Beijing Site: Status and Application
Production Manager Tools (New Architecture)
The LHCb Computing Data Challenge DC06
Presentation transcript:

N. De Filippis - LLR-Ecole Polytechnique CMS tools for distributed analysis N. De Filippis - LLR-Ecole Polytechnique

CMS tools tutorial organization morning session: overview of CMS tools concepts of the CMS computing/analysis model overview of the analysis workflow and tools overview of the analysis job monitoring system afternoon session: practical examples demonstration of data discovery demonstration of job submission job monitoring and trouble-shooting

CMS tools: overview session

CMS dataflow and workflow Data are collected, filtered online, stored, reconstructed with HLT information at Tier-0 and registered in Data Bookeeping Service (DBS) at CERN. CERN Computer Centre Tier 0 DBS RECO data are moved from Tier-0 to Tier 1 via PhEDEx PhEDEx PhEDEx PhEDEx France Regional Centre (IN2P3 Italy Regional Centre (CNAF) FermiLab Tier 1 Data are filtered (in a reduced AOD format) at Tier-1 according to the physics analysis group selection for skimming; skim output are shipped to Tier-2 via PhEDEx PhEDEx PhEDEx PhEDEx GRIF LNL Rome Wisconsin Tier 2 LLR Data are analysed at Tier-2 and Tier-3 by physics group users and published in a DBS instance dedicated to the physics group. Tier 3

CMS analysis model (CTDR) In the computing model CMS stated that the analysis resources are: Central Analysis Facility (CAF) at CERN: intended for specific varieties of analysis with requirements of low latency access to the data (data are on disk pool of CASTOR) CAF is a very large resource but policy decisions about the users and the use of the resource processing at CAF: calibration and alignment, few physics analyses Tier-2 resources: two groups of analysis users to be supported in Tier-2 support for analysis groups/specific analyses support for local communities 1. is addressed by using the distributed analysis system 2. is addressed via local batch queues/access and distributed systems Tier-3 resources: local and interactive analysis, private resources

Tier-2 resources A nominal Tier-2 should be: 0.9MSI2k of computing power, 200TB of disk and 1Gb/s of WAN. that means several hundred batch slots and disk for large skim samples A reasonably large fraction of the nominal is devoted to analysis group activities and the remainder is assignable to the local community Proposal under discussion is ~50% of nominal processing resources for simulation, ~40% of the nominal resources for analysis groups (specific and well organized analyses) , and the remainder for local users 10% of the local storage for simulation, 60% for analysis groups and 30% for local communities • for 2008 lower guideline of 60TB of disk and 0.4MSI2k for analysis groups

Distributed analysis in Tier-2 via virtual organization (VO) authorized access via data discovery (DBS) via job analysis builder (CRAB) via data, job and resources monitoring (DashBoard) via automatic storage allocation for physics groups and local user data via data movement tools (PhEDeX)

CMS Data Discovery The Dataset Bookkeeping System (DBS) provides the means to define, discover and use CMS event data. The main features that DBS provides are: Data Description: keeps dataset definition along with attributes characterising the dataset, the type of content resulting from a degree of processing applied to the data (RAW, RECO, etc) The DBS also provides information regarding the “provenance” of the data it describes. Data Discovery: stores information about (real and simulated) CMS data in a queryable format. The supported queries allow users to discover available data and how they are organized (logically) in term of packaging units (files and file-blocks). Answers the question: “Which data exist?” Data location: provides the means to locate replicas of data in the distributed computing system by providing the names of Storage Elements of sites hosting the data. Answers the question “Where data exist?”

CMS Data Discovery: DBS Data discovery page: https://cmsweb.cern.ch/dbs_discovery/_navigator?userMode=user A sample of data is identified by a string called: datasetpath /Primarydataset/Processeddataset/DataTier Ex: /Njet_2j_80_140-alpgen/CMSSW_1_6_7-CSA07-1200571375/RECO Primary dataset: name that describes the physics channel Processed dataset: name that describe the kind of processing applied Data Tier: describes the kind of event information stored from each step in the simulation and reconstruction chain. Examples: RAW and RECO, and for MC, GEN, SIM and DIGI. File-related concepts: Logical File Name (LFN): a site-independent name for a file. It doesn't contain either the actual protocol used to read the file or any of the site-specific information about the place where it is located. Physical File Name (PFN): site-dependent name for a file to allow local access to a file at a site. Logical file names are mapped into the physical file names via the local trivial file catalog, (TFC)

CMS Data Discovery (2) Logical file names Data discovery page: https://cmsweb.cern.ch/dbs_discovery/_navigator?userMode=user Logical file names Site: SE name click on this

CRAB (CMS remote analysis builder) CRAB is a user oriented tool for: job preparation, submission, (basic) monitoring of CMS analysis jobs in remove sites by using the GRID infrastructure (EGEE and OSG) Features: User Settings provided via a configuration file (dataset, data type) Data discovery querying DBS for remote sites Job splitting automatic (n. of jobs or events per job to be provided) Jobs will run where the data are GRID details mostly hidden to the user status monitoring, job tracking and output management of jobs publishing of the output Use cases supported: Official and private code analysis of official CMS data But also private production and skimming The aim of CRAB is to hide as much as possible the grid complexity to the user

The user analysis basic model

The user analysis workflow The user provides: Dataset (runs,#event,..) taken by DBS Private CMSSW code UI DataSet Catalogue DBS CRAB discoveries data and sites hosting them by querying DBS CRAB Job submission tool CRAB prepares, splits and submits jobs to the Resource Broker/WMS Workload Management System The RB/WMS sends jobs at sites hosting the data provided the CMS software was installed Resource Broker (RB/WMS) CRAB retrieves automatically logs and the output files of the jobs; it’s possible to store output files into SE (best solution) CMSSW Computing Element CRAB can publish the output of the jobs in DBS to make output data available officially for subsequent processing. Worker node Storage Element

CRAB standalone and server

CRAB documentation and support CRAB home page: http://cmsdoc.cern.ch/cms/ccs/wm/www/Crab/ CRAB twiki: https://twiki.cern.ch/twiki/bin/view/CMS/CRAB

CRAB frequently asked questions https://twiki.cern.ch/twiki/bin/view/CMS/CrabFaq

CMS job monitoring: Dashboard Most of the CMS Job Submission Systems , including Crab, are instrumented to send monitoring information to the CMS Dashboard. The CMS Job Submission Systems Dashboard collects information from the Grid Monitoring systems. Monitoring data is stored in the central data base and there is a web interface running on top of it and allowing CMS users to follow the progress of their jobs. Web interface at link: http://arda-dashboard.cern.ch/cms/

Monitor of analysis jobs (last month) sites Users dataset

Monitor of analysis tasks Choose your identity in the "Select a User" window, select the time window to define the tasks submitted during a given time range, you should get at the screen the list of all your tasks submitted over the time range you have chosen.

Monitor of site availability for analysis http://lxarda16.cern.ch/dashboard/request.py/samvisualization Simple analysis test is run continuosly on any site to check the availability for analysis

Mostly for storage problems at sites Statistics of failures All Tiers: ~400 Kjobs: Failure 20% Mostly for storage problems at sites Tier2 only (32%): ~130 Kjobs: Failure 20% https://twiki.cern.ch/twiki/bin/view/CMS/JobExitCodes

Storage management at Tier-2 CMS users/physics groups can produce and store large quantities of data: due to re-processing of skim due to common preselection processing and iterated reprocessing due to private productions (especially fast simulation) due to end-user analysis ROOT-ples  care on how to manage the utilization of the storage Tier-2 disk-based storage will be fixed size and so care is needed to ensure every site can meet their obligations with respect to the collaboration and analysis users are treated fairly: user quota, policy are under definition physics analysis data namespace user data namespace Official Physics groups Private user

Physics analysis data namespace

User analysis data namespace

Request a replica of a dataset in one site CMS data movement http://cmsdoc.cern.ch/cms/aprom/phedex/ request samples to be transferred. Requests to be approved by site administrators according to policies of physics groups Request a replica of a dataset in one site

Glossary

Demonstration about CMS analysis workflow in the afternoon