Introduction to CMS computing CMS for summer students 7/7/09 Oliver Gutsche, Fermilab.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
1 User Analysis Workgroup Update  All four experiments gave input by mid December  ALICE by document and links  Very independent.
Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
Oliver Gutsche - CMS / Fermilab Analyzing Millions of Gigabyte of LHC Data for CMS - Discover the Higgs on OSG.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
A tool to enable CMS Distributed Analysis
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
WLCG/8 July 2010/MCSawley WAN area transfers and networking: a predictive model for CMS WLCG Workshop, July 7-9, 2010 Marie-Christine Sawley, ETH Zurich.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
Claudio Grandi INFN Bologna CMS Operations Update Ian Fisk, Claudio Grandi 1.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
1 Kittikul Kovitanggoon*, Burin Asavapibhop, Narumon Suwonjandee, Gurpreet Singh Chulalongkorn University, Thailand July 23, 2015 Workshop on e-Science.
LHC Computing Review - Resources ATLAS Resource Issues John Huth Harvard University.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
Introduction to CMS computing J-Term IV 8/3/09 Oliver Gutsche, Fermilab.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
7April 2000F Harris LHCb Software Workshop 1 LHCb planning on EU GRID activities (for discussion) F Harris.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
Meeting, 5/12/06 CMS T1/T2 Estimates à CMS perspective: n Part of a wider process of resource estimation n Top-down Computing.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
CERN – IT Department CH-1211 Genève 23 Switzerland t Working with Large Data Sets Tim Smith CERN/IT Open Access and Research Data Session.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
CMS Computing Model Simulation Stephen Gowdy/FNAL 30th April 2015CMS Computing Model Simulation1.
The CMS CERN Analysis Facility (CAF) Peter Kreuzer (RWTH Aachen) - Stephen Gowdy (CERN), Jose Afonso Sanches (UERJ Brazil) on behalf.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
Claudio Grandi INFN Bologna CERN - WLCG Workshop 13 November 2008 CMS - Plan for shutdown and data-taking preparation Claudio Grandi Outline: Global Runs.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks CRAB: the CMS tool to allow data analysis.
INFSO-RI Enabling Grids for E-sciencE CRAB: a tool for CMS distributed analysis in grid environment Federica Fanzago INFN PADOVA.
The ATLAS Computing Model and USATLAS Tier-2/Tier-3 Meeting Shawn McKee University of Michigan Joint Techs, FNAL July 16 th, 2007.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
1 Andrea Sciabà CERN The commissioning of CMS computing centres in the WLCG Grid ACAT November 2008 Erice, Italy Andrea Sciabà S. Belforte, A.
Handling of T1D0 in CCRC’08 Tier-0 data handling Tier-1 data handling Experiment data handling Reprocessing Recalling files from tape Tier-0 data handling,
Maria Girone, CERN CMS Experiment Status, Run II Plans, & Federated Requirements Maria Girone, CERN XrootD Workshop, January 27, 2015.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
GDB, 07/06/06 CMS Centre Roles à CMS data hierarchy: n RAW (1.5/2MB) -> RECO (0.2/0.4MB) -> AOD (50kB)-> TAG à Tier-0 role: n First-pass.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Computing Model José M. Hernández CIEMAT, Madrid On behalf of the CMS Collaboration XV International Conference on Computing in High Energy and Nuclear.
Operational Experience with CMS Tier-2 Sites I. González Caballero (Universidad de Oviedo) for the CMS Collaboration.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
1 June 11/Ian Fisk CMS Model and the Network Ian Fisk.
Oct 16, 2009T.Kurca Grilles France1 CMS Data Distribution Tibor Kurča Institut de Physique Nucléaire de Lyon Journées “Grilles France” October 16, 2009.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
Grid technologies for large-scale projects N. S. Astakhov, A. S. Baginyan, S. D. Belov, A. G. Dolbilov, A. O. Golunov, I. N. Gorbunov, N. I. Gromova, I.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
Artem Trunov and EKP team EPK – Uni Karlsruhe
N. De Filippis - LLR-Ecole Polytechnique
Gridifying the LHCb Monte Carlo production system
LHCb thinking on Regional Centres and Related activities (GRIDs)
The LHCb Computing Data Challenge DC06
Presentation transcript:

Introduction to CMS computing CMS for summer students 7/7/09 Oliver Gutsche, Fermilab

7/7/09Oliver Gutsche - Introduction to CMS Computing 2 Basic of computing for HEP Particle physics is an event based processing problem The detector records collisions and stores the detector signals per collision This is what we call an event After the event recording, a lot of computing is needed to Store the data permanently Process the data to reconstruct event properties Access different forms of data to analyze events and extract physics results Monte Carlo simulation (MC) is also event based In addition to the processing, a lot of computing is needed to simulate a collision in the first place Processing, simulation and analysis can be performed with very high grade of parallelism In general, every event can be handled in parallel

7/7/09Oliver Gutsche - Introduction to CMS Computing 3 CMS in a nutshell (current planning basis) Data recording rate: 300 Hz Size per event: RAW: 1.5 MB (RAW+SIM 2 MB) RECO: 0.5 MB AOD: 0.1 MB Processing power per event: Simulation (including reconstruction) : 1000 HS HS06 = 1 event simulated and reconstructed in 100s on 3 GHz core Reconstruction: 100 HS HS06 = 1 event reconstructed in 10s on 3 GHz core DATA

7/7/09Oliver Gutsche - Introduction to CMS Computing /2010 We all are eagerly awaiting the start of data taking in 2009 Following estimates for the data taking year have been determined: “2009-run”: Oct’09 - Mar’10: 726M events “2010-run”: Apr’10 - Sep’10: 1555M events Translates to: 3.42 PB (yes, PetaByte) RAW data, 1.14 PB RECO data One single 3 HGz core would take ~100 years to reconstruct all of it This only refers to collision data, MC is not included

7/7/09Oliver Gutsche - Introduction to CMS Computing 5 Parallel processing We exploit the parallel nature of event processing by using multiple standard computers in parallel Processes only need access to input data and produce output data No communication between different computers during processing of events needed For the computing needs of CMS, we need many 10,000 computers (cores) Very interesting and challenging computational problem which governs the way we are doing computing

7/7/09Oliver Gutsche - Introduction to CMS Computing 6 From Single COmputer to the GRID 1. User is running processes on individual machines Has to log in to every machine, start the processes/jobs by hand Does scale for 1-3 users, but is very work intensive 2. Batch system: Scheduling system with access to many cores (farm) User adds his jobs to the queue Scheduling system starts jobs on one of the cores System can prioritize job execution and can impose fair share mechanism to support many users ➞ scales to many users Limited by farm size Batch system

7/7/09Oliver Gutsche - Introduction to CMS Computing 7 From Single COmputer to the GRID 3. The GRID, an interconnected network of batch farms Why not a single huge batch farm for CMS: Leverage local facilities, national/institutional computing resources and expertise Computing infrastructure for science is publicly funded, tax payers money stays in their respective country Reuse existing infrastructure and share with other disciplines Follow the international nature of the CMS collaboration Too big to put in one place even if we wanted to Several smaller farms are much easier to manage than one large one

7/7/09Oliver Gutsche - Introduction to CMS Computing 8 The GRID Transparent access to all distributed resources (like a very big batch system) Authentication via GRID certificates: User accounts of all users cannot be distributed to all GRID sites GRID certificates and derived short-living GRID proxies authenticate a user globally At the site, the GRID proxy is mapped to a generic local user account CMS uses the Virtual Organization Membership Service (VOMS) on top of GRID proxies to have a more fine grained role based user prioritization (e.g. “production role” gets priority) CMS uses following GRIDs: EGEE (European GRID), OSG (American GRID), NorduGRID Important for users in CMS: get your GRID certificate!

7/7/09Oliver Gutsche - Introduction to CMS Computing 9 Storage CMS will produce an enormously large volume of data Cheapest and most reliable technique to store these volumes of data: tape Simple technology with very high reliability (tape robot feeding tapes stored in large shelf systems to tape drives) Slow in writing and access (remember the old cassette players?) Storage capacity: 10 to 100 of PB Approach to overcome slow access: disk caches Access to data is not handled directly, disk cache systems are sitting between the processing farm and the tape system Disk cache system organize large amounts of disks into larger entities and provide access to the cached files, take care of deleting not or very rarely used files Systems have > 1000 individual disks Some disk cache systems you might come across: Castor, dCache, DPM, Storm, Hadoop Storage capacity: ~1 to 10 PB

7/7/09Oliver Gutsche - Introduction to CMS Computing 10 Computing for CMS How does CMS organize its computing to provide sufficient computing power to handle all data and MC?

7/7/09Oliver Gutsche - Introduction to CMS Computing 11 CMS Computing model: tiers T0T0 T1: USA T1: Italy T1: France T1: Spain T1: Germany T1: Taiwan T1: UK T2T2T2T2T2T2T2T2 T2T2 T2T2 T2T2 T2T2 T2T2 T2T2 T2T2T2T2T2T2T2T2T2T2T2T2 T2T2 T2T2 T2T2 T2T2 T2T2 T2T2

7/7/09Oliver Gutsche - Introduction to CMS Computing 12 CMS Computing model Tier 0 (T0) at CERN (20% of all CMS computing resources) Record and prompt reconstruct collision data Calculate condition and alignment constants Store data on tape (only archival copy, no access) Only central processing, no user access Tier 1 (T1): regional centers in 7 countries (40% of all CMS computing resources) Store data fraction on tape (served copy) Every T1 site gets a fraction of the data according to its respective size Archive fraction of produced MC on tape Skim data to reduce data size and make data more easily handleable Rereconstruct data with newer software and conditions/alignment constants Only central processing, no user access Tier 2 (T2): local computing centers at Universities and Laboratories (40% of all CMS computing resources) Simulate MC events User access to data for analysis

7/7/09Oliver Gutsche - Introduction to CMS Computing 13 CMS distributed computing MODEL “Data driven” computing model Data and MC samples are distributed centrally Jobs (processing, analysis) “go” to the data Requires very fast network connections between the different centers: T0 ➞ T1: handled via the LHC-OPN (Optical Private Network) consisting of dedicated 10 Gbit/s network links Distributes the recorded data for storage on tape at T1 sites T1 ➞ T1: also handled via the OPN Redistribute parts of the data produced during rereconstruction T1 ➞ T2: handled via national high speed network links Transfer datasets for analysis to T2 sites T2 ➞ T1: handled via national high speed network links Transfer produced MC to T1 for storage on tape

7/7/09Oliver Gutsche - Introduction to CMS Computing 14 T0 Online system of the detector records event data and stores it in binary files (streamer files) There are 2 processing paths in the T0: Prompt: latency of several days The T0 repacks the binary files into ROOT files and splits the events into Primary Datasets according to trigger selections and stores them on tape This is called the RAW data tier The T0 then reconstructs the RAW data for the first time (Prompt Reconstruction) and stores the output on tape as well This is called the RECO data tier During reconstruction, a subset of information sufficient to perform 90% of all physics analysis is extracted and stored separately This is called the AOD (Analysis Object Data) data tier Also special alignment and calibration datasets are produced and copied directly to the CERN Analysis Facility (CAF) All RAW, RECO and AOD data is then transferred to the T1 sites for storage on tape Express: latency of 1 hour All the steps above combined into a single process run on 10% of all events selected from all the recorded data Output is copied to the CAF for express alignment and calibration workflows and prompt feedback by physics analysis

7/7/09Oliver Gutsche - Introduction to CMS Computing 15 T1 All T1 centers store a fraction of the recorded data on tape This is called the custodial data fraction MC produced at the T2 sites is archived on tape here as well The T1 processing resources are used for Skimming of data to reduce the data size by physics pre-selections to make the datasets more easily handleable by the physicists Rereconstruction of the data with new software versions and updated alignment and calibrations The T1 sites serve data to the T2 sites for analysis Users, physics groups or also central operations can request transfers of datasets Currently up to 10,000 jobs in parallel on the T1 sites

7/7/09Oliver Gutsche - Introduction to CMS Computing 16 T T2 sites serve 2000 physicists and provide access to data and MC samples for analysis Each CMS physicist can access data at all the T2 sites Still there is some regional association between physicist and T2 site to support local resources for storage of user files There is also association between physics groups (top, higgs, …) and T2 sites to organize data and MC sample placement A typical T2 sites has ~800 batch slots and 300 TB of disk space T2 sites don’t have tape Half of the resources of all T2 sites are reserved for MC production which is handled centrally

7/7/09Oliver Gutsche - Introduction to CMS Computing 17 USER How does the user interact with the CMS computing infrastructure?

7/7/09Oliver Gutsche - Introduction to CMS Computing 18 Transfer tool: phedex PhEDEx is CMS’ tool to request and manage data transfers Every user can request the transfer of a data sample to a T2 site for analysis Every T2 site (also the T1 sites and the T0) have data managers which approve or disapprove transfer requests according to global policies and available storage space

7/7/09Oliver Gutsche - Introduction to CMS Computing 19 Dataset Bookkeeping system (DBS) DBS handles to bookkeeping of datasets A dataset name is composed of: / / / Primary dataset name: specifies the physics content of the sample Processed dataset name: specifies the processing conditions and data taking or MC production period Data tier: specifies the format of content of the files (RAW, RECO, AOD, … ) Primary tool to look up and discovery datasets and their location on the T2 level for your analysis

7/7/09Oliver Gutsche - Introduction to CMS Computing 20 GRID submission tool: CRAB CMS Remote Analysis Builder Enables every user to send her/his analysis code to the T2 sites to process stored data and MC samples Represents a wrapper to the GRID tools used to execute jobs on the GRID

7/7/09Oliver Gutsche - Introduction to CMS Computing 21 Summary CMS computing needs are significant CMS expects to record in 2009/2010 alone More than 6 PetaByte of collision data All this data has to be processed and accessed via the distributed tiered computing infrastructure Especially for the average user: Documentation: Getting help via Hypernews: look for most appropriate forum, send a fully detailed report (too much is better than too little). Dont be shy to ask questions!