Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.

Slides:



Advertisements
Similar presentations
1 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna A. Fanfani University of Bologna MC Production System & DM catalogue.
Advertisements

31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
Introduction to CMS computing CMS for summer students 7/7/09 Oliver Gutsche, Fermilab.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
ATLAS computing in Geneva Szymon Gadomski, NDGF meeting, September 2009 S. Gadomski, ”ATLAS computing in Geneva", NDGF, Sept 091 the Geneva ATLAS Tier-3.
A tool to enable CMS Distributed Analysis
1 INDIACMS-TIFR TIER-2 Grid Status Report IndiaCMS Meeting, Sep 27-28, 2007 Delhi University, India.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Claudio Grandi INFN Bologna CMS Operations Update Ian Fisk, Claudio Grandi 1.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
How to Install and Use the DQ2 User Tools US ATLAS Tier2 workshop at IU June 20, Bloomington, IN Marco Mambelli University of Chicago.
Cosener’s House – 30 th Jan’031 LHCb Progress & Plans Nick Brook University of Bristol News & User Plans Technical Progress Review of deliverables.
11/30/2007 Overview of operations at CC-IN2P3 Exploitation team Reported by Philippe Olivero.
Stefano Belforte INFN Trieste 1 CMS DM and Monitor CMS Data Management and CMS Monitoring (emphasis on T2 perspective) CMS data organization Data names,
Costin Grigoras ALICE Offline. In the period of steady LHC operation, The Grid usage is constant and high and, as foreseen, is used for massive RAW and.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
São Paulo Regional Analysis Center SPRACE Status Report 22/Aug/2006 SPRACE Status Report 22/Aug/2006.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
UMD TIER-3 EXPERIENCES Malina Kirn October 23, 2008 UMD T3 experiences 1.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
LOGO Development of the distributed computing system for the MPD at the NICA collider, analytical estimations Mathematical Modeling and Computational Physics.
Stefano Belforte INFN Trieste 1 CMS Simulation at Tier2 June 12, 2006 Simulation (Monte Carlo) Production for CMS Stefano Belforte WLCG-Tier2 workshop.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
DBS/DLS Data Management and Discovery Lee Lueking 3 December, 2006 Asia and EU-Grid Workshop 1-4 December, 2006.
Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.
Tier 3 Status at Panjab V. Bhatnagar, S. Gautam India-CMS Meeting, July 20-21, 2007 BARC, Mumbai Centre of Advanced Study in Physics, Panjab University,
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks CRAB: the CMS tool to allow data analysis.
INFSO-RI Enabling Grids for E-sciencE CRAB: a tool for CMS distributed analysis in grid environment Federica Fanzago INFN PADOVA.
CMS Computing Model summary UKI Monthly Operations Meeting Olivier van der Aa.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Data Analysis w ith PROOF, PQ2, Condor Data Analysis w ith PROOF, PQ2, Condor Neng Xu, Wen Guan, Sau Lan Wu University of Wisconsin-Madison 30-October-09.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
Computing Model José M. Hernández CIEMAT, Madrid On behalf of the CMS Collaboration XV International Conference on Computing in High Energy and Nuclear.
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
ALICE Physics Data Challenge ’05 and LCG Service Challenge 3 Latchezar Betev / ALICE Geneva, 6 April 2005 LCG Storage Management Workshop.
Gestion des jobs grille CMS and Alice Artem Trunov CMS and Alice support.
瑞士 CMS 实验网格计算 站点建设与运行 陈志凌. CMS 实验物理目标 Large Hadron Collider (LHC): – Design values: E CM = 14 TeV Luminosity = cm -2 s -1 Compact Muon Spectrometer.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
1-2 March 2006 P. Capiluppi INFN Tier1 for the LHC Experiments: ALICE, ATLAS, CMS, LHCb.
Simulation Production System
Xiaomei Zhang CMS IHEP Group Meeting December
Overview of the Belle II computing
Chapter 2: System Structures
Model (CMS) T2 setup for end users
Computing Board Report CHIPP Plenary Meeting
Artem Trunov and EKP team EPK – Uni Karlsruhe
Ákos Frohner EGEE'08 September 2008
N. De Filippis - LLR-Ecole Polytechnique
R. Graciani for LHCb Mumbay, Feb 2006
The CMS Beijing Site: Status and Application
Production Manager Tools (New Architecture)
The LHCb Computing Data Challenge DC06
Presentation transcript:

Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009

2 Outline  Intro of CMS Computing Model  Setup of Swiss CMS Tier-3 at PSI  Working on the Swiss CMS Tier-3  Operational Experience

LCG Tier Organization 3 T0 (CERN) Filter farm Raw data custodial Prompt reconstruction 7 T1s Raw data custodial (shared) Re-reconstruction Skimming, calibration ~ 40 T2s Central scheduled MC production Analysis and MC Simulation for all CMS Users ~ Many T3s at institutes Local institutes’ users Final-stage analysis and MC Simulation optimized for users' analysis needs Swiss Tier-2 for ATLAS, CMS, LHCb …. Tier-3 for Swiss CMS community

CMS Data Organization Physicist’s View Event collection Dataset A set of event collections that would naturally be grouped for analysis To process Events: Find Transfer Access System View Files File Blocks Files grouped into blocks of reasonable size or logical content. To operate files: files blocks Stored in Grid Transfer and Access files in different storage system Manage replicas 4

CMS Data Organization Physicist’s View Find “What data exist?” “Where are data located?” Transfer Access CMS Data management Data Bookkeeping System Standardized and queryable info of event data mapping from event collections to files/file blocks Data Location Service maps file blocks to locations PhEDEx Data Transfer and Placement System LCG commands SRM and POSIX-I/O 5 Map Physicist and system views

CMS Tier-3 Local Site Globe CMS Data Management Service CMS Analysis work flow 6 Data Book- keeping DB DBS Data Book- keeping DB DBS Data Location DB DLS Data Location DB DLS LHC Grid Computing Tier-3 Local Cluster Tier-3 Local Cluster Tier-3 Storage Element PhEDEx Globe Data Transfer Agents and Database PhEDEx Globe Data Transfer Agents and Database PhEDEx Local Data Transfer Agents PhEDEx Local Data Transfer Agents File Transfer Service CRAB is a Python program to simplify the process of creation and submission of CMS analysis jobs into a grid environment.

Overview of Swiss CMS Tier-3 7  For CMS members of ETHZ, University of Zurich and PSI  Located at PSI  Try to adapt best to the users' analysis needs  running in test mode in October 2008, and in production mode since November 2008  30 registered physicist users  Manager: Dr. Derek Feichtinger, Assistant: Zhiling Chen

Hardware of Swiss CMS Tier-3 8 No. of Work Nodes Processors Cores/NodeTotal Cores 82*Xeon E No. of File Servers Type Space/Node (TB) Total Space (TB) 6SUN X Present Computing Power Present Storage

Layout of Swiss CMS Tier-3 at PSI 9 User Interface CMS VoBox (PhEDEx) CMS VoBox (PhEDEx) Storage Element (t3se01.psi.c h) [dcache admin, dcap, SRM, gridftp, resource info provider] Storage Element (t3se01.psi.c h) [dcache admin, dcap, SRM, gridftp, resource info provider] NFS Server NFS (home and shared software directories: CMSSW, CRAB, Glite) NFS Server NFS (home and shared software directories: CMSSW, CRAB, Glite) DB Server [postgres, pnfs, dcache pnfs cell] DB Server [postgres, pnfs, dcache pnfs cell] Computing Element [Sun Grid Engine ] Computing Element [Sun Grid Engine ] Dispatch/Collection Batch Jobs Submit/retrieve Batch Jobs Access Home/Software Directory Access Local SE : SRM, gridftp, dcap … Submit/retrieve LCG jobs User login Access PheEDEx Central DB Access Remote SE Accessed by LCG CMS Tier-3 at PSI Monitoring [ganglia collector, ganglia web front end ] Monitoring [ganglia collector, ganglia web front end ] Network connectivity: PSI has a 1Gb/s uplink to CSCS.

Setup of Swiss CMS Tier-3  User Interface (8 cores): t3ui01.psi.ch A fully operational LCG UI. It enables users to:  login from outside  Submit/Manage local jobs on the Tier-3 local cluster  Interact with the LCG Grid: Submit Grid jobs, access storage elements, etc.  Interact with AFS, CVS …  Test users’ Jobs  Local batch cluster(8 Work Nodes * 8 Cores):  Batch System: Sun Grid Engine

Setup of Swiss CMS Tier-3 (cont.)  Storage Element (SE): t3se01.psi.ch A fully equipped LCG storage element running a dCache. It allows users to:  Access files by local jobs (dcap, srmcp, gridftp etc.) in Tier-3  Access files (srmcp, gridftp) from other sites  Give users extra space in addition to the space in CSCS Tier-2  NFS Server (for small storage)  Hosts users’ home directories: analysis code, jobs output  Shared software: CMSSW, CRAB, Glite …  Easy to access, but not for huge files Note: If you need large storage space for longer time, you should use SE. 11

Setup of Swiss CMS Tier-3 (cont.)  CMS VoBox (PhEDEx):  Users can order datasets to Tier-3 SE  Admin can manage datasets with PhDEDx  Monitoring:  Status of batch system  Accounting  Worker nodes load  Free storage space  Network Activities …

13 Working on Swiss CMS Tier-3 Before Submit jobs: Order dataset  Check currently stored data sets for the Tier-3 from DBS Data Discovery Page  If the data sets are not stored on Tier-3, Order data sets to T3_CH_PSI by PhEDEx central web page

Work Flow on Tier-3 Working on Swiss CMS Tier-3 Submit and Manage batch jobs CRAB CRAB module for SGE Simplify creation and submission of CMS analysis jobs Consistent way to submit jobs to Grid or Tier-3 Local Cluster Sun Grid Engine More flexible More powerful controls Priority Job Dependency… Command line and GUI 14

Operational Experience  User acceptance of the T3 services seems to be quite good  Our CRAB SGE-scheduler module works well with SGE batch system.  SGE provides flexible and versatile way to submit and manage jobs on Tier-3 local cluster  Typical Problems in “Bad” jobs:  CMSSW jobs produce huge output file with tons of debug messages -> Fill up home directory quickly, cluster stalled  Set Quota for every user  Jobs initiate too many requests to SE in parallel -> Overload SE, jobs waiting  Users should beware 15

Upgrade Plan PhaseyearCPU/kCINT200 0 Disk/TB A (Plan) A (Archived) B (Plan)End of Hardware Upgrade Software Upgrade: Regular upgrade Glite CMS Software: CMSSW, CRAB … Upgrade under discussion: using a parallel file system instead of NFS Better performance than NFS Good for the operational of large root files

Documents and User Support  Request Account: Send to cms-  Users mailing list:  e Swiss CMS Tier-3 Wiki page 17

19

CMS Event Data Flow 20 FormatContentEvent Size [MB] RAWDetector data after online formatting; Result of HLT selections (~ 5 PB/Year) 1.5 RECOCMSSW Data Format containing the relevant output of reconstruction. (tracks, vertices, jets, electrons, muons, hits/clusters) 0.25 AODderived from the RECO information. They are in a convenient, compact format, enough information about the event to support all the typical usage patterns of a physics analysis 0.05 Event Data Flow Tier-0 Online system tape RAW,RECO AOD First pass reconstruction O(50) primary datasets O(10) streams (RAW) Tier-1 -1 Scheduled data processing (skim & reprocessing) tape RAW RECO AOD RECO, AOD Tier Analysis MC simulation Tier Analysis MC simulation Tier-3 Analysis Mc sinmulation Based on the hierarchy of computing tiers from LHC Computing Grid