Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.

Slides:



Advertisements
Similar presentations
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
Advertisements

Physics with SAM-Grid Stefan Stonjek University of Oxford 6 th GridPP Meeting 30 th January 2003 Coseners House.
Introduction to CMS computing CMS for summer students 7/7/09 Oliver Gutsche, Fermilab.
CMS Applications Towards Requirements for Data Processing and Analysis on the Open Science Grid Greg Graham FNAL CD/CMS for OSG Deployment 16-Dec-2004.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
Oxford Jan 2005 RAL Computing 1 RAL Computing Implementing the computing model: SAM and the Grid Nick West.
Sep Donatella Lucchesi 1 CDF Status of Computing Donatella Lucchesi INFN and University of Padova.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
YuChul Yang kr October 21, 2005The Korean Physical Society The Current Status of CDF Grid 양유철 *, 한대희, 공대정, 김지은, 서준석, 장성현, 조기현, 오영도,
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
HEP Experiment Integration within GriPhyN/PPDG/iVDGL Rick Cavanaugh University of Florida DataTAG/WP4 Meeting 23 May, 2002.
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
LcgCAF:CDF submission portal to LCG Federica Fanzago for CDF-Italian Computing Group Gabriele Compostella, Francesco Delli Paoli, Donatella Lucchesi, Daniel.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
Grid Job and Information Management (JIM) for D0 and CDF Gabriele Garzoglio for the JIM Team.
BaBar MC production BaBar MC production software VU (Amsterdam University) A lot of computers EDG testbed (NIKHEF) Jobs Results The simple question:
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
YuChul Yang Oct KPS 2006 가을 EXCO, 대구 The Current Status of KorCAF and CDF Grid 양유철, 장성현, 미안 사비르 아메드, 칸 아딜, 모하메드 아즈말, 공대정, 김지은,
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
CHEP 2003Stefan Stonjek1 Physics with SAM-Grid Stefan Stonjek University of Oxford CHEP th March 2003 San Diego.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
1 st December 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
Interactive Job Monitor: CafMon kill CafMon tail CafMon dir CafMon log CafMon top CafMon ps LcgCAF: CDF submission portal to LCG resources Francesco Delli.
SAMGrid as a Stakeholder of FermiGrid Valeria Bartsch Computing Division Fermilab.
SAM and D0 Grid Computing Igor Terekhov, FNAL/CD.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
GridPP18 Glasgow Mar 07 DØ – SAMGrid Where’ve we come from, and where are we going? Evolution of a ‘long’ established plan Gavin Davies Imperial College.
DØ Computing Model & Monte Carlo & Data Reprocessing Gavin Davies Imperial College London DOSAR Workshop, Sao Paulo, September 2005.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
International Workshop on HEP Data Grid Nov 9, 2002, KNU Data Storage, Network, Handling, and Clustering in CDF Korea group Intae Yu*, Junghyun Kim, Ilsung.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
22 nd September 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
4 March 2004GridPP 9th Collaboration Meeting SAMGrid:JIM and CDF Development CDF Accepts the Need for the Grid –Requirements How to Meet the Need –Status.
1 DØ Grid PP Plans – SAM, Grid, Ceiling Wax and Things Iain Bertram Lancaster University Monday 5 November 2001.
DØSAR a Regional Grid within DØ Jae Yu Univ. of Texas, Arlington THEGrid Workshop July 8 – 9, 2004 Univ. of Texas at Arlington.
SAM - Sequential Data Access via Metadata Schema Metadata Functionality Workshop Glasgow University April 26-28,2004.
May 12, 2005Batch Workshop HEPiX Karlsruhe 1 Preparing for the Grid— Changes in Batch Systems at Fermilab HEPiX Batch System Workshop.
Amber Boehnlein, FNAL Accelerator Based Physics: ATLAS CDF CMS DO STAR Amber Boehnlein OSG Consortium Meeting January 24, 2006.
Outline: Tasks and Goals The analysis (physics) Resources Needed (Tier1) A. Sidoti INFN Pisa.
ATLAS is a general-purpose particle physics experiment which will study topics including the origin of mass, the processes that allowed an excess of matter.
Evolution of a High Performance Computing and Monitoring system onto the GRID for High Energy Experiments T.L. Hsieh, S. Hou, P.K. Teng Academia Sinica,
GridPP11 Liverpool Sept04 SAMGrid GridPP11 Liverpool Sept 2004 Gavin Davies Imperial College London.
19 February 2004SAMGrid Project Review SAMGrid: Future Plans CDF Accepts the Need for the Grid –Requirements D0 Relies on the Grid –Requirements How to.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Pilot Factory using Schedd Glidein Barnett Chiu BNL
May Donatella Lucchesi 1 CDF Status of Computing Donatella Lucchesi INFN and University of Padova.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
International Workshop on HEP Data Grid Aug 23, 2003, KNU Status of Data Storage, Network, Clustering in SKKU CDF group Intae Yu*, Joong Seok Chae Department.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Eileen Berman. Condor in the Fermilab Grid FacilitiesApril 30, 2008  Fermi National Accelerator Laboratory is a high energy physics laboratory outside.
Parag Mhashilkar Computing Division, Fermi National Accelerator Laboratory.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
D0 File Replication PPDG SLAC File replication workshop 9/20/00 Vicky White.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
Integration of Physics Computing on GRID S. Hou, T.L. Hsieh, P.K. Teng Academia Sinica 04 March,
A Data Handling System for Modern and Future Fermilab Experiments Robert Illingworth Fermilab Scientific Computing Division.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
DØ Grid Computing Gavin Davies, Frédéric Villeneuve-Séguier Imperial College London On behalf of the DØ Collaboration and the SAMGrid team The 2007 Europhysics.
IFAE Apr CDF Computing Experience - Gabriele Compostella1 IFAE Apr CDF Computing Experience Gabriele Compostella, University.
LcgCAF:CDF submission portal to LCG
Dynamic Deployment of VO Specific Condor Scheduler using GT4
Presentation transcript:

Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London

CDF experiment  located at Fermilab close to Chicago  proton/anti-proton collisions at the Tevatron of an energy of 1.2 TeV CDF  multipurpose detector with discovery potential for the Higgs, studies of b physics and measurement of standard model parameters  luminosity of about 1fb -1 per year

Principle of data analysis 40MB/s 2TB/day assign particle momentum, tracks etc. raw datareco data user selection user data MC monte carlo simulation of the events user analysis analysis performed by ~800 physicists in ~60 institutes

CDF – data handling requirements  The experiment has ~ 800 physicists of which ~ 50 are in the UK.  The experiment produces large amounts of data which is stored in the US  ~ 1000 Tb per year  ~ 2000 Tb data stored to date and expect this to rise to 10,000 by 2008  UK physicists:  need to be able to copy datasets ( ~ Tb) quickly to the UK  create MC data within the UK  other UK physicists and other CDF physicists worldwide

data handling numbers CDF has acquired produces nowadays 1Pb/year, expected to rise to 10Pb by 2008 Fermilab alone is serving about 18 Tb/day 590 TBRaw data 660 TBReconstructed data 280 TBMC 1530 TBtotal Bytes read TBytes

CDF batch computing 2 types of activities –organized processing raw data reconstruction data reduction for different physics groups MC production –user analysis need to be able to copy datasets (0.5-10Tb)  both use large amount of CPU  use the same tools for all

CDF Grid Philosophy CDF has adopted Grid concepts quite late during run time while it already had a mature software  look & feel of the old data handling system maintained  reliability main issue  use existing infrastructure as portal and change software underneath

CDF Analysis Farm (CAF) Submit and forget until receiving a mail Does all the job handling and negotiation with the data handling system without the user knowing CDF batch job contains a tar ball with all the needed scripts, binaries and shared libraries and send tarball to output location user need to authenticate with their kerberos ticket

CAF -evolution over time CDF used several batch systems and distribution mechanisms FBSNG Condor Condor with Globus gLite WMS  CAF was able to be distributed, run on non-dedicated resources  glite WMS helps to run on EGEE sites Grid based Used as Production systems

Condor-based GRID CAF Collector User priorities Negotiator User jobs Schedd Globus User Job Grid nodes Starter Negotiator assigns nodes to jobs Globus assigns nodes to VOs Glide-ins Pull Model

gLite WMS-based GRID CAF Push Model User jobs Schedd Globus User Job Grid nodes Resource Broker

Pros & Cons Condor based Grid CAF Pros: Globally managed user and job priorities within CDF Broken nodes kill condor daemons, not user jobs Resource selection done after a batch slot is secured Cons: Uses a single service proxy for all jobs to enter Grid sites Requires outgoing connectivity gLite WMS-based Grid CAF Pros: LCG-backed tools No need for external connectivity Grid sites can manage users Cons: No global fair share for CDF

gLite WMS-based GRID CAF at FNAL: CAF worker nodes used to have CDF software distribution NFS mounted, but not an option in the Grid world all production jobs are now self-contained trying Parrot to distribute CDF software over HTTP in analysis jobs

FNAL remote dedicated resources Condor based Grid CAFs LCGCaf FermiGrid avg. usable VMs (Virtual Machine) Number of jobs on the CAF Some numbers

Data handling system SAM SAM manages file storage –Data files are stored in tape systems at FNAL and elsewhere (most use ENSTORE at FNAL) –Files are cached around the world for fast access SAM manages file delivery –Users at FNAL and remote sites retrieve files transparently out of file storage. SAM handles caching for efficiency SAM manages file cataloging –SAM DB holds meta-data for each file transparent to the user SAM manages analysis bookkeeping –SAM remembers what files you ran over, what files you processed, what applications you ran, when you ran them and where

world wide distribution of SAM stations selected SAM stations FNAL CDF:  10k/20k Files declared/day  15k Files consumed/day  8 TByte of Files cons./day  main consumption of data  still central  remote use on the rise testdeployment 300Tb Total CDF Files To User

summary & outlook UCL-HEP cluster deployed UCL-CCC cluster still to come need a better integration of SAM and the CAF user feedback needs to be collated