B A B AR and the GRID Roger Barlow for Fergus Wilson GridPP 13 5 th July 2005, Durham.

Slides:



Advertisements
Similar presentations
S.L.LloydATSE e-Science Visit April 2004Slide 1 GridPP – A UK Computing Grid for Particle Physics GridPP 19 UK Universities, CCLRC (RAL & Daresbury) and.
Advertisements

1 ALICE Grid Status David Evans The University of Birmingham GridPP 14 th Collaboration Meeting Birmingham 6-7 Sept 2005.
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
GridPP9 – 5 February 2004 – Data Management DataGrid is a project funded by the European Union GridPP is funded by PPARC GridPP2: Data and Storage Management.
BaBarGrid GridPP10 Meeting CERN June 3 rd 2004 Roger Barlow Manchester University 1: Simulation 2: Data Distribution: The SRB 3: Distributed Analysis.
1 ALICE Grid Status David Evans The University of Birmingham GridPP 16 th Collaboration Meeting QMUL June 2006.
Your university or experiment logo here BaBar Status Report Chris Brew GridPP16 QMUL 28/06/2006.
Status Report University of Bristol 3 rd GridPP Collaboration Meeting 14/15 February, 2002Marc Kelly University of Bristol 1 Marc Kelly University of Bristol.
Partner Logo Tier1/A and Tier2 in GridPP2 John Gordon GridPP6 31 January 2003.
GridPP Building a UK Computing Grid for Particle Physics A PPARC funded project.
1Oxford eSc – 1 st July03 GridPP2: Application Requirement & Developments Nick Brook University of Bristol ALICE Hardware Projections Applications Programme.
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
12th September 2002Tim Adye1 RAL Tier A Tim Adye Rutherford Appleton Laboratory BaBar Collaboration Meeting Imperial College, London 12 th September 2002.
Service Data Challenge Meeting, Karlsruhe, Dec 2, 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Plans and outlook at GridKa Forschungszentrum.
The GATE-LAB system Sorina Camarasu-Pop, Pierre Gueth, Tristan Glatard, Rafael Silva, David Sarrut VIP Workshop December 2012.
GridPP From Prototype to Production David Britton 21/Sep/06 1.Context – Introduction to GridPP 2.Performance of the GridPP/EGEE/wLCG Grid 3.Some Successes.
LHCb Computing Activities in UK Current activities UK GRID activities RICH s/w activities.
GridPP News NeSC opening “Media” dissemination Tier 1/A hardware Web pages Collaboration meetings Nick Brook University of Bristol.
EasyGrid: the job submission system that works! James Cunha Werner GridPP18 Meeting – University of Glasgow.
McFarm: first attempt to build a practical, large scale distributed HEP computing cluster using Globus technology Anand Balasubramanian Karthik Gopalratnam.
Grid in action: from EasyGrid to LCG testbed and gridification techniques. James Cunha Werner University of Manchester Christmas Meeting
Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
Production Planning Eric van Herwijnen Thursday, 20 june 2002.
The B A B AR G RID demonstrator Tim Adye, Roger Barlow, Alessandra Forti, Andrew McNab, David Smith What is BaBar? The BaBar detector is a High Energy.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
ATLAS-Specific Activity in GridPP EDG Integration LCG Integration Metadata.
JetWeb on the Grid Ben Waugh (UCL), GridPP6, What is JetWeb? How can JetWeb use the Grid? Progress report The Future Conclusions.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
GLAST LAT ProjectDOE/NASA Baseline-Preliminary Design Review, January 8, 2002 K.Young 1 LAT Data Processing Facility Automatically process Level 0 data.
Computing and LHCb Raja Nandakumar. The LHCb experiment  Universe is made of matter  Still not clear why  Andrei Sakharov’s theory of cp-violation.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Cosener’s House – 30 th Jan’031 LHCb Progress & Plans Nick Brook University of Bristol News & User Plans Technical Progress Review of deliverables.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
Nick Brook Current status Future Collaboration Plans Future UK plans.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
25th October 2006Tim Adye1 RAL Tier A Tim Adye Rutherford Appleton Laboratory BaBar UK Physics Meeting Queen Mary, University of London 25 th October 2006.
GridPP Building a UK Computing Grid for Particle Physics Professor Steve Lloyd, Queen Mary, University of London Chair of the GridPP Collaboration Board.
The Experiments – progress and status Roger Barlow GridPP7 Oxford 2 nd July 2003.
…building the next IT revolution From Web to Grid…
GridPP11 Liverpool Sept04 SAMGrid GridPP11 Liverpool Sept 2004 Gavin Davies Imperial College London.
BaBar and the Grid Roger Barlow Dave Bailey, Chris Brew, Giuliano Castelli, James Werner, Fergus Wilson and Will Roethel GridPP18 Glasgow March 20 th 2007.
A B A B AR InterGrid Testbed Proposal for discussion Robin Middleton/Roger Barlow Rome: October 2001.
11th November 2002Tim Adye1 Distributed Analysis in the BaBar Experiment Tim Adye Particle Physics Department Rutherford Appleton Laboratory University.
May Donatella Lucchesi 1 CDF Status of Computing Donatella Lucchesi INFN and University of Padova.
PERFORMANCE AND ANALYSIS WORKFLOW ISSUES US ATLAS Distributed Facility Workshop November 2012, Santa Cruz.
BaBarGrid UK Distributed Analysis Roger Barlow Montréal collaboration meeting June 22 nd 2006.
BaBar and the GRID Tim Adye CLRC PP GRID Team Meeting 3rd May 2000.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
11th September 2002Tim Adye1 BaBar Experience Tim Adye Rutherford Appleton Laboratory PPNCG Meeting Brighton 11 th September 2002.
Grid development at University of Manchester Hardware architecture: - 1 Computer Element and 10 Work nodes Software architecture: - EasyGrid to submit.
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
18/12/03PPD Christmas Lectures 2003 Grid in the Department A Guide for the Uninvolved PPD Computing Group Christmas Lecture 2003 Chris Brew.
BaBar-Grid Status and Prospects
GridPP10 Meeting CERN June 3 rd 2004
Eleonora Luppi INFN and University of Ferrara - Italy
Universita’ di Torino and INFN – Torino
DØ MC and Data Processing on the Grid
MonteCarlo production for the BaBar experiment on the Italian grid
The LHCb Computing Data Challenge DC06
Presentation transcript:

B A B AR and the GRID Roger Barlow for Fergus Wilson GridPP 13 5 th July 2005, Durham

Fergus Wilson 2 Outline Personnel. Current BaBar Computing Model Monte Carlo Data Reconstruction User Analyses Projections of required resources. BaBar GRID effort and planning. Monte Carlo User Analysis

5th July 2005, Durham Fergus Wilson 3 BaBar GRID Personnel (2.5 FTEs) James Werner Manchester GridPP funded Giuliano Castelli RAL GridPP funded Chris Brew RAL 50% GRID Roger Barlow Manchester BaBar GRID PI We do not have an infinite number of monkeys… our goals are therefore constrained Fergus Wulson RAL

5th July 2005, Durham Fergus Wilson 4 BaBar Computing Model – Monte Carlo Monte Carlo is generated at ~25 sites around the world. Database driven production. ~20KBytes per event. ~10 seconds per event. 2.8 billion events generated last year. 99.5% efficient. Need million events per week. MC datasets (ROOT files) are merged and sent to SLAC. MC datasets are distributed from SLAC to any Tier 1/2/3 that wants them.

5th July 2005, Durham Fergus Wilson 5 BaBar Computing Model - Data 10 Mbytes/sec to tape at SLAC. Reconstructed at Padova (1.5 fb -1 /day). Skimmed into datasets at Karlsruhe. Skimmed datasets (ROOT files) sent to SLAC. Datasets are distributed from SLAC to any Tier 1/2/3 that wants them. An analysis can be run on a laptop.

5th July 2005, Durham Fergus Wilson 6 BaBar Computing Model – User Analysis Location of datasets provided by mySQL/Oracle database. Data/Monte Carlo datasets accessed via Xrootd file server (load-balancing, fault-tolerant, disk or tape interface). Conditions accessed from proprietary Objectivity database. User Code XrootdObjectivity Files Tier 1/2/3 mySQL

5th July 2005, Durham Fergus Wilson 7 Current Status at RAL Tier 1 RAL imports data and Monte Carlo every night. RAL has the full data and Monte Carlo for 4 out 15 of the Analysis Working Group. All disk and tape are full. Importing has stopped. We will have to delete our backups of the data. Moving to a disk/tape staging system but unlikely to keep up with demand. CPU underused at the moment.

5th July 2005, Durham Fergus Wilson 8 BaBar Projections Bottom-up planning driven by luminosity: Double dataset by 2006 (500 fb -1 ) Quadruple dataset by 2008 (1000 fb -1 )

5th July 2005, Durham Fergus Wilson 9 BaBar Monte Carlo on the GRID We have already produced 30 million Monte Carlo events on the GRID at Bristol/RAL/Manchester/RHUL (2004 using globus). Now using LCG at RAL: Software is installed via an RPM at sites (provided by BaBar Italian GRID groups). Job submission/control from RAL. 1.2 million events per week during June This is 7.5% of BaBar weekly production (during a slow period). Will aim to soak up 25% of our Tier 1 allocation with SP as requested by GridPP. Should do 3-6 million per week at RAL.

5th July 2005, Durham Fergus Wilson 10 BaBar Monte Carlo on the GRID – Tier 2 We are merging the QMUL, Birmingham and Bristol BaBar farms: 240 slow (866MHz) cpus. We will setup regional Objectivity servers that can be accessed over WAN. This means Objectivity is not needed at every Tier site. We need a large stable Tier 2 if we are to roll this out beyond RAL. We dont have the manpower to develop the MC and manage lots of small sites.

5th July 2005, Durham Fergus Wilson 11 BaBar GRID Data Analysis We now have a standard generic initialisation script for all GRID sites. Sets up BaBar environment. Sets up xrootd/objectivity. Identifies what software releases are available. Identifies what conditions are available. Identifies what collections of datasets are available. Identifies if site is setup and/or validated for Monte Carlo production.

5th July 2005, Durham Fergus Wilson 12 BaBar GRID Data Analysis Prototype Job Submission System (EasyGrid): interfaces to mySQL database to identify required datasets and allocates them to jobs. Submits jobs Resubmits jobs when they fail. Resubmits jobs when they fail again. Monitors progress. Retrieves output (usually root files). Have analysed 60 million events this way with jobs submitted from Manchester to RAL.

5th July 2005, Durham Fergus Wilson 13 BaBar GRID Data Analysis The Data Analysis works if you know that the data exists at a particular site. Datasets are not static: MC always being generated. Billions of events. Millions of files. Thousands (currently 36000) collections of datasets (arranged by processing release and physics process). The challenge will be to Interrogate sites about their available data. Allocate jobs according to available data and site resources. Monitor it all. First Step: Shortly the local mySQL database that identifies the locally available datasets will also know about the availability of datasets at every other site. Can then form the backend of an RLS.

5th July 2005, Durham Fergus Wilson 14 Conclusion We are already doing Monte Carlo production on the GRID. We have met all our deliverables. We will start major production at RAL. We need some large Tier 2 sites if this is to go anywhere in the UK. We are already doing Data Analysis on the GRID. We have met all our deliverables. Concentrate on sites with BaBar infrastructure and local datasets. Provide WAN-accessible servers. We have a prototype data analysis GRID interface. Still many GRID issues to be tackled before allowing normal people near it. BUT…the GRID still has prove it can provide a production quality service on the time scale of running experiments.