21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.

Slides:



Advertisements
Similar presentations
Tony Doyle - University of Glasgow GridPP EDG - UK Contributions Architecture Testbed-1 Network Monitoring Certificates & Security Storage Element R-GMA.
Advertisements

B A B AR and the GRID Roger Barlow for Fergus Wilson GridPP 13 5 th July 2005, Durham.
12th September 2002Tim Adye1 RAL Tier A Tim Adye Rutherford Appleton Laboratory BaBar Collaboration Meeting Imperial College, London 12 th September 2002.
UCL HEP Computing Status HEPSYSMAN, RAL,
Computing Infrastructure
Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
Distributed IT Infrastructure for U.S. ATLAS Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen1 Offline Mass Data Processing using Online Computing Resources at HERA-B José Hernández DESY-Zeuthen.
Jean-Yves Nief, CC-IN2P3 Wilko Kroeger, SCCS/SLAC Adil Hasan, CCLRC/RAL HEPiX, SLAC October 11th – 13th, 2005 BaBar data distribution using the Storage.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
TRIUMF Site Report for HEPiX, SLAC, October 10-14,2005 TRIUMF SITE REPORT Corrie Kost Update since Hepix Spring 2005.
TRIUMF SITE REPORT – Corrie Kost April Catania (Italy) Update since last HEPiX/HEPNT meeting.
EU funding for DataGrid under contract IST is gratefully acknowledged GridPP Tier-1A Centre CCLRC provides the GRIDPP collaboration (funded.
Edinburgh Site Report 1 July 2004 Steve Thorn Particle Physics Experiments Group.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
Terabyte IDE RAID-5 Disk Arrays David A. Sanders, Lucien M. Cremaldi, Vance Eschenburg, Romulus Godang, Christopher N. Lawrence, Chris Riley, and Donald.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
April 2001HEPix/HEPNT1 RAL Site Report John Gordon CLRC, UK.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
The SLAC Cluster Chuck Boeheim Assistant Director, SLAC Computing Services.
30-Jun-04UCL HEP Computing Status June UCL HEP Computing Status April DESKTOPS LAPTOPS BATCH PROCESSING DEDICATED SYSTEMS GRID MAIL WEB WTS.
Group Computing Strategy Introduction and BaBar Roger Barlow June 28 th 2005.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
An Overview of PHENIX Computing Ju Hwan Kang (Yonsei Univ.) and Jysoo Lee (KISTI) International HEP DataGrid Workshop November 8 ~ 9, 2002 Kyungpook National.
Paul Scherrer Institut 5232 Villigen PSI HEPIX_AMST / / BJ95 PAUL SCHERRER INSTITUT THE PAUL SCHERRER INSTITUTE Swiss Light Source (SLS) Particle accelerator.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
23 Oct 2002HEPiX FNALJohn Gordon CLRC-RAL Site Report John Gordon CLRC eScience Centre.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
CDF Offline Production Farms Stephen Wolbers for the CDF Production Farms Group May 30, 2001.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Laboratório de Instrumentação e Física Experimental de Partículas GRID Activities at LIP Jorge Gomes - (LIP Computer Centre)
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH OS X Home server AFS using openafs 3 DB servers Kerberos 4 we will move.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
WP8 Meeting Glenn Patrick1 LHCb Grid Activities in UK Grid WP8 Meeting, 16th November 2000 Glenn Patrick (RAL)
D0 Taking Stock 11/ /2005 Calibration Database Servers.
19th September 2003Tim Adye1 RAL Tier A Status Tim Adye Rutherford Appleton Laboratory BaBar UK Collaboration Meeting Royal Holloway 19 th September 2003.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
Nov. 8, 2000RIKEN CC-J RIKEN CC-J (PHENIX Computing Center in Japan) Report N.Hayashi / RIKEN November 8, 2000 PHENIX Computing
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
November 10, 1999PHENIX CC-J Updates in Nov.991 PHENIX CC-J Updates in Nov New Hardware - N.Hayashi / RIKEN November 10, 1999 PHENIX Computing Meeting.
International Workshop on HEP Data Grid Aug 23, 2003, KNU Status of Data Storage, Network, Clustering in SKKU CDF group Intae Yu*, Joong Seok Chae Department.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
January 30, 2016 RHIC/USATLAS Computing Facility Overview Dantong Yu Brookhaven National Lab.
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
MC Production in Canada Pierre Savard University of Toronto and TRIUMF IFC Meeting October 2003.
US ATLAS Tier 1 Facility Rich Baker Deputy Director US ATLAS Computing Facilities October 26, 2000.
CNAF Database Service Barbara Martelli CNAF-INFN Elisabetta Vilucchi CNAF-INFN Simone Dalla Fina INFN-Padua.
15-Feb-02Steve Traylen, RAL WP6 Test Bed Report1 RAL/UK WP6 Test Bed Report Steve Traylen, WP6 PPGRID/RAL, UK
Batch Software at JLAB Ian Bird Jefferson Lab CHEP February, 2000.
A UK Computing Facility John Gordon RAL October ‘99HEPiX Fall ‘99 Data Size Event Rate 10 9 events/year Storage Requirements (real & simulated data)
Oct. 6, 1999PHENIX Comp. Mtg.1 CC-J: Progress, Prospects and PBS Shin’ya Sawada (KEK) For CCJ-WG.
W.A.Wojcik/CCIN2P3, HEPiX at SLAC, Oct CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
11th September 2002Tim Adye1 BaBar Experience Tim Adye Rutherford Appleton Laboratory PPNCG Meeting Brighton 11 th September 2002.
UK GridPP Tier-1/A Centre at CLRC
Universita’ di Torino and INFN – Torino
Presentation transcript:

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second Large Scale Cluster Computing Workshop

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 Overview ● Distributed Collaboration ● Computing Sites ● Production ● Analysis ● BaBar GRID

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 Distributed Collaboration

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 Distributed Collaboration (Cont.) ● 560 Collaborations in 76 institutions in 9 countries – Important to allow remote development ● AFS & CVS – Started tools to allow remote collaboration ● SoftRelTools & SiteConfig – BaBar tools used to allow software to run at any site ● At last count (2 years ago) over 300 collaborations had contributed to software

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 Computing Sites ● Multi-tier system – Tier-As: large concentration of resources – Tier-Bs: not observered – Tier-Cs: small sites/laptops ● Currently 4 Tier-As ● O(20) serious Tier-C sites – Could have a few people's laptops

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 SLAC ● Main computing site – Complete data set(s) on tape – Till 1.5 years ago all production – All users require SLAC account ● Client hardware – 900 Sun Solaris8 440MHz SparcIIIs – 512 VA RedHat7.2 Dual 866MHz P-IIIs – 256 (+128) Rackable RedHat7.2 Dual 1.4GHz P-IIIs

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 Sun T1s VA 1220s Rackables (back-to-back) E420R + T3 disks Journal & Lock

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 SLAC (Cont.) ● Server hardware – ~100 small servers (lock, journal etc) ● Sun Solaris8 440MHz SparcIIIs ● 100Mb Ethernet – ~50 data servers (2 or 4 CPUs with 1-2TB disk) ● Gigabit Ethernet ● HPSS used to manage tertiary storage ● Switched network

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 IN2P3 ● First external Tier-A centre – Started early 2001 ● Shared facility with 30+ experiments ● BaBar allocation (varies when needed) – 11 data servers (22TB disk) ● Currently all Sun machines ● Heavily rely on HPSS to stage data – ~100 IBM eServers Dual P-IIIs clients ● Mix of 750MHz and 1.2GHz – ~40 Sun Solaris8 clients

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 IBM Linux clients Sun data servers

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 INFN/Padova ● 120 IBM eServer Dual 1.26GHz P-IIIs clients – 100Mbit Ethernet ● 8 Linux servers with 1TB each for production ● 28TB of disk in total ● LTO Tape library – Currently contains ~80TB of Xtc files ● Output from online system ● Will be complete off site backup

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 IBM Clients EIDE 3Ware RAID

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 RAL ● 104 RedHat7.2 Dual 1.4GHz P-IIIs – In addition share old farm of 156 slower machines – 100Mb Ethernet ● 20TB of disk utilising IDE RAID arrays – Using Gigabit Ethernet

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 Linux Clients IDE disk array

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 Production ● Data Reconstruction – Prompt Reconstruction – Reprocessing ● Simulation Production ● Skimming ● Alternative data format production – Primary data format is BaBar Data Store/Objectivity – Format for university distribution is Kanga/ROOT

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 Data Reconstruction ● Calibrations done with data ● Prior model used Rolling Calibrations – One defined the constants for reconstrucing the next ● Now split calibration step Prompt Calibration Event Reconstruction Event Reconstruction Event Reconstruction

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 Reprocessing ● Started two weeks ago ● Reprocesses full statistics – Should finish in February ● Prompt Calibration – SLAC Farm ● Event Reconstruction – 3 (or 4) Padova Farms ● Actually two SLAC – Second feeds two SLAC ER farms Prompt Calibration Event Reconstruction Event Reconstruction Event Reconstruction SLAC 16 Dual 1.4GHz P-IIIs Padova Each farm 40 Dual 1.26GHz P-IIIs

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 Prompt Reconstruction ● PEP-II restarts 15 th November – Expect collisions within week ● Start with two ER farms at SLAC ● Add Padova farms – When reprocessing done Prompt Calibration Event Reconstruction Event Reconstruction Event Reconstruction SLAC 16 Dual 1.4GHz P-IIIs Each Farm; 32 Dual 1.4GHz P-IIIs Padova Farm(s)

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 Simulation Production ● Largly Distributed system – 20 sites produce Monte Carlo – 48 Dual 866MHz P-IIIs at SLAC – Home grown system to centrally manage, ProdTools ● Moving towards larger statistics – Generate ~1σ continuum/taus – ~2σ B mesons – Increase to 3σ for Bs ● Utilise under used cycles on analysis and data production farms – Required increase in memory on client nodes (2GB)

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 Analysis ● Three of the Tier-A site currently provide analysis capacity – SLAC and IN2P3 provides Objectivity based data – RAL provides ROOT based data ● Almost all RAL & IN2P3 for analysis ● SLAC provides 20k SpecINT 95s

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar GRID ● Three applications in development – Simulation Production ● Currently use ~8 FTEs for production – Analysis ● Allow better use of world wide resources ● Allow users to work interactively where they wish – Currently users need to login to Tier-A sites to use their resources – Data Distribution ● User can initiate request from their site and athenticate to nearest resource with data

21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 Conclusions ● BaBar development has been distributed since inception ● During last two years moved to use distributed computing resources ● Moving towards using Grid Techonolgy to add transparency