Hall D Computing Facilities Ian Bird 16 March 2001.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
Randall Sobie The ATLAS Experiment Randall Sobie Institute for Particle Physics University of Victoria Large Hadron Collider (LHC) at CERN Laboratory ATLAS.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
LIGO- GXXXXXX-XX-X Advanced LIGO Data & Computing Material for the breakout session NSF review of the Advanced LIGO proposal Albert Lazzarini Caltech,
DATA PRESERVATION IN ALICE FEDERICO CARMINATI. MOTIVATION ALICE is a 150 M CHF investment by a large scientific community The ALICE data is unique and.
1 Towards an Upgrade TDR: LHCb Computing workshop May 2015 Introduction to Upgrade Computing Session Peter Clarke Many peoples ideas Vava, Conor,
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
Design & Management of the JLAB Farms Ian Bird, Jefferson Lab May 24, 2001 FNAL LCCWS.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
LHC Computing Review - Resources ATLAS Resource Issues John Huth Harvard University.
PPDG and ATLAS Particle Physics Data Grid Ed May - ANL ATLAS Software Week LBNL May 12, 2000.
LHC Computing Plans Scale of the challenge Computing model Resource estimates Financial implications Plans in Canada.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Data Logistics in Particle Physics Ready or Not, Here it Comes… Prof. Paul Sheldon Vanderbilt University Prof. Paul Sheldon Vanderbilt University.
December 10,1999: MONARC Plenary Meeting Harvey Newman (CIT) Phase 3 Letter of Intent (1/2)  Short: N Pages è May Refer to MONARC Internal Notes to Document.
Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
Ian Bird GDB; CERN, 8 th May 2013 March 6, 2013
…building the next IT revolution From Web to Grid…
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory Review of U.S. LHC Software and Computing Projects Fermi National Laboratory November.
ESFRI & e-Infrastructure Collaborations, EGEE’09 Krzysztof Wrona September 21 st, 2009 European XFEL.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
1D. Olson, SDM-ISIC Mtg, 26 Mar 2002 Scientific Data Management: An Incomplete Experimental HENP Perspective D. Olson, LBNL 26 March 2002 SDM-ISIC Meeting.
23.March 2004Bernd Panzer-Steindel, CERN/IT1 LCG Workshop Computing Fabric.
U.S. Grid Projects and Involvement in EGEE Ian Foster Argonne National Laboratory University of Chicago EGEE-LHC Town Meeting,
LHC Computing, CERN, & Federated Identities
U.S. ATLAS Computing Facilities Overview Bruce G. Gibbard Brookhaven National Laboratory U.S. LHC Software and Computing Review Brookhaven National Laboratory.
Tier 1 at Brookhaven (US / ATLAS) Bruce G. Gibbard LCG Workshop CERN March 2004.
The ATLAS Computing Model and USATLAS Tier-2/Tier-3 Meeting Shawn McKee University of Michigan Joint Techs, FNAL July 16 th, 2007.
U.S. ATLAS Computing Facilities DOE/NFS Review of US LHC Software & Computing Projects Bruce G. Gibbard, BNL January 2000.
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
U.S. ATLAS Computing Facilities U.S. ATLAS Physics & Computing Review Bruce G. Gibbard, BNL January 2000.
Storage Management on the Grid Alasdair Earl University of Edinburgh.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
05/14/04Larry Dennis, FSU1 Scale of Hall D Computing CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
1 GlueX Software Oct. 21, 2004 D. Lawrence, JLab.
10-Feb-00 CERN HepCCC Grid Initiative ATLAS meeting – 16 February 2000 Les Robertson CERN/IT.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
LHC collisions rate: Hz New PHYSICS rate: Hz Event selection: 1 in 10,000,000,000,000 Signal/Noise: Raw Data volumes produced.
Computing requirements for the experiments Akiya Miyamoto, KEK 31 August 2015 Mini-Workshop on ILC Infrastructure and CFS for Physics and KEK.
Evolution of storage and data management
HEP LTDP Use Case & EOSC Pilot
Ian Bird WLCG Workshop San Francisco, 8th October 2016
Grid site as a tool for data processing and data analysis
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
Pasquale Migliozzi INFN Napoli
LHC experiments Requirements and Concepts ALICE
for the Offline and Computing groups
WLCG: TDR for HL-LHC Ian Bird LHCC Referees’ meting CERN, 9th May 2017.
Russian Regional Center for LHC Data Analysis
LHCb computing in Russia
UK GridPP Tier-1/A Centre at CLRC
Readiness of ATLAS Computing - A personal view
The LHC Computing Grid Visit of Her Royal Highness
Fabric and Storage Management
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Scientific Computing At Jefferson Lab
LHCb thinking on Regional Centres and Related activities (GRIDs)
Development of LHCb Computing Model F Harris
The LHC Computing Grid Visit of Professor Andreas Demetriou
Presentation transcript:

Hall D Computing Facilities Ian Bird 16 March 2001

Overview Comparisons – Hall D computing Estimates of needs –As an illustration – but actual needs require a model –Costs –Staffing –Timeline Other projects – Data Grids Some comments

Some comparisons: Hall D vs other HENP Data Volumes (tape) TB/year Data rates MB/s Disk Cache TB CPU SI95/year People CMS2 000 (total) ~1800 US Atlas (Tier 1) ~500 (?) STAR20040>207000~300 D0/CDF Run II 300~500 BaBar300~500 Not just an issue of equipment. These experiments all have the support of – large dedicated computing groups within the experiments – well defined computing models JLAB– current ~240 (CLAS) Hall D

Process For CDR; computing/analysis chapter –Define the Hall-D computing model Distributed architecture (facilities) Data model Software architecture Collaboratory tools and infrastructure Estimate of costs and funding profile, and management plan Can set bounds (best/worst case) based on technology guesses –All this must be based on: Analysis models – e.g. how will the PWA be done, etc. –Needs strong management – hire “now” a computing professional to lead this Write a Computing Technical Design Report –This can come after the CDR, fixes the ideas from the CDR, provides a detailed implementation plan

JLAB Facilities for Hall D Some crude estimates – –No computing model – has to come first –Maybe too soon to fix technologies

Mass storage – at JLAB How much needs to be on tape – depends on computing model and how well managed the activities are: Assume: –0.75 PB/year raw data –0.75 PB/year reconstructed –0.3 PB/year other –All simulated data stored off-site –1.8 PB/year (minimum) to be stored 300 GB/tape = 6000 tapes = 1 silo 750 GB/tape = 2400 tapes = ½ silo –Need other tapes (DST on fast access, lower density) –Keep data available for 2 years: Pessimistic – 2 silos, Optimistic – 1 silo (for Hall D) –Realistic guess – Hall D should have at least 2 dedicated silos –Number of drives – depends on technology and access model Experience shows need at least 30 drives Lab needs more for other parts of the program

Other storage Disk –Again amount and type depends strongly on the computing model –Not unreasonable to expect to want 20% of data on disk 200 TB ? –Current costs – 1TB/ $10K (IDE), - expect  10? Cost and type depends on requirements

CPU & Networks Not a computing problem for reconstruction –All significant computing is in simulation – most not at JLAB? Level 3 trigger farm –It will be cheaper to compute more and store (and move) less –Conservative assumption – 500 SI95/processor 2 procs in 1u rack = 40,000 SI95/rack Networking –Will be of critical importance to success of Hall D Distributed computing model Transparent access to all data for all users –Expect 10-Gigabit Ethernet (perhaps first deployment of subsequent generation) –Assume JLAB will have OC12 (622 Mb/s) to ESNet Even today just a configuration change

Staffing Experiment needs to have a strong dedicated computing group Computer Center – needs depends on facilities – depends on computing model Estimate: –Support of Hall D Level 3 farm:0.5 –Support of offline MSS, farm:3.0 –Additional network support:0.5 –Development/experiment support:2.0 »Total 6.0

Costs Real cost will be >> $3M in report – probably closer to $5-6M –Cf. RHIC computing facilities was $12M project over 5 years Costs cannot be defined without a clear vision for the computing model New Computer Center is already in lab building plan

Integration Technologies will be there Challenge is in software (middleware) in integrating all the distributed pieces into a seamless system that is useable and responsive

Development activities Grid computing, collaboratory environments and Data Grids

LHC Concept of Computing Hierarchy – Data Grid LHC Grid Hierarchy Example Tier0: CERN Tier1: National “Regional” Center Tier2: Regional Center Tier3: Institute Workgroup Server Tier4: Individual Desktop Total 5 Levels

Data Grid activities Particle Physics Data Grid (PPDG) –DOE funded – labs (inc JLAB) + universities GriPhyN (Grid Physics Network) –NSF funded Computing grids are heavily funded –US, Europe, Japan, –LHC computing relies on these technologies –Not just academic interest - industry

PPDG Has been funded for last 2 years New PPDG proposal using DOE SciDac funds – just submitted –Other (complementary) proposals relevant to JLAB or Hall D: FSU/IU proposal – Hall D portal FIU proposal LQCD PPDG will: –“…provide a distributed (grid-enabled) data access and management service for the large collaborations of current and future particle and nuclear physics experiments. It is a collaborative effort between physicists and computer scientists at several DOE laboratories and universities. This is accomplished by applying existing grid middleware to current problems and providing feedback to middleware developers on additional features required or shortcomings in the current implementations.” –For JLAB will provide directly useful services for current program These funds could be targeted next year for Hall D development activities (needs a context first)

Comments Absolute requirement: –Need a clear vision for the computing/analysis model –Computing requires a dedicated group within Hall D – the leader of that group should be found now Management –Badly managed computing costs real money Well managed – calibration & reconstruction are immediate – need less long-term storage; do not need to keep simulated data,… Badly managed software architecture will kill the L3 trigger – you have to trust it –Computing task is not trivial, but not overwhelming, but is at least as complex as the detector Must be recognized and treated as such by the collaboration