9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.

Slides:



Advertisements
Similar presentations
The RHIC-ATLAS Computing Facility at BNL HEPIX – Edinburgh May 24-28, 2004 Tony Chan RHIC Computing Facility Brookhaven National Laboratory.
Advertisements

Computing Infrastructure
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
Belle computing upgrade Ichiro Adachi 22 April 2005 Super B workshop in Hawaii.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
PetaByte Storage Facility at RHIC Razvan Popescu - Brookhaven National Laboratory.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
CHEP 2004 September 2004Richard P. Mount, SLAC Huge-Memory Systems for Data-Intensive Science Richard P. Mount SLAC CHEP, September 29, 2004.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
Simulating Quarks and Gluons with Quantum Chromodynamics February 10, CS635 Parallel Computer Architecture. Mahantesh Halappanavar.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
The SLAC Cluster Chuck Boeheim Assistant Director, SLAC Computing Services.
Design & Management of the JLAB Farms Ian Bird, Jefferson Lab May 24, 2001 FNAL LCCWS.
1 Computing & Networking User Group Meeting Roy Whitney Andy Kowalski Sandy Philpott Chip Watson 17 June 2008.
Scientific Computing Experimental Physics Lattice QCD Sandy Philpott May 20, 2011 IT Internal Review 12GeV Readiness.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
1 U.S. Department of the Interior U.S. Geological Survey Contractor for the USGS at the EROS Data Center EDC CR1 Storage Architecture August 2003 Ken Gacke.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Presented by Leadership Computing Facility (LCF) Roadmap Buddy Bland Center for Computational Sciences Leadership Computing Facility Project.
SoCal Infrastructure OptIPuter Southern California Network Infrastructure Philip Papadopoulos OptIPuter Co-PI University of California, San Diego Program.
Laboratório de Instrumentação e Física Experimental de Partículas GRID Activities at LIP Jorge Gomes - (LIP Computer Centre)
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
JLab Scientific Computing: Theory HPC & Experimental Physics Thomas Jefferson National Accelerator Facility Newport News, VA Sandy Philpott.
Integrating JASMine and Auger Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Probe Plans and Status SciDAC Kickoff July, 2001 Dan Million Randy Burris ORNL, Center for.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility (formerly CEBAF - The Continuous Electron Beam Accelerator Facility)
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National Laboratory.
Sep 02 IPP Canada Remote Computing Plans Pekka K. Sinervo Department of Physics University of Toronto 4 Sep IPP Overview 2 Local Computing 3 Network.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
PHENIX Computing Center in Japan (CC-J) Takashi Ichihara (RIKEN and RIKEN BNL Research Center ) Presented on 08/02/2000 at CHEP2000 conference, Padova,
IDE disk servers at CERN Helge Meinhard / CERN-IT CERN OpenLab workshop 17 March 2003.
1D. Olson, SDM-ISIC Mtg, 26 Mar 2002 Scientific Data Management: An Incomplete Experimental HENP Perspective D. Olson, LBNL 26 March 2002 SDM-ISIC Meeting.
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
1 Cluster Development at Fermilab Don Holmgren All-Hands Meeting Jefferson Lab June 1-2, 2005.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
January 30, 2016 RHIC/USATLAS Computing Facility Overview Dantong Yu Brookhaven National Lab.
Tier 1 at Brookhaven (US / ATLAS) Bruce G. Gibbard LCG Workshop CERN March 2004.
US ATLAS Tier 1 Facility Rich Baker Deputy Director US ATLAS Computing Facilities October 26, 2000.
Batch Software at JLAB Ian Bird Jefferson Lab CHEP February, 2000.
The Worldwide LHC Computing Grid Frédéric Hemmer IT Department Head Visit of INTEL ISEF CERN Special Award Winners 2012 Thursday, 21 st June 2012.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
26. Juni 2003Bernd Panzer-Steindel, CERN/IT1 LHC Computing re-costing for for the CERN T0/T1 center.
Jefferson Lab Site Report Sandy Philpott HEPiX Fall 07 Genome Sequencing Center Washington University at St. Louis.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Hall D Computing Facilities Ian Bird 16 March 2001.
Cluster Active Archive
LHC Computing re-costing for
OffLine Physics Computing
Scientific Computing At Jefferson Lab
Nuclear Physics Data Management Needs Bruce G. Gibbard
TeraScale Supernova Initiative
Lee Lueking D0RACE January 17, 2002
QMUL Site Report by Dave Kant HEPSYSMAN Meeting /09/2019
Presentation transcript:

9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird

9/16/2000Ian Bird/JLAB2 Overview Present facilities overview Planned growth – 3 years, 5 years Projects  PPDG, Globus, Grids, Planning for Hall D  Issues  Facilities, infrastructure

9/16/2000Ian Bird/JLAB3

9/16/2000Ian Bird/JLAB4 Existing facilities Storage  Tape  STK silo – 6000 slots  8 Redwood drives – MB/s, helical, expensive, very sensitive, unreliable, many failures  drives – MB/s, linear, mid-load, work very reliably, 1/5 cost of Redwood, tape cost is the same  Disk  Large (5 TB) NFS RAID 5 (Symbios) – 18 – 9 c/MB Load can kill them – upgrades not available and expensive  Stage disks – host attached – 2 TB  Linux file servers (Cache and DST) – RAID 0 – 3c/MB Dual PIII, 12 73GB disks, GB Ethernet Excellent match performance/network I/O/Capacity 5 x 800 GB, 4 x 400 GB Near future expansion of disk farm

9/16/2000Ian Bird/JLAB5 Existing facilities – 2 Computing  Experimental program  Farm ~ 6000 SPECint95 – 250 Linux CPU Mostly rack mounted dual processor, 100 MB ethernet  Lattice QCD  Alpha – myrinet, MPI  Small brother (1024 DSP) to Columbia-BNL-Riken  Use essentially no storage or network bandwidth Networks  Local  GB Ethernet backbone everywhere, 100 MB switched to desktops  GB (trunked) between storage servers and farm (24 port switches), interactive servers  WAN  OC-3 to Esnet installed (+ 2 weeks), OC-12 capable

9/16/2000Ian Bird/JLAB6 Existing facilities – 3 Space & power  Space limited – CC sufficient for next few years(?)  Power – UPS upgrade – sufficient for anything could install in available space Software  Storage  OSM – replacement in hand Tapeserver – disk pool/cache managers, remote file copies, –Expand to wide area, parallel file copies  Batch  LSF + wrappers  PBS in LQCD clusters (development for wide area clusters)

9/16/2000Ian Bird/JLAB7 Expected expansion – current program Experimental program  Storage  Add drives (9840 higher capacity,rate as available) -> 30, replace Redwoods  Double disk storage yearly – SAN (back end) if feasible, cost effective and useful  CPU  Level now is as required – modest increase and replacement of older systems  Network  No real changes anticipated – add more GB links, trunk

9/16/2000Ian Bird/JLAB8 Expansion – 2 LQCD & FEL  256-node (300 Gigaflop) cluster proposed (1 yr)  Aim for >1 Teraflop LQCD & > 1 Teraflop FEL  >= FY02  Context of a wider DOE advanced scientific computing program

9/16/2000Ian Bird/JLAB9 Associated projects Grids  Particle Physics Data Grid  Other – Eurogrid, alliance Supporting technologies  Globus  LSF, Condor, etc

9/16/2000Ian Bird/JLAB10 Particle Physics Data Grid Goals  The delivery of an infrastructure for very widely distributed analysis of particle physics data at multi-petabyte scales by hundreds to thousands of physicists,  The acceleration of the development of network and middleware infrastructure aimed broadly at data-intensive collaborative science. Method  Design, develop, and deploy a network and middleware infrastructure capable of supporting data analysis and data flow patterns common to the many particle physics experiments represented.  Application-specific software will be adapted to operate in this wide-area environment and to exploit this infrastructure.

9/16/2000Ian Bird/JLAB11 Planning for Hall D Issues:  Data rate  Storage capacity – tape/disk  Data access and distribution  CPU requirements  Networking  Physical facilities

9/16/2000Ian Bird/JLAB12 Data rates 100 MB/s – is done today at RHIC  Could do now with parallel 10 MB/s  Expect drives of 40-60MB/s – 1,2 drives as now for CLAS  Same timescale LHC will have GB/s

9/16/2000Ian Bird/JLAB13 CPU & Network needs CPU  Not an issue  May need farms of > 500 CPU  No big difference from current & planned  Size is dropping – 2 proc in 1U, 4,8-way systems Networking  10 GB Ethernet soon  ESnet plans sufficient for needs  University and overseas links will be OK too (LHC)

9/16/2000Ian Bird/JLAB14 Storage Disk  Capacity, I/O rates will increase, costs drop  Suspect it will not be an issue Tape  Largest uncertainty – capacity and rates  Recent experience is not as expected (slower)  STK roadmap (Feb 2000)  >300 GB tapes  60 MB/s

9/16/2000Ian Bird/JLAB15 Tape storage STK silos (6000 slots)  Assume 300 GB -> 2 silo/year  Thus need 4 for comfort  Rest of lab needs 2  Tape access - # drives depends on speed and tape capacity Options:  STK silos – 4-6 with ~ 10 drives each  “FNAL” – expensive ADIC silo with lots of “commodity” drives – AIT, DLT, Mammoth, etc..  Reliability of drives, cost of silo

9/16/2000Ian Bird/JLAB16 Space A new Computer Center building is in the plan for 5 years  Recognized as essential for 12 GeV upgrade Lab could have significant computing facilities:  Hall D + CLAS-2 + others  5 silos, 800 node farms  LQCD + FEL 5 – 10 Tflop each  ~500 nodes each Essential to get building planned soon

9/16/2000Ian Bird/JLAB17 Conclusions Facilities – commodity components Need to initiate planning now – buildings and physical infrastructure Wide-area access, export – have not done well  Encourage collaborative development projects  Practical uses now – solve future needs  PPDG – we will be more effective if we use the technology – our local tapeserver technology