1 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Mass Storage For BaBar at SLAC Andrew Hanushevsky Stanford.

Slides:



Advertisements
Similar presentations
Computing Infrastructure
Advertisements

Andrew Hanushevsky7-Feb Andrew Hanushevsky Stanford Linear Accelerator Center Produced under contract DE-AC03-76SF00515 between Stanford University.
Chapter 5: Server Hardware and Availability. Hardware Reliability and LAN The more reliable a component, the more expensive it is. Server hardware is.
CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000.
Applying Data Grids to Support Distributed Data Management Storage Resource Broker Reagan W. Moore Ian Fisk Bing Zhu University of California, San Diego.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Jean-Yves Nief, CC-IN2P3 Wilko Kroeger, SCCS/SLAC Adil Hasan, CCLRC/RAL HEPiX, SLAC October 11th – 13th, 2005 BaBar data distribution using the Storage.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
PetaByte Storage Facility at RHIC Razvan Popescu - Brookhaven National Laboratory.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
Richard P. Mount CHEP 2000Data Analysis for SLAC Physics Richard P. Mount CHEP 2000 Padova February 10, 2000.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
Interposed Request Routing for Scalable Network Storage Darrell Anderson, Jeff Chase, and Amin Vahdat Department of Computer Science Duke University.
25 February 2000Tim Adye1 Using an Object Oriented Database to Store BaBar's Terabytes Tim Adye Particle Physics Department Rutherford Appleton Laboratory.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
An Overview of PHENIX Computing Ju Hwan Kang (Yonsei Univ.) and Jysoo Lee (KISTI) International HEP DataGrid Workshop November 8 ~ 9, 2002 Kyungpook National.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
23 Oct 2002HEPiX FNALJohn Gordon CLRC-RAL Site Report John Gordon CLRC eScience Centre.
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
Linux Servers with JASMine K. Edwards, A. Kowalski, S. Philpott HEPiX May 21, 2003.
20-22 September 1999 HPSS User Forum, Santa Fe CERN IT/PDP 1 History  Test system HPSS 3.2 installation in Oct 1997 IBM AIX machines with IBM 3590 drives.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Experience with the Thumper Wei Yang Stanford Linear Accelerator Center May 27-28, 2008 US ATLAS Tier 2/3 workshop University of Michigan, Ann Arbor.
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
JLab Scientific Computing: Theory HPC & Experimental Physics Thomas Jefferson National Accelerator Facility Newport News, VA Sandy Philpott.
Integrating JASMine and Auger Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
HPSS for Archival Storage Tom Sherwin Storage Group Leader, SDSC
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility (formerly CEBAF - The Continuous Electron Beam Accelerator Facility)
8 October 1999 BaBar Storage at CCIN2P3 p. 1 Rolf Rumler BaBar Storage at Lyon HEPIX and Mass Storage SLAC, California, U.S.A. 8 October 1999 Rolf Rumler,
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
PHENIX Computing Center in Japan (CC-J) Takashi Ichihara (RIKEN and RIKEN BNL Research Center ) Presented on 08/02/2000 at CHEP2000 conference, Padova,
Performance and Scalability of xrootd Andrew Hanushevsky (SLAC), Wilko Kroeger (SLAC), Bill Weeks (SLAC), Fabrizio Furano (INFN/Padova), Gerardo Ganis.
CASPUR Site Report Andrei Maslennikov Lead - Systems Amsterdam, May 2003.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Xrootd Present & Future The Drama Continues Andrew Hanushevsky Stanford Linear Accelerator Center Stanford University HEPiX 13-October-05
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
Andrew Hanushevsky17-Mar-991 Pursuit of a Scalable High Performance Multi-Petabyte Database 16th IEEE Symposium on Mass Storage Systems Andrew Hanushevsky.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
January 30, 2016 RHIC/USATLAS Computing Facility Overview Dantong Yu Brookhaven National Lab.
Xrootd Proxy Service Andrew Hanushevsky Heinz Stockinger Stanford Linear Accelerator Center SAG September-04
W.A.Wojcik/CCIN2P3, Nov 1, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
Focus 1 July 1999Summary of April 99 HepiX mass storage meeting 1 Summary of April 1999 HepiX mass storage meeting Focus 1 July 1999 H.Renshall PDP/IT.
A UK Computing Facility John Gordon RAL October ‘99HEPiX Fall ‘99 Data Size Event Rate 10 9 events/year Storage Requirements (real & simulated data)
W.A.Wojcik/CCIN2P3, HEPiX at SLAC, Oct CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Cofax Scalability Document Version Scaling Cofax in General The scalability of Cofax is directly related to the system software, hardware and network.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
12/19/01MODIS Science Team Meeting1 MODAPS Status and Plans Edward Masuoka, Code 922 MODIS Science Data Support Team NASA’s Goddard Space Flight Center.
CCIN2P3 Site Report - BNL, Oct 18, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center.
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
PC Farms & Central Data Recording
The LHC Computing Challenge
VNX Storage Report Project: Sample VNX Report Project ID:
Unity Storage Array Profile
San Diego Supercomputer Center
Using an Object Oriented Database to Store BaBar's Terabytes
Lee Lueking D0RACE January 17, 2002
Presentation transcript:

1 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Mass Storage For BaBar at SLAC Andrew Hanushevsky Stanford Linear Accelerator Center Produced under contract DE-AC03-76SF00515 between Stanford University and the Department of Energy

2 Andrew Hanushevsky - HEPiX, October 6-8, 1999 BaBar & The B-Factory n Use big-bang energies to create B meson particles u Look at collision decay products u Answer the question “where did all the anti-matter go?” n 800 physicists collaborating from >80 sites in 10 countries u USA, Canada, China, France, Germany, Italy, Norway, Russia, UK, Taiwan n Data reconstruction & analysis requires lots of cpu power u Need over 250 Ultra 5’s to just to find particle tracks in the data n The experiment also produces large quantities of data u TBytes/year for 10 years u Data stored as objects using Objectivity u Backed up offline on tape in HPSS u Distributed to regional labs across the world

3 Andrew Hanushevsky - HEPiX, October 6-8, 1999 HPSS Milestones n Production HPSS 4.1 deployed in May, 1999 u B-factory data taking begins u Solaris Mover is working n To date, ~12TBs data stored u Over 10,000 files written n STK 9840 tapes used exclusively u Over 300 tapes written

4 Andrew Hanushevsky - HEPiX, October 6-8, 1999 HPSS Core Server n RS6000/F50 running AIX u 4 cpus u 1Gb RAM u 12 x 9Gb disk for Encina/SFS, etc n Use tape only storage heirarchy u Only use pftp to access data n One problem with BFS u symptom: pftp_client file open failures u two circumventions added to BFS

5 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Solaris Tape Movers n SLAC port of mover using HPSS version 4.1 n Solaris machine configuration u Ultra-250 with 2 cpus, 512Mb RAM, Gigabit ethernet u Solaris 2.6, DCE 2.0, Encina TX4.2 u Three 9840 tape drives, each on separate Ultra SCSI bus n Observed peak load u CPU 60% busy u Aggregate I/O 26Mb/sec

6 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Solaris Disk Movers n Does not use HPSS disk cache u Performance & Reliability F HPSS latency too high for small block transfers F Disk cache maintenance rather complex n Solaris machine configuration u E4500 & Ultra 450, 4 cpus, 1Gb RAM, Gigabit ethernet F A3500’s, RAID-5, 5-way striped, 2 controllers, 500 to 1TB u Ultra 250, 2 cpus, 512Mb RAM, Gigabit ethernet F A1000’s, RAID-5, 5-way striped, 2 controllers, 100 to 200TB u Solaris 2.6, DCE 2.0, Encina TX4.2 (DCE/Encina not necessary) n Observed peak load u CPU 65% busy u Aggregate I/O 10Mb/sec (no migration or staging at the time)

7 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Mass Storage Architecture File & catalog management Staging Manager AMS (unix fs I/o) Gateway daemon HPSS Server Disk Pool Gateway Requests Migration daemon PFTP (data) Disk Server (Solaris) HPSS Mover Tape Robot PFTP (control) Purge daemon Prestage daemon

8 Andrew Hanushevsky - HEPiX, October 6-8, 1999 HPSS Configuration

9 Andrew Hanushevsky - HEPiX, October 6-8, 1999 SLAC Detailed Hardware Configuration

10 Andrew Hanushevsky - HEPiX, October 6-8, 1999 HPSS Total Space Used

11 Andrew Hanushevsky - HEPiX, October 6-8, 1999 HPSS Total Files

12 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Summary n HPSS is very stable u Mass Storage architecture has proven to be highly flexible n Solaris mover is a success n 9840 working well for new technology n Software upgrades will be a problem n Disk Space is always an issue u Will be getting 1TB/Month for the next year (total of about 25TB) n Tape drive contention concerns u Will be getting 12 more this year (for a total of 24)