PetaByte Storage Facility at RHIC Razvan Popescu - Brookhaven National Laboratory.

Slides:



Advertisements
Similar presentations
The RHIC-ATLAS Computing Facility at BNL HEPIX – Edinburgh May 24-28, 2004 Tony Chan RHIC Computing Facility Brookhaven National Laboratory.
Advertisements

Computing Infrastructure
CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000.
Vorlesung Speichernetzwerke Teil 2 Dipl. – Ing. (BA) Ingo Fuchs 2003.
Novell Server Linux vs. windows server 2008 By: Gabe Miller.
IBM RS6000/SP Overview Advanced IBM Unix computers series Multiple different configurations Available from entry level to high-end machines. POWER (1,2,3,4)
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
1 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Mass Storage For BaBar at SLAC Andrew Hanushevsky Stanford.
Storage Networking. Storage Trends Storage growth Need for storage flexibility Simplify and automate management Continuous availability is required.
Session 3 Windows Platform Dina Alkhoudari. Learning Objectives Understanding Server Storage Technologies Direct Attached Storage DAS Network-Attached.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
The SLAC Cluster Chuck Boeheim Assistant Director, SLAC Computing Services.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
An Overview of PHENIX Computing Ju Hwan Kang (Yonsei Univ.) and Jysoo Lee (KISTI) International HEP DataGrid Workshop November 8 ~ 9, 2002 Kyungpook National.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
1 U.S. Department of the Interior U.S. Geological Survey Contractor for the USGS at the EROS Data Center EDC CR1 Storage Architecture August 2003 Ken Gacke.
20-22 September 1999 HPSS User Forum, Santa Fe CERN IT/PDP 1 History  Test system HPSS 3.2 installation in Oct 1997 IBM AIX machines with IBM 3590 drives.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Developing & Managing A Large Linux Farm – The Brookhaven Experience CHEP2004 – Interlaken September 27, 2004 Tomasz Wlodek - BNL.
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
KLOE Computing Update Paolo Santangelo INFN LNF KLOE General Meeting University of Rome 2, Tor Vergata 2002, December
Large Scale Parallel File System and Cluster Management ICT, CAS.
HPSS for Archival Storage Tom Sherwin Storage Group Leader, SDSC
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
CASTOR: CERN’s data management system CHEP03 25/3/2003 Ben Couturier, Jean-Damien Durand, Olof Bärring CERN.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Probe Plans and Status SciDAC Kickoff July, 2001 Dan Million Randy Burris ORNL, Center for.
CHEP 2000: 7-11 February, 2000 I. SfiligoiData Handling in KLOE 1 CHEP 2000 Data Handling in KLOE I.Sfiligoi INFN LNF, Frascati, Italy.
8 October 1999 BaBar Storage at CCIN2P3 p. 1 Rolf Rumler BaBar Storage at Lyon HEPIX and Mass Storage SLAC, California, U.S.A. 8 October 1999 Rolf Rumler,
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
PHENIX Computing Center in Japan (CC-J) Takashi Ichihara (RIKEN and RIKEN BNL Research Center ) Presented on 08/02/2000 at CHEP2000 conference, Padova,
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
Status SC3 SARA/Nikhef 20 juli Status & results SC3 throughput phase SARA/Nikhef Mark van de Sanden.
 The End to the Means › (According to IBM ) › 03.ibm.com/innovation/us/thesmartercity/in dex_flash.html?cmp=blank&cm=v&csr=chap ter_edu&cr=youtube&ct=usbrv111&cn=agus.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
January 30, 2016 RHIC/USATLAS Computing Facility Overview Dantong Yu Brookhaven National Lab.
May 10, 2000PHENIX CC-J Updates1 PHENIX CC-J Updates - Preparation For Opening - N.Hayashi / RIKEN May 10, 2000 PHENIX Computing
Batch Software at JLAB Ian Bird Jefferson Lab CHEP February, 2000.
Storage Networking. Storage Trends Storage grows %/year, gets more complicated It’s necessary to pool storage for flexibility Intelligent storage.
01. December 2004Bernd Panzer-Steindel, CERN/IT1 Tape Storage Issues Bernd Panzer-Steindel LCG Fabric Area Manager CERN/IT.
Tackling I/O Issues 1 David Race 16 March 2010.
Production Mode Data-Replication Framework in STAR using the HRM Grid CHEP ’04 Congress Centre Interlaken, Switzerland 27 th September – 1 st October Eric.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Oct. 6, 1999PHENIX Comp. Mtg.1 CC-J: Progress, Prospects and PBS Shin’ya Sawada (KEK) For CCJ-WG.
W.A.Wojcik/CCIN2P3, HEPiX at SLAC, Oct CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
NASA Langley Research Center’s Distributed Mass Storage System (DMSS) Juliet Z. Pao Guest Lecturing at ODU April 8, 1999.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
PC Farms & Central Data Recording
Storage Networking.
Bernd Panzer-Steindel, CERN/IT
Direct Attached Storage Overview
Large Scale Test of a storage solution based on an Industry Standard
Storage Networking.
San Diego Supercomputer Center
Nuclear Physics Data Management Needs Bruce G. Gibbard
CASTOR: CERN’s data management system
RHIC Computing Facility Processing Systems
Lee Lueking D0RACE January 17, 2002
Presentation transcript:

PetaByte Storage Facility at RHIC Razvan Popescu - Brookhaven National Laboratory

CHEP PadovaPetaByte Storage Facility at RHIC2 Who are we? t Relativistic Heavy-Ion BNL –Four experiments: Phenix, Star, Phobos, Brahms. –1.5PB per year. –~500MB/sec. –>20,000SpecInt95. t Startup in May 2000 at 50% capacity and ramp up to nominal parameters in 1 year.

CHEP PadovaPetaByte Storage Facility at RHIC3 Overview t Data Types: –Raw: very large volume (1.2PB/yr.), average bandwidth (50MB/s). –DST: average volume (500TB), large bandwidth (200MB/s). –mDST: low volume (<100TB), large bandwidth (400MB/s).

CHEP PadovaPetaByte Storage Facility at RHIC4 Data Flow (generic) RHIC File Servers (DST/mDST) Reconstruction Farm (Linux) Analysis Farm (Linux) Archive raw DST mDST DST 35MB/s 50MB/s 10MB/s 200MB/s 400MB/s

CHEP PadovaPetaByte Storage Facility at RHIC5 The Data Store t HPSS (ver patch level 2) –Deployed in –After overcoming some growth difficulties we consider the present implementation successful. –One major/total reconfiguration to adapt to new hardware (and system understanding). –Flexible enough for our needs. One shortage: preemptable priority schema. –Very high performance.

CHEP PadovaPetaByte Storage Facility at RHIC6 The HPSS Archive t Constraints - large capacity & high bandwidth: –Two types of tape technology: SD-3 (best $/GB) & 9840 (best $/MB/s). –Two tape layers hierarchies. Easy management of the migration. t Reliable and fast disk storage: –FC attached RAID disk. t Platform compatible with HPSS: –IBM, SUN, SGI.

CHEP PadovaPetaByte Storage Facility at RHIC7 Present Resources t Tape Storage: –(1) STK Powderhorn silo (6000 cart.) –(11) SD-3 (Redwood) drives. –(10) 9840 (Eagle) drives. t Disk Storage: –~8TB of RAID disk. 1TB for HPSS cache. 7TB Unix workspace. t Servers: –(5) RS/6000 H50/70 for HPSS. –(6) E450&E4000 for file serving and data mining.

CHEP PadovaPetaByte Storage Facility at RHIC8

CHEP PadovaPetaByte Storage Facility at RHIC9

CHEP PadovaPetaByte Storage Facility at RHIC10

CHEP PadovaPetaByte Storage Facility at RHIC11 HPSS Structure t (1) Core Server: –RS/6000 Model H50 –4x CPU –2GB RAM –Fast Ethernet (control) –OS mirrored storage for metadata (6pv.)

CHEP PadovaPetaByte Storage Facility at RHIC12 HPSS Structure t (3) Movers: –RS/6000 Model H70 –4x CPU –1GB RAM –Fast Ethernet (control) –Gigabit Ethernet (data) (1500&9000MTU) –2x FC attached RAID - 300GB - disk cache –(3-4) SD-3 “Redwood” tape transports –(3-4) 9840 “Eagle” tape transports

CHEP PadovaPetaByte Storage Facility at RHIC13 HPSS Structure t Guarantee availability of resources for a specific user group separate resources separate PVRs & movers. t One mover per user group total exposure to single-machine failure. t Guarantee availability of resources for Data Acquisition stream separate hierarchies. t Result: 2PVR&2COS&1Mvr per group.

CHEP PadovaPetaByte Storage Facility at RHIC14 HPSS Structure

CHEP PadovaPetaByte Storage Facility at RHIC15 HPSS Topology M3M2M1Core Net 2 - Control (100baseT) Net 1 - Data (1000baseSX) STK 10baseT N x PVR pftpd Client (Routing)

CHEP PadovaPetaByte Storage Facility at RHIC16 HPSS Performance t 80 MB/sec for the disk subsystem. t ~1 CPU per 40MB/sec for TCPIP Gbit 1500MTU or 9000MTU t >9MB/sec per SD-3 transport. t ~10MB/sec per 9840 transport.

CHEP PadovaPetaByte Storage Facility at RHIC17 I/O Intensive Systems t Mining and Analysis systems. t High I/O & moderate CPU usage. t To avoid large network traffic merge file servers with HPSS movers: –Major problem with HPSS support on non-AIX platforms. –Several (Sun) SMP machines or Large (SGI) Modular System.

CHEP PadovaPetaByte Storage Facility at RHIC18 Problems t Short lifecycle of the SD-3 heads. –~ 500 hours < 2 average usage. (6 of 10 drives in 10 months). –Built a monitoring tool to try to predict transport failure (based of soft error frequency). t Low throughput interface (F/W) for SD-3: high slot consumption. t SD-3 production discontinued?! t 9840 ???

CHEP PadovaPetaByte Storage Facility at RHIC19 Issues t Tested the two tape layer hierarchies: –Cartridge based migration. –Manually scheduled reclaim. t Work with large files. Preferable ~1GB. Tolerable >200MB. –Is this true with 9840 tape transports? t Don’t think at NFS. Wait for DFS/GPFS? –We use exclusively pftp.

CHEP PadovaPetaByte Storage Facility at RHIC20 Issues t Guarantee avail. of resources for specific user groups: –Separate PVRs & movers. –Total exposure to single-mach. failure ! t Reliability: –Distribute resources across movers share movers (acceptable?). –Inter-mover traffic: 1 CPU per 40MB/sec TCPIP per adapter: Expensive!!!

CHEP PadovaPetaByte Storage Facility at RHIC21 Inter-Mover Traffic - Solutions t Affinity. –Limited applicability. t Diskless hierarchies (not for DFS/GPFS). –Not for SD-3. Not enough tests on t High performance networking: SP switch. (This is your friend.) –IBM only. t Lighter protocol: HIPPI. –Expensive hardware. t Multiply attached storage (SAN). Most promising! See STK’s talk. Requires HPSS modifications.

CHEP PadovaPetaByte Storage Facility at RHIC22 Summary t HPSS works for us. t Buy an SP2 and the SP switch. –Simplified admin. Fast interconnect. Ready for GPFS. t Keep an eye on the STK’s SAN/RAIT. t Avoid SD-3. (not a risk anymore) t Avoid small file access. At least for the moment.

Thank you! Razvan Popescu