GStore: GSI Mass Storage ITEE-Palaver GSI 26.6.2007 Horst Göringer, Matthias Feyerabend, Sergei Sedykh

Slides:



Advertisements
Similar presentations
NGAS – The Next Generation Archive System Jens Knudstrup NGAS The Next Generation Archive System.
Advertisements

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.
XenData SXL-5000 LTO Archive System Turnkey video archive system with near-line LTO capacities scaling from 210 TB to 1.18 PB, designed for the demanding.
The GSI Mass Storage for Experiment Data DVEE-Palaver GSI Darmstadt Feb. 15, 2005 Horst Göringer, GSI Darmstadt
Vorlesung Speichernetzwerke Teil 2 Dipl. – Ing. (BA) Ingo Fuchs 2003.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Server Platforms Week 11- Lecture 1. Server Market $ 46,100,000,000 ($ 46.1 Billion) Gartner.
NWfs A ubiquitous, scalable content management system with grid enabled cross site data replication and active storage. R. Scott Studham.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
PetaByte Storage Facility at RHIC Razvan Popescu - Brookhaven National Laboratory.
Backup Rationalisation Reorganisation of the CERN Computer Centre Backups David Asbury IT/DS Friday 6 December 2002.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
1 Objectives Discuss the Windows Printer Model and how it is implemented in Windows Server 2008 Install the Print Services components of Windows Server.
File Systems and N/W attached storage (NAS) | VTU NOTES | QUESTION PAPERS | NEWS | VTU RESULTS | FORUM | BOOKSPAR ANDROID APP.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
DMF Configuration for JCU HPC Dr. Wayne Mallett Systems Manager James Cook University.
The GSI Mass Storage System TAB GridKa, FZ Karlsruhe Sep. 4, 2002 Horst Göringer, GSI Darmstadt
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
1 © 2010 Overland Storage, Inc. © 2012 Overland Storage, Inc. Overland Storage The Storage Conundrum Neil Cogger Pre-Sales Manager.
Virtualization in the NCAR Mass Storage System Gene Harano National Center for Atmospheric Research Scientific Computing Division High Performance Systems.
Building Advanced Storage Environment Cheng Yaodong Computing Center, IHEP December 2002.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
Nov 1, 2000Site report DESY1 DESY Site Report Wolfgang Friebel DESY Nov 1, 2000 HEPiX Fall
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
9-Sept-2003CAS2003, Annecy, France, WFS1 Distributed Data Management at DKRZ Distributed Data Management at DKRZ Wolfgang Sell Hartmut Fichtel Deutsches.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
M.Lautenschlager (WDCC, Hamburg) / / 1 Semantic Data Management for Organising Terabyte Data Archives Michael Lautenschlager World Data Center.
Storage Tank in Data Grid Shin, SangYong(syshin, #6468) IBM Grid Computing August 23, 2003.
POW : System optimisations for data management 11 November 2004.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Large Scale Parallel File System and Cluster Management ICT, CAS.
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
CASTOR: CERN’s data management system CHEP03 25/3/2003 Ben Couturier, Jean-Damien Durand, Olof Bärring CERN.
Storage and Storage Access 1 Rainer Többicke CERN/IT.
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015.
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
The KLOE computing environment Nuclear Science Symposium Portland, Oregon, USA 20 October 2003 M. Moulson – INFN/Frascati for the KLOE Collaboration.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
January 30, 2016 RHIC/USATLAS Computing Facility Overview Dantong Yu Brookhaven National Lab.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
SATA In Enterprise Storage Ron Engelbrecht Vice President and General Manager Engineering and Manufacturing Operations September 21, 2004.
W.A.Wojcik/CCIN2P3, Nov 1, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
CASTOR project status CASTOR project status CERNIT-PDP/DM October 1999.
1.3 ON ENHANCING GridFTP AND GPFS PERFORMANCES A. Cavalli, C. Ciocca, L. dell’Agnello, T. Ferrari, D. Gregori, B. Martelli, A. Prosperini, P. Ricci, E.
CERN IT Department CH-1211 Genève 23 Switzerland t The Tape Service at CERN Vladimír Bahyl IT-FIO-TSI June 2009.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
GDB meeting - Lyon - 16/03/05 An example of data management in a Tier A/1 Jean-Yves Nief.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
Status of GSDC, KISTI Sang-Un Ahn, for the GSDC Tier-1 Team
Centre de Calcul de l’Institut National de Physique Nucléaire et de Physique des Particules Data storage services at CC-IN2P3 Jean-Yves Nief.
An Introduction to GPFS
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
CERN IT-Storage Strategy Outlook Alberto Pace, Luca Mascetti, Julien Leduc
CASTOR: possible evolution into the LHC era
Tivoli Storage Manager 4.1
Bernd Panzer-Steindel, CERN/IT
Brookhaven National Laboratory Storage service Group Hironori Ito
Ákos Frohner EGEE'08 September 2008
The INFN Tier-1 Storage Implementation
Large Scale Test of a storage solution based on an Industry Standard
CASTOR: CERN’s data management system
Presentation transcript:

gStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh

ITEE-Palaver gStore - GSI Mass Storage2 Overview 1.adsmcli: ended after 10 years of operation 2.tsmcli: modern concept 3.gStore (gstore): unified user interface 4.rearrangement of storage 5.gStore projects 6.final remarks

ITEE-Palaver gStore - GSI Mass Storage3 adsmcli: initial system Software: 1.ADSM: Adstar Storage manager commercial handles ATL and tapes 2.GSI Software: –Interface to users –API to ADSM

ITEE-Palaver gStore - GSI Mass Storage4 adsmcli: initial system Hardware 1996: –AIX server –ATL: IBM tape drives IBM 3590: 14 MB/s, 10 GB/volume max 23 TByte – few GB disk (write) cache ADSM – 80 GB read cache (1998)

ITEE-Palaver gStore - GSI Mass Storage5 adsmcli: overview

ITEE-Palaver gStore - GSI Mass Storage6 adsmcli: early usage

ITEE-Palaver gStore - GSI Mass Storage7 adsmcli: the best year

ITEE-Palaver gStore - GSI Mass Storage8 adsmcli limitations Restrictions: –bottleneck server –no scalability data capacity (cache) I/O bandwidth –missing write cache frozen since 2001 –only read cache upgrade 2003: 1.2 TB

ITEE-Palaver gStore - GSI Mass Storage9 tsmcli: concepts Concepts: separation of control and data flow: –data flow: Data Mover –control flow: TSM Server, Entry Server many DMs => many parallel data streams SAN: Storage Area Network Cache Manager: read and write cache direct DAQ connection to gStore

ITEE-Palaver gStore - GSI Mass Storage10 tsmcli concept

ITEE-Palaver gStore - GSI Mass Storage11 tsmcli: storage view

ITEE-Palaver gStore - GSI Mass Storage12 tsmcli: usage tsmcli in production since January 2003 –in parallel to adsmcli –initially only for 'large' experiments write cache: since February 2005 –for 'normal' clients: command tsmcli RFIO API –for DAQ clients (RFIO, write only)

ITEE-Palaver gStore - GSI Mass Storage13 tsmcli hardware 2007 server: Windows 2000 cluster ATL: Sun StorageTek L700 –9 tape drives LTO2: –35 MByte/s, 200 GByte/vol –max 140 TByte data mover: –10 Windows (gsidm0-9), 4 TB disk cache – 5 Linux (slxdm01-5), 13 TB disk cache

ITEE-Palaver gStore - GSI Mass Storage14 tsmcli usage 2006

ITEE-Palaver gStore - GSI Mass Storage15 tsmcli usage 2007

ITEE-Palaver gStore - GSI Mass Storage16 gStore top load top data transfer in 2006: Dec 31 overall: 9.6 TB in 24 h –111 MB/s on average slxdm01: 2.9 TB in 24 h –33.6 MB/s on average

ITEE-Palaver gStore - GSI Mass Storage17 common mass storage interface coexistence of 2 mass storage systems: intermediary solution (ca 4 years) => common new interface gstore: replacing adsmcli and tsmcli successfully in operation since May 23 (considerable) enhancement of tsmcli SW: access to 2 independent TSM servers and attached DMs/disk caches –further scalability aspect!

ITEE-Palaver gStore - GSI Mass Storage18 storage status mid ATLs: 1.IBM 3494 (3590 tapes): –50 TB experiment data (adsmcli) –15 TB backup data –nearly filled 2.Sun StorageTek L700 (LTO2 tapes): –120 TB experiment data (tsmcli) –max 140 TB => nearly filled 3.Sun StorageTek L700 (LTO1 tapes): –38 TB backup data –max 70 TB

ITEE-Palaver gStore - GSI Mass Storage19 requirements 1.substantially more data capacity –4 new tape drives IBM 3592 for 3494 ATL 2.separate experiment and backup data: experiment data -> IBM 3494 backup data -> LTO2 ATL 3.safe long term storage upgrade LTO1 ATL -> LTO3 deploy in 'remote RZ'

ITEE-Palaver gStore - GSI Mass Storage20 gstore hardware server: AIX –ATL: IBM tape drives IBM 3592: 100 MByte/s, 700 GByte/vol max 1.6 PB –data mover: 5 Linux (slxdm01-5), 13 TB disk cache 3 Linux (slxdm06-8), 17 TB disk cache 2.server: Windows

ITEE-Palaver gStore - GSI Mass Storage21 actions: move all existing experiment data to IBM 3592 tapes in 3494 ATL –50 TB from 3590 media: finished (adsmcli data) => old 3590 hardware/media replaced –130 TB from LTO2 media: 40 TB done (tsmcli data) write all new experiment data to 3494 ATL: –since May 23

ITEE-Palaver gStore - GSI Mass Storage22 actions: redirect all new backup data to LTO2 media –new pair of Linux TSM servers –in work move actual backup data to LTO2 media –mainly user archives –from LTO1 and 3590 media –still to be done

ITEE-Palaver gStore - GSI Mass Storage23 open projects xrootd: –in test environments gStore access for xrootd clients available –still open: stability xrdcp, functionality Posix ls

ITEE-Palaver gStore - GSI Mass Storage24 open projects Grid SRM (Storage Resource Manager): –several types of SRMs installed worldwide –common: no general mass storage interface –currently under investigation for connection with gStore: Berkeley SRM ('BeStMan')

ITEE-Palaver gStore - GSI Mass Storage25 open projects 2nd level DM: –no SAN connection –filled via LAN from 1st level DM –inexpensive extension gStore read cache: for data needed online for longer time scales (weeks/months) –no NFS: use gstore query/retrieve e.g. xrootd: enable full file information for new /d file servers !?

ITEE-Palaver gStore - GSI Mass Storage26 user requests gStore enhancements: –staging large sets of files: equal distribution on all DMs (1st or 2nd level) stage –distr stage –distr –L2 –recursive access query/stage/retrieve –r path –rename path/file –files > 2 GB –...

ITEE-Palaver gStore - GSI Mass Storage27 Final Remarks I currently ca. 180 TB of exp. data on tape (+50 TB raw data backup) 1.6 – 2 PB max tape capacity I/O bandwidth –> 1 GB/s cache clients – tape Hades DAQ end 2008: 200 MB/s => more tape drives needed 35 TB disk cache (1st level)

ITEE-Palaver gStore - GSI Mass Storage28 Final Remarks II gStore fully scalable in data capacity and I/O bandwidth supports several TSM servers gStore fully flexible in hardware (TSM) in the past years: –managed growth of > order of magnitude –handled various hardwares and platforms gStore prepared for further growth (FAIR) gStore adaptable for cooperation with external software packages