Andrei Moskalenko Storage team, Centre de Calcul de l’ IN2P3. HPSS – The High Performance Storage System Storage at the Computer Centre of the IN2P3 HEPiX.

Slides:



Advertisements
Similar presentations
© 2006 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice Enigma Data’s SmartMove.
Advertisements

CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000.
Distributed Database Management Systems
Jean-Yves Nief, CC-IN2P3 Wilko Kroeger, SCCS/SLAC Adil Hasan, CCLRC/RAL HEPiX, SLAC October 11th – 13th, 2005 BaBar data distribution using the Storage.
Chapter 12 Distributed Database Management Systems
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
Session 3 Windows Platform Dina Alkhoudari. Learning Objectives Understanding Server Storage Technologies Direct Attached Storage DAS Network-Attached.
© 2010 IBM Corporation Kelly Beavers Director, IBM Storage Software Changing the Economics of Storage.
CT NIKHEF June File server CT system support.
IBM TotalStorage ® IBM logo must not be moved, added to, or altered in any way. © 2007 IBM Corporation Break through with IBM TotalStorage Business Continuity.
CC - IN2P3 Site Report Hepix Fall meeting 2009 – Berkeley
Building Advanced Storage Environment Cheng Yaodong Computing Center, IHEP December 2002.
High Performance Storage System Harry Hulen
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
File and Object Replication in Data Grids Chin-Yi Tsai.
DATABASE MANAGEMENT SYSTEMS IN DATA INTENSIVE ENVIRONMENNTS Leon Guzenda Chief Technology Officer.
Week 5 Lecture Distributed Database Management Systems Samuel ConnSamuel Conn, Asst Professor Suggestions for using the Lecture Slides.
Storage Tank in Data Grid Shin, SangYong(syshin, #6468) IBM Grid Computing August 23, 2003.
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
20-22 September 1999 HPSS User Forum, Santa Fe CERN IT/PDP 1 History  Test system HPSS 3.2 installation in Oct 1997 IBM AIX machines with IBM 3590 drives.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
Author - Title- Date - n° 1 Partner Logo WP5 Summary Paris John Gordon WP5 6th March 2002.
National Partnership for Advanced Computational Infrastructure San Diego Supercomputer Center Persistent Management of Distributed Data Reagan W. Moore.
Large Scale Parallel File System and Cluster Management ICT, CAS.
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
 CASTORFS web page - CASTOR web site - FUSE web site -
CASTOR: CERN’s data management system CHEP03 25/3/2003 Ben Couturier, Jean-Damien Durand, Olof Bärring CERN.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY HPSS Features and Futures Presentation to SCICOMP4 Randy Burris ORNL’s Storage Systems Manager.
8 October 1999 BaBar Storage at CCIN2P3 p. 1 Rolf Rumler BaBar Storage at Lyon HEPIX and Mass Storage SLAC, California, U.S.A. 8 October 1999 Rolf Rumler,
INTRODUCTION TO DBS Database: a collection of data describing the activities of one or more related organizations DBMS: software designed to assist in.
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
 Introduction  Architecture NameNode, DataNodes, HDFS Client, CheckpointNode, BackupNode, Snapshots  File I/O Operations and Replica Management File.
Padova, 5 October StoRM Service view Riccardo Zappi INFN-CNAF Bologna.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
W.A.Wojcik/CCIN2P3, Nov 1, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
CASTOR project status CASTOR project status CERNIT-PDP/DM October 1999.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Hands-On Microsoft Windows Server 2008 Chapter 7 Configuring and Managing Data Storage.
12 Mars 2002LCG Workshop: Disk and File Systems1 12 Mars 2002 Philippe GAILLARDON IN2P3 Data Center Disk and File Systems.
New Development Efforts in GridFTP Raj Kettimuthu Math & Computer Science Division, Argonne National Laboratory, Argonne, IL 60439, U.S.A.
GDB meeting - Lyon - 16/03/05 An example of data management in a Tier A/1 Jean-Yves Nief.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
CC-IN2P3 Pierre-Emmanuel Brinette Benoit Delaunay IN2P3-CC Storage Team 17 may 2011.
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
Data Infrastructure in the TeraGrid Chris Jordan Campus Champions Presentation May 6, 2009.
An Introduction to GPFS
CommVault Architecture
SRB at KEK Yoshimi Iida, Kohki Ishikawa KEK – CC-IN2P3 Meeting on Grids at Lyon September 11-13, 2006.
CCIN2P3 Site Report - BNL, Oct 18, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center.
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
CASTOR: possible evolution into the LHC era
GridOS: Operating System Services for Grid Architectures
StoRM: a SRM solution for disk based storage systems
Introduction to Data Management in EGI
SAM at CCIN2P3 configuration issues
The INFN Tier-1 Storage Implementation
Research Data Archive - technology
Kirill Lozinskiy NERSC Storage Systems Group
CASTOR: CERN’s data management system
INFNGRID Workshop – Bari, Italy, October 2004
IBM Tivoli Storage Manager
STATEL an easy way to transfer data
Presentation transcript:

Andrei Moskalenko Storage team, Centre de Calcul de l’ IN2P3. HPSS – The High Performance Storage System Storage at the Computer Centre of the IN2P3 HEPiX Spring Meeting 2006

HPSS is a Highly Scalable Storage System that provides hierarchical storage management (HSM) Quality of Services global name space, ACLs, security (DCE, Kerberos, GSS API) control and programming interfaces: DMAPI, POSIX API (Extended POSIX API) disk and tape data striping; transfer striping (over multiple TCP connections) data replication (double copy, etc.) Design-wise scalable architecture (achieved by adding more storage and control elements) network-centric (LAN and SAN). What's HPSS ? HPSS – The High Performance Storage System

free disk spaces through an over allocation mechanism transparency by hiding complex machinery from users Hierarchical Storage Management provides HPSS – The High Performance Storage System HSM might also provide Dynamic Resource management : resources allocated when they needed and where they needed Optimisation/Performance/Administration ( designed to handle different types of resources) Possibility to integrate multiple storage tiers, not just two (HPSS handles 5 levels)

HPSS 5.1 architecture HPSS – The High Performance Storage System CORE SERVER PVLPVR DISK MOVER TAPE MOVER LSGK NDCG MIGRATION / PURGE SERVER DB2 DCE

HPSS Configuration HPSS – The High Performance Storage System Classes of Service, Hierarchies, Storage Classes COS 10 for files< 30 MB COS 11 < 300 MB COS 12 < 4 GB COS 13 < 16 GB COS 14 < 80 GB Migration and Purge policies (Dynamically tunable) Name Space and Storage Subsystems (each subsystem controls its part/parts of the global name space and storage resources) Filesets, Junctions Families (assure that logically aggregated data is localised on media to reduce tape mounts) Machinery: automatic repack, reclaim, etc.

Storage Hierarchies HPSS – The High Performance Storage System Hierarchy 1 SClass 1 SClass 2 SClass 3 COS 1COS 11

Storage Hierarchies HPSS – The High Performance Storage System Hierarchy 1 SClass 1 SClass 2 SClass 3 COS 1COS 11 Hierarchy 2 SClass 1 SClass 2SClass 3 COS 2

Examples of COS definitions HPSS - The High Performance Storage System Disks only with two tiers of disks Tapes only Disks + Tapes ( two tiers) Disks + Tapes (with double copy) Disks + Tapes (multiple tiers)

Subsystems and Global Name Space HPSS – The High Performance Storage System CORE 3: fs3 / HPSS_ROOT CORE 2: fs2 CORE 1: fs1 junction CORE 3: fs6 junction CORE 1: fs5 junction CORE 2: fs4 junction

HPSS IBM web resources HPSS – The High Performance Storage System HPSS Technology HPSS Brochure Introductory Presentation for the HPSS Administrators Course High Performance Storage System Scalability: Architecture Implementation and Experience Storage Area Networks and the High Performance Storage System High Availability, etc

MSS and Computing Infrastructure at the CC-IN2P3 HPSS – The High Performance Storage System batch worker dCache XRootd bbFTP GridFTP SRB HPSS batch worker HSI transfers communicationsrfio servers

at our site: HPSS – The High Performance Storage System RFIO RFIO API POSIX like interface + extensions (readlist, writelist, setcos) RFIO commands: rfcp uses the readlist/writelist inteface that allows the data to flow directly from the client to the allocated hpss server (disk or tape mover) Clients Xrootd, dCache (via RFIO commands) bbftp, gridftp (via RFIO API) SRB (via HPSS-API, NDCG)

HPSS at the CC (IN2P3). Facts. HPSS – The High Performance Storage System Data Volume doubles annually: 1.5 PB now, 2.5 PB at the end of 2006 Up to 30TB of data transfers per day: 100MB/s with rfcp Up to tape mounts per day cartridges; B/200GB; /20GB 3 Subsystems. More in the future + a cluster for the Core Servers 32 disk servers: 36 TB/1100 disks/ 238 movers (600 movers); 28 tape servers: B and drives

Evolution... HPSS – The High Performance Storage System From 1 to 3 Subsystems (more in the future) Extra external control (sophistication of BQS resource definition: hpss, u_hpss_cms_xxx, u_rfio_cms_xxx). MSS-BQS autoregulation mechanism? RFIO connections watch-dog Internal control: fewer disk per software mover/ more movers per node Development of more sophisticated repack mechanism/policy Introduction of a 3rd tier (Disk to Tape to Tape) Tests: crash scenarios, system limits, error messages reproduction/correlations 9840 to 9940B migration

HPSS 5.1 – Daily routine and technical skills HPSS – The High Performance Storage System Skills UNIX + DCE + LAN + SAN (administration, optimisation, etc.) DB2 DBA (backup, recovery, optimisation + optimisation + optimisation) Robotics (tons of tape mounts per day, incidents, earthquakes, etc.) Some knowledge of Java, CAs, etc. special: data modelling, batch system resource modelling, etc. Routine 550 Tape/Drive incidents (prem. EOMs and worse) for 2005 RFIO log analysis (always something new) Troubleshooting occasional anomalies due to complexity (rfio bugs, hidden timeouts, etc.) User/Client Support + Supervision (But admins also need support, doughnuts and coffee…) Small files …… Resources planning and allocation Repacks: s (9840 ->9940B); ~ Bs

Bottlenecks, traps and inefficiencies HPSS – The High Performance Storage System Data isolation/clustering/aggregation (ideally on a per user/function/usage/type basis… ) is the most crucial task. New data should not get on the same storage media as old data (repacks) Easy to write, but difficult to read (but writes/reads = 15/85) Repacked Volume ~ Stored Volume (if data badly clustered) Repacks and Deletes make DB2 tables volatile. Volatile tables with 20 million entries are an optimisation nightmare. Badly isolated data => too many mounts If we let users choose they tend to make a lot of mistakes (wrong COS, small files, MD5 integrity tests run by users, etc.) Weak support for tape errors, insufficient error message reporting, not so many administrative tools, non dynamic configuration...

HPSS Enterprise HSM Services. HPSS 6.2 Client Access Overview HPSS – The High Performance Storage System. A slide from Harry Hulen’s presentation of HPSS,

Interfacing with HPSS HPSS – The High Performance Storage System RFIO at our site HPSS API = POSIX Client API (CLAPI) extends POSIX API to include COSs, striping, etc. (LINUX VFS will provide access to the HPSS CLAPI ) Hierarchical Storage Interface. HSI can provide information about file locations (ex., tape id and the exact position) FTP, Parallel File Transfer Protocol (PFTP) Interface Data Management API (DMAPI) (will be interfaced with Linux XFS,GPFS) Linux NFSv4 and SAMBA GridFTP native support, etc

at our site: HPSS – The High Performance Storage System RFIO RFIO API POSIX like interface + extensions (readlist, writelist, setcos) RFIO commands: rfcp uses the readlist/writelist inteface that allows the data to flow directly from the client to the allocated hpss server (disk or tape mover) Xrootd, dCache (via RFIO commands) SRB (via HPSS-API, NDCG) bbftp, gridftp (via RFIO API)

Conclusions HPSS – The High Performance Storage System HPSS is an excelent HMS system that provides highly scalable storage and archival services (SUs) provides (SAN-centred) global file system functionalities is capable to host 10s of PB of data (100s PB?) provides support for scalable, parallel I/O operations scales to 10s of TB daily throughput (100s TB ) does not impose any unreasonable restriction on your storage models is highly modular (new technology, evolution) is very robust (data replication, HA)

Conclusions HPSS – The High Performance Storage System HPSS is an excellent HMS system that provides highly scalable storage and archival services (SUs) provides (SAN-centred) global file system functionalities is capable to host 10s and 100s PB of data provides support for scalable, parallel I/O operations scales to 10s and 100s TB daily throughput does not impose any unreasonable restriction on your storage models is highly modular (new technology, evolution) is very robust (data replication, HA) What one would possibly like to see is more sophisticated migration/purge policies (dCache-like) tools/utilities for data migration (exploiting meta-data, not just copy) more sophisticated inter-subsystem name space partitioning take advantage of all SAN functionalities as soon as possible a better incident and error message control