HPSS The High Performance Storage System Developed by IBM, LANL, LLNL, ORNL, SNL, NASA Langley, NASA Lewis, Cornell, MHPCC, SDSC, UW with funding from.

Slides:



Advertisements
Similar presentations
The Globus Striped GridFTP Framework and Server Bill Allcock 1 (presenting) John Bresnahan 1 Raj Kettimuthu 1 Mike Link 2 Catalin Dumitrescu 2 Ioan Raicu.
Advertisements

Archive Task Team (ATT) Disk Storage Stuart Doescher, USGS (Ken Gacke) WGISS-18 September 2004 Beijing, China.
DISTRIBUTED FILE SYSTEMS Computer Engineering Department Distributed Systems Course Asst. Prof. Dr. Ahmet Sayar Kocaeli University - Fall 2013.
XenData SX-520 LTO Archive Servers A series of archive servers based on IT standards, designed for the demanding requirements of the media and entertainment.
XenData SX-10 LTO Archive Appliance An Archive Appliance based on IT standards, designed for the demanding requirements of the media and entertainment.
High Performance Cluster Computing Architectures and Systems Hai Jin Internet and Cluster Computing Center.
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
What to expect.  Linux  Windows Server (2008 or 2012)
CS-550: Distributed File Systems [SiS]1 Resource Management in Distributed Systems: Distributed File Systems.
GridFTP Introduction – Page 1Grid Forum 5 GridFTP Steve Tuecke Argonne National Laboratory.
Network-Attached Storage
 Introduction Originally developed by Open Software Foundation (OSF), which is now called The Open Group ( Provides a set of tools and.
Vorlesung Speichernetzwerke Teil 2 Dipl. – Ing. (BA) Ingo Fuchs 2003.
File System Implementation
Copyright © Clifford Neuman - UNIVERSITY OF SOUTHERN CALIFORNIA - INFORMATION SCIENCES INSTITUTE CS582: Distributed Systems Lecture 19 - November.
INTRODUCTION OS/2 was initially designed to extend the capabilities of DOS by IBM and Microsoft Corporations. To create a single industry-standard operating.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
Introduction to the new mainframe: Large-Scale Commercial Computing © Copyright IBM Corp., All rights reserved. Chapter 6: Accessing large amount.
Storage Area Network (SAN)
Storage Networking. Storage Trends Storage growth Need for storage flexibility Simplify and automate management Continuous availability is required.
Data Storage Willis Kim 14 May Types of storages Direct Attached Storage – storage hardware that connects to a single server Direct Attached Storage.
RAID-x: A New Distributed Disk Array for I/O-Centric Cluster Computing Kai Hwang, Hai Jin, and Roy Ho.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
To provide the world with a next generation storage platform for unstructured data, enabling deployment of mobile applications, virtualization solutions,
Chapter 8 Implementing Disaster Recovery and High Availability Hands-On Virtual Computing.
Virtualization in the NCAR Mass Storage System Gene Harano National Center for Atmospheric Research Scientific Computing Division High Performance Systems.
Latest Relevant Techniques and Applications for Distributed File Systems Ela Sharda
What is a Distributed File System?? Allows transparent access to remote files over a network. Examples: Network File System (NFS) by Sun Microsystems.
Mark A. Magumba Storage Management. What is storage An electronic place where computer may store data and instructions for retrieval The objective of.
20-22 September 1999 HPSS User Forum, Santa Fe CERN IT/PDP 1 History  Test system HPSS 3.2 installation in Oct 1997 IBM AIX machines with IBM 3590 drives.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY HPSS Features and Futures Presentation to SCICOMP4 Randy Burris ORNL’s Storage Systems Manager.
Storage and Storage Access 1 Rainer Többicke CERN/IT.
Distributed Object Frameworks DCE and CORBA. Distributed Computing Environment (DCE) Architecture proposed by OSF Goal: to standardize an open UNIX envt.
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
11 CLUSTERING AND AVAILABILITY Chapter 11. Chapter 11: CLUSTERING AND AVAILABILITY2 OVERVIEW  Describe the clustering capabilities of Microsoft Windows.
VMware vSphere Configuration and Management v6
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition File System Implementation.
Click to add text Introduction to the new mainframe: Large-Scale Commercial Computing © Copyright IBM Corp., All rights reserved. Chapter 6: Accessing.
1 Status of HPSS New Features, Requirements, and Installations Otis Graf IBM Global Services - Federal Houston, Texas October 1999.
WINDOWS SERVER 2003 Genetic Computer School Lesson 12 Fault Tolerance.
Andrei Moskalenko Storage team, Centre de Calcul de l’ IN2P3. HPSS – The High Performance Storage System Storage at the Computer Centre of the IN2P3 HEPiX.
Chapter 9: Networking with Unix and Linux. Objectives: Describe the origins and history of the UNIX operating system Identify similarities and differences.
© 2006 EMC Corporation. All rights reserved. The Host Environment Module 2.1.
Storage Why is storage an issue? Space requirements Persistence Accessibility Needs depend on purpose of storage Capture/encoding Access/delivery Preservation.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
CASTOR project status CASTOR project status CERNIT-PDP/DM October 1999.
CEG 2400 FALL 2012 Linux/UNIX Network Operating Systems.
Hands-On Microsoft Windows Server 2008 Chapter 7 Configuring and Managing Data Storage.
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
NASA Langley Research Center’s Distributed Mass Storage System (DMSS) Juliet Z. Pao Guest Lecturing at ODU April 8, 1999.
An Introduction to GPFS
Network-Attached Storage. Network-attached storage devices Attached to a local area network, generally an Ethernet-based network environment.
XenData SX-10 LTO Archive Appliance
Chapter 1: Introduction
Storage Networking.
Introduction to Data Management in EGI
File System Implementation
Chapter 12: Mass-Storage Structure
Unit OS10: Fault Tolerance
Research Data Archive - technology
Storage Virtualization
XtreemFS Olga Rocheeva
Storage Networking.
XenData SX-550 LTO Archive Servers
San Diego Supercomputer Center
AGENT OS.
Outline Operating System Organization Operating System Examples
Improving performance
Presentation transcript:

HPSS The High Performance Storage System Developed by IBM, LANL, LLNL, ORNL, SNL, NASA Langley, NASA Lewis, Cornell, MHPCC, SDSC, UW with funding from DoE, NASA & NSF Presented by Christopher Ho, CSci 599

Motivation n In last 10 years, processor speeds have increased 50-fold n Disk transfer rates have increased < 4 X u RAID now successful, inexpensive n Tape speeds have increased < 4 X u tape striping not widespread n Performance gap is widening! n Bigger & bigger files (10s, 100s of GB, soon TB) n => Launch scalable storage initiative

IEEE Mass Storage Reference Model n Defines layers of abstraction & transparency u device, location independence n Separation of policy and mechanism n Logical separation of control and data flow n Defines common terminology u compliance does not imply inter-operability n Scalable, Hierarchical Storage Management n see

Introduction: Hierarchical Storage n Storage pyramid Magnetic Tape Memory Disk Optical disk Decreasing cost & speed, Increasing capacity

HPSS Objectives n Scalable u transfer rate, file size, name space, geography n Modular u software subsystems replaceable, network/tape technologies updateable, API access n Portable u multiple vendor platforms, no kernel modifications, multiple storage technologies, standards-based, leverage commercial products

HPSS Objectives (cont) n Reliable u distributed software and hardware components u atomic transactions u mirror metadata u failed/restarted servers can reconnect u storage units can be varied on/offline

Access into HPSS n FTP u protocol already supports 3rd party transfers u new: partial file transfer (offset & size) n Parallel FTP u pget, pput, psetblocksize, psetstripewidth n NFS version 2 u most like traditional file system, slower than FTP n PIOFS u parallel distributed FS on IBM SP2 MPP n futures: AFS/DCE DFS, DMIG-API

HPSS architecture Network Attached Disk Network Attached Tape HiPPI/FC/ATM data network I/O node MPP interconnect Processing node HPSS server Storage System Mgmt I/O node NFS FTP DMIG-API - NETWORK Control Network

Software infrastructure n Encina transaction processing manager u two-phase commit, nested transactions u guarantees consistency of metadata, server state n OSF Distributed Computing Environment u RPC calls for control messages u Thread library u Security (registry & privilege service) F Kerberos authentication n 64 bit Arithmetic functions u file sizes up to 2^64 bytes u 32 bit platforms, big/little endian architectures

Software components n Name server u map POSIX filenames to internal file, directory or link n Migration/Purge policy manager u when/where to migrate to next level in hierarchy u after migrated, when to purge copy on this level F purge initiated when usage exceeds administrator- configured high-water mark F each file evaluated by size, time since last read u migration, purge can also be manually initiated

Software components (cont) n Bitfile server u provides abstraction of bitfiles to client u provides scatter/gather capability u supports access by file offset, length u supports random and parallel reads/writes u works with file segment abstraction (see Storage server)

Software components (cont) n Storage server u map segments onto virtual volumes, virtual volumes onto physical volumes u virtual volumes allow tape striping n Mover u transfers data from a source to a sink F tape, disk, network, memory u device control: seek, load/unload, write tape mark, etc.

Software components (cont) n Physical Volume Library u map physical volume to cartridge, cartridge to PVR n Physical Volume Repository u control cartridge mount/dismount functions u modules for Ampex D2, STK 4480/90 & SD-3, IBM 3480 & 3590 robotic libraries n Repack server u deletions leave gaps on sequential media F read live data, rewrite on new sequential volume, free up previous volume

Software components (cont) n Storage system management u GUI to monitor/control HPSS u stop/start software servers u monitor events and alarms, manual mounts u vary devices on/offline

Parallel transfer protocol - goals n Provide parallel data exchange between heterogeneous systems and devices n Support different combinations of parallel and sequential source/sink n Support gather/scatter and random access u combinations of stripe width, both regular and irregular data block size n Scalable I/O bandwidth n Transport independent (TCP/IP, HiPPI, FCS, ATM)

Gather/scatter lists S1S2S3AB logical window ABABABxyzxyzxyz D1D2D3

Parallel transport architecture S1SnD1Dn parallel data flow control connections client control connections

Parallel FTP transfer (pget) Mover Storage server Bitfile server Name serverParallel FTPd Client mover Parallel FTP client

Summary n High performance u up to 1 GB/s aggregate transfer rates n Scalable storage u parallel architecture u terabyte-sized files u petabytes in archive n Robust u transaction processing manager n Portable u IBM, Sun implementations available

Conclusion n Feasability has been demonstrated for large, scalable storage n Software exists, is shipping, and is actively used in the national labs on a daily basis n Distributed architecture and parallel capabilities mesh well with grid computing