Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.

Slides:



Advertisements
Similar presentations
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft File Systems for your Cluster Selecting a storage solution for tier 2 Suggestions and experiences.
Advertisements

Archive Task Team (ATT) Disk Storage Stuart Doescher, USGS (Ken Gacke) WGISS-18 September 2004 Beijing, China.
GridKa May 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Installing dCache into an existing Storage environment at GridKa Forschungszentrum.
A quick introduction to SANs and Panasas ActivStor
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.
Introduction to Storage Area Network (SAN) Jie Feng Winter 2001.
IWR Ideen werden Realität Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Institut für Wissenschaftliches Rechnen Status of Database Services.
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
Chapter 5: Server Hardware and Availability. Hardware Reliability and LAN The more reliable a component, the more expensive it is. Server hardware is.
Smart Storage and Linux An EMC Perspective Ric Wheeler
Storage area Network(SANs) Topics of presentation
vBrownBag – May 20, 2015 vSphere 6 Foundation Exam Section 3 – Storage
NWfs A ubiquitous, scalable content management system with grid enabled cross site data replication and active storage. R. Scott Studham.
VMware Infrastructure Alex Dementsov Tao Yang Clarkson University Feb 28, 2007.
STORAGE Virtualization
Storage Networking Technologies and Virtualization Section 2 DAS and Introduction to SCSI1.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Development of an API Standard in Interoperable Storage Networking Benjamin F. Kuo Troika Networks, Inc.
National Energy Research Scientific Computing Center (NERSC) The GUPFS Project at NERSC GUPFS Team NERSC Center Division, LBNL November 2003.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
Elad Hayun Agenda What's New in Hyper-V 2012 Storage Improvements Networking Improvements VM Mobility Improvements.
11 Capacity Planning Methodologies / Reporting for Storage Space and SAN Port Usage Bob Davis EMC Technical Consultant.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Managing Storage Lesson 3.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
Windows Server 2012 VSP Windows Kernel Applications Non-Hypervisor Aware OS Windows Server 2008, 2012 Windows Kernel VSC VMBus Emulation “Designed for.
Introducing Snap Server™ 700i Series. 2 Introducing the Snap Server 700i series Hardware −iSCSI storage appliances with mid-market features −1U 19” rack-mount.
CERN IT Department CH-1211 Geneva 23 Switzerland t Experience with NetApp at CERN IT/DB Giacomo Tenaglia on behalf of Eric Grancher Ruben.
Appendix B Planning a Virtualization Strategy for Exchange Server 2010.
Slide 1 DESIGN, IMPLEMENTATION, AND PERFORMANCE ANALYSIS OF THE ISCSI PROTOCOL FOR SCSI OVER TCP/IP By Anshul Chadda (Trebia Networks)-Speaker Ashish Palekar.
Virtualization in the NCAR Mass Storage System Gene Harano National Center for Atmospheric Research Scientific Computing Division High Performance Systems.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
DAC-FF The Ultimate Fibre-to-Fibre Channel External RAID Controller Solution for High Performance Servers, Clusters, and Storage Area Networks (SAN)
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
Sensitivity of Cluster File System Access to I/O Server Selection A. Apon, P. Wolinski, and G. Amerson University of Arkansas.
November 2, 2000HEPiX/HEPNT FERMI SAN Effort Lisa Giacchetti Ray Pasetes GFS information contributed by Jim Annis.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
Clustering In A SAN For High Availability Steve Dalton, President and CEO Gadzoox Networks September 2002.
1 Public DAFS Storage for High Performance Computing using MPI-I/O: Design and Experience Arkady Kanevsky & Peter Corbett Network Appliance Vijay Velusamy.
KIT – The cooperation of Forschungszentrum Karlsruhe GmbH und Universität Karlsruhe (TH) Hadoop on HEPiX storage test bed at FZK Artem Trunov.
MATSUMOTO Hitoshi SCSI support on Xen MATSUMOTO Hitoshi Fujitsu Ltd.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Tier-2 storage A hardware view. HEP Storage dCache –needs feed and care although setup is now easier. DPM –easier to deploy xrootd (as system) is also.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
KIT – University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association STEINBUCH CENTRE FOR COMPUTING - SCC
SA1 operational policy training, Athens 20-21/01/05 Presentation of the HG Node “Isabella” and operational experience Antonis Zissimos Member of ICCS administration.
BNL Oracle database services status and future plans Carlos Fernando Gamboa, John DeStefano, Dantong Yu Grid Group, RACF Facility Brookhaven National Lab,
VSphere 5 – Maximums – Virtual Machine Compute vCPUs per VM 32 Memory vRAM per VM 1TB Swap per VM 1TB Storage SCSI Adapto rs per VM 4 SCSI Targets per.
GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache Implementation at FZK Forschungszentrum Karlsruhe.
1 CEG 2400 Fall 2012 Network Servers. 2 Network Servers Critical Network servers – Contain redundant components Power supplies Fans Memory CPU Hard Drives.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
An Introduction to GPFS
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
12/19/01MODIS Science Team Meeting1 MODAPS Status and Plans Edward Masuoka, Code 922 MODIS Science Data Support Team NASA’s Goddard Space Flight Center.
INFN Site Report R.Gomezel October 9-13, 2006 Jefferson Lab, Newport News.
IT-DB Physics Services Planning for LHC start-up
Tape Drive Testing.
Cluster Active Archive
Christof Hanke, HEPIX Spring Meeting 2008, CERN
Scalable Database Services for Physics: Oracle 10g RAC on Linux
Direct Attached Storage and Introduction to SCSI
ASM-based storage to scale out the Database Services for Physics
Scalable Database Services for Physics: Oracle 10g RAC on Linux
High-Performance Storage System for the LHCb Experiment
Cost Effective Network Storage Solutions
Presentation transcript:

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel IWR, Forschungszentrum Karlsruhe Germany

CHEP04 The GridKa centre operates cluster computer for: D0, BaBar, CDF, Compass, LHC experiments (tier 1 for LHC) 500 dual CPU nodes, 220 TB disk, 400 TB tape expect growth to 1.6 PB disk, 4 PB tape in 2008 tape storage via dCache and Tivoli Storage Manager backend disk storage via NFS/GPFS

CHEP04 Overview Storage components at GridKa Cluster file system implementation Integration with Linux On line storage management Load balancing

CHEP04 Storage components (1) IO servers –dual Xeon 2.4 GHz, 1.5 GB RAM, Broadcom Ethernet –failover host bus adapter driver (Qlogic version 6.01) –RedHat 8, kernel on production cluster –RedHat ES 3 (Scientific Linux) on test cluster disks and RAID –disk 136 GB, 10 krpm –9 * 10 units of 14 disks: 1260 (36 hot spare) –arranged as RAID-5 volumes of 957 GB –stripe size 256 KB

CHEP04 Storage components (2) disk controllers (IBM FastT700) –to disk: 9 * 4 independent 2 Gb FC connections –to servers: 9 * 4 independent 2 Gb FC connections –reset or failure of (access to) one controller is handled without service interruption parallel cluster file system (GPFS) –each node of the storage cluster sees each disk –a partition is striped over 2 or more RAID volumes –file systems are exported via NFS –maximum size of single LUN is 1 TB

CHEP04 Cluster to storage connection GPFS cluster Worker nodes on NFS Disk collection Fibre channel switch SAN Ethernet

CHEP04 Linux parts SCSI driver –allows for hot adding disks/LUNs –no fixed relation between LUN ID and SCSI numbering. HBAs support persistent binding Fibre Channel driver –failover driver selects functional path –maximum number of LUNs on Qlogic FC HBA is 128 nfs server and nfs client –server side optimized, client default autofs and program maps –version (autofs4)

CHEP04 Maintenance and management Disk storage supports: –global hot-spares –on-line replaceable parts: controllers (incl fw), batteries, power supplies –background disk scrubbing LVM of GPFS allows for: –online replacement of volumes –expansion of file systems –online rebalancing after expansion

CHEP04 Storage load balancing At file system level –data transfers are striped over several raid volumes –storage is re-balanced on-line after expansion At server level –clients select servers at random –combination of autofs and DNS –introduce selection criteria (server capacity, service groups)

CHEP04 Server level load balancing read and write activity of last 24 hrs summed over all file servers read activity of production file servers

CHEP04 Presented solution benefits scalable size (4 PB) and large (15 TB) file spaces scalable performance (100 MB/server on single GE) native OS syscall API, no application code change needed on-line replaceable components reduce down time on-line storage expansion dynamic load balancing server load policies allows different server HW native Linux components on clients

CHEP04 Work to do get GPFS/NFS working on RH ES 3.0 integrate dCache into existing storage environment get DC to CERN and peer tier 1’s rolling start experimenting with SATA connect NFS servers with second Ethernet via Ethernet bonding introduce load policies

CHEP04 Thank you and colleagues from the GIS, GES and DASI departments at Institute for scientific computing (IWR), Karlsruhe