Performance, Reliability, and Operational Issues for High Performance NAS Matthew O’Keefe, Alvarri CMG Meeting February 2008.

Slides:



Advertisements
Similar presentations
Archive Task Team (ATT) Disk Storage Stuart Doescher, USGS (Ken Gacke) WGISS-18 September 2004 Beijing, China.
Advertisements

Proposed Storage Area Network Facilities For Discussion.
XenData SX-520 LTO Archive Servers A series of archive servers based on IT standards, designed for the demanding requirements of the media and entertainment.
Ddn.com ©2012 DataDirect Networks. All Rights Reserved. GridScaler™ Overview Vic Cornell Application Support Consultant.
NAS vs. SAN 10/2010 Palestinian Land Authority IT Department By Nahreen Ameen 1.
Consolidating Data Management with Enterprise Storage and Backup Systems Barry Ribbeck – Director, Systems, Architecture & Infrastructure, Rice University.
MUNIS Platform Migration Project WELCOME. Agenda Introductions Tyler Cloud Overview Munis New Features Questions.
Architecture and Implementation of Lustre at the National Climate Computing Research Center Douglas Fuller National Climate Computing Research Center /
Introduction to DBA.
2 June 2015 © Enterprise Storage Group, Inc. 1 The Case for File Server Consolidation using NAS Nancy Marrone Senior Analyst The Enterprise Storage Group,
NWfs A ubiquitous, scalable content management system with grid enabled cross site data replication and active storage. R. Scott Studham.
SAN Design Considerations Hylton Leigh Senior Consultant Novell Consulting, UK Stuart Thompson Senior Consultant Novell Consulting, UK.
Storage Area Network (SAN)
How to Cluster both Servers and Storage W. Curtis Preston President The Storage Group.
Designing Storage Architectures for Preservation Collections Library of Congress, September 17-18, 2007 Preservation and Access Repository Storage Architecture.
Session 3 Windows Platform Dina Alkhoudari. Learning Objectives Understanding Server Storage Technologies Direct Attached Storage DAS Network-Attached.
Data Storage Willis Kim 14 May Types of storages Direct Attached Storage – storage hardware that connects to a single server Direct Attached Storage.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
“Better together” PowerVault virtualization solutions
© 2010 IBM Corporation Kelly Beavers Director, IBM Storage Software Changing the Economics of Storage.
1© Copyright 2013 EMC Corporation. All rights reserved. EMC and Microsoft SharePoint Server Performance Name Title Date.
IBM TotalStorage ® IBM logo must not be moved, added to, or altered in any way. © 2007 IBM Corporation Break through with IBM TotalStorage Business Continuity.
Storage Area Networks The Basics. Storage Area Networks SANS are designed to give you: More disk space Multiple server access to a single disk pool Better.
Understand what’s new for Windows File Server Understand considerations for building Windows NAS appliances Understand how to build a customized NAS experience.
Bologna Aprile Atempo Product Suite Atempo Time Navigator™ Secure, highly scalable protection of heterogeneous data in complex, mission-critical.
Windows Server MIS 424 Professor Sandvig. Overview Role of servers Performance Requirements Server Hardware Software Windows Server IIS.
Chapter 10 : Designing a SQL Server 2005 Solution for High Availability MCITP Administrator: Microsoft SQL Server 2005 Database Server Infrastructure Design.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
© Copyright 2010 Hewlett-Packard Development Company, L.P. 1 HP + DDN = A WINNING PARTNERSHIP Systems architected by HP and DDN Full storage hardware and.
SGI Proprietary SGI Update IDC HPC User Forum September, 2008.
XenDesktop Built on FlexPod Flexible IT Infrastructure for Desktop Virtualization.
Chapter 8 Implementing Disaster Recovery and High Availability Hands-On Virtual Computing.
Meeting the Data Protection Demands of a 24x7 Economy Steve Morihiro VP, Programs & Technology Quantum Storage Solutions Group
Ceph Storage in OpenStack Part 2 openstack-ch,
Hadoop Hardware Infrastructure considerations ©2013 OpalSoft Big Data.
Page 110/12/2015 CSE 30341: Operating Systems Principles Network-Attached Storage  Network-attached storage (NAS) is storage made available over a network.
Protect Your Business-Critical Data in the Cloud with SoftNAS, a Full-Featured, Highly Available Solution for the Agile Microsoft Azure Platform MICROSOFT.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
Microsoft Azure Storage. Networking Compute Storage Virtual Machine Operating System Applications Data & Access Runtime Provision.
1 U.S. Department of the Interior U.S. Geological Survey Contractor for the USGS at the EROS Data Center EDC CR1 Storage Architecture August 2003 Ken Gacke.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
IST Storage & Backup Group 2011 Jack Shnell Supervisor Joe Silva Senior Storage Administrator Dennis Leong.
Using NAS as a Gateway to SAN Dave Rosenberg Hewlett-Packard Company th Street SW Loveland, CO 80537
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
VMware vSphere Configuration and Management v6
Click to add text Introduction to the new mainframe: Large-Scale Commercial Computing © Copyright IBM Corp., All rights reserved. Chapter 6: Accessing.
 The End to the Means › (According to IBM ) › 03.ibm.com/innovation/us/thesmartercity/in dex_flash.html?cmp=blank&cm=v&csr=chap ter_edu&cr=youtube&ct=usbrv111&cn=agus.
Ashish Prabhu Douglas Utzig High Availability Systems Group Server Technologies Oracle Corporation.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
SATA In Enterprise Storage Ron Engelbrecht Vice President and General Manager Engineering and Manufacturing Operations September 21, 2004.
Experiments in Utility Computing: Hadoop and Condor Sameer Paranjpye Y! Web Search.
STORAGE ARCHITECTURE/ MASTER): Where IP and FC Storage Fit in Your Enterprise Randy Kerns Senior Partner The Evaluator Group.
GPFS: A Shared-Disk File System for Large Computing Clusters Frank Schmuck & Roger Haskin IBM Almaden Research Center.
Tackling I/O Issues 1 David Race 16 March 2010.
1 CEG 2400 Fall 2012 Network Servers. 2 Network Servers Critical Network servers – Contain redundant components Power supplies Fans Memory CPU Hard Drives.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Architecture of a platform for innovation and research Erik Deumens – University of Florida SC15 – Austin – Nov 17, 2015.
An Introduction to GPFS
System Storage TM © 2007 IBM Corporation IBM System Storage™ DS3000 Series Jüri Joonsaar Tartu.
PHD Virtual Technologies “Reader’s Choice” Preferred product.
MERANTI Caused More Than 1.5 B$ Damage
iSCSI Storage Area Network
CERN Lustre Evaluation and Storage Outlook
Large Scale Test of a storage solution based on an Industry Standard
Storage Virtualization
Real IBM C exam questions and answers
Building continuously available systems with Hyper-V
PerformanceBridge Application Suite and Practice 2.0 IT Specifications
IBM Tivoli Storage Manager
Presentation transcript:

Performance, Reliability, and Operational Issues for High Performance NAS Matthew O’Keefe, Alvarri CMG Meeting February 2008

Cray’s Storage Strategy Background –Broad range of HPC requirements – big file I/O, small file I/O, scalability across multiple dimensions, data management, heterogeneous access… –Rate of improvement in I/O performance lags significantly behind Moore’s Law Direction –Move away from “one solution fits all” approach –Use cluster file system for supercomputer scratch space and focus on high performance –Use scalable NAS, combined with data management tools and new hardware technologies for shared and managed storage

3 Cray Advanced Storage Architecture (CASA) High Performance Scratch Files: Shared file system RAID disk Servers, Clusters, and Workstations 10G/1G Ethernet Network Fast NAS Fast NAS Fast NAS Fast NAS HSM/ Backup Server HSM/ Backup Server HSM/Backup Tape Robot/ Archive Virtual Tape Archive FC fabric

4 CASA Partners High Performance Scratch Files: Shared file system RAID disk Servers, Clusters, and Workstations 10G/1G Ethernet Network Fast NAS Fast NAS Fast NAS Fast NAS CFS – Lustre DDN, Engenio – RAID disk COPAN – VTL (MAID) BlueArc – scalable NAS ADIC – tapes ADIC – HSM TBD – backup HSM/Backup Tape Robot/ Archive HSM/ Backup Server HSM/ Backup Server FC fabric Virtual Tape Archive

5 CASA Lab Chippewa Falls Opteron Cluster Cisco 6509 switch Engenio Storage COPAN MAID

6 CASA Lab Opteron Cluster Related Storage

7 CASA Lab Blue Arc Titan Servers Engenio Storage

8 Blue Arc Titan-2 Dual Heads

9 CASA Lab COPAN Revolution System

10 How will this help? Use a cluster file system for big file (bandwidth) I/O for scalable systems –Focus on performance for applications Use commercial NAS products to provide solid storage for home directories and shared files –Vendors looking at NFS performance, scalability Use new technologies – nearline disk, virtual tape – in addition to or instead of physical tape for backup and data migration –Higher reliability and performance

11 Major HPC Storage Issue Too many HPC RFPs (esp for supercomputers) treat storage as secondary consideration –Storage “requirements” are incomplete or ill- defined Only performance requirement and/or benchmark is maximum aggregate bandwidth –No small files, no IOPS, metadata ops Requires “HSM” or “backup” with insufficient details No real reliability requirements –Selection criteria don’t give credit for a better storage solution Vendor judged on whether storage requirements are met or not

12 Why NFS? NFS is the basis of the NAS storage market (but CIFS important as well) –Highly successful, adopted by all storage vendors –Full ecosystem of data management and administration tools proven in commercial markets –Value propositions – ease of install and use, interoperability NAS vendors are now focusing on scaling NAS –Various technical approaches for increasing client and storage scalability Major weakness – performance –Some NAS vendors have been focusing on this –We see opportunities for improving this

13 CASA Lab Benchmarking CASA Lab in Chippewa Falls provides testbed to benchmark, configure and test CASA components –Opteron cluster (30 nodes) running Suse Linux –Cisco 6509 switch –BlueArc Titan — dual-heads, 6x1 Gigabit Ethernet on each head –Dual-fabric Brocade SAN with 4 FC controllers and 1 SATA controller –Small Cray XT3

14 Test and Benchmarking Methodology Used Bringsel tool (J. Kaitschuck — see CUG paper) –Measure reliability, uniformity, scalability and performance –Creates large, symmetric directory trees, varying file sizes, access patterns, block sizes –Allows testing of the operational behavior of a storage system: behavior under load, reliability, uniformity of performance Executed nearly 30 separate tests –Increasing complexity of access patterns and file distributions –Goal was to observe system performance across varying workloads

May 15Cray Technical Review 15 Quick Summary of Benchmarking Results  A total of over 400 TB of data has been written without data corruption or access failures  There have been no major hardware failures since testing began in August 2006 –predictable and relatively uniform.  with some exceptions, the BlueArc aggregate performance generally scales with the number of clients

Summary (continued)  Recovery from injected faults was fast and relatively transparent to clients 32 test cases have been prepared, about 28 of varying length have been run, all file checksums to date have been valid  Early SLES9 NFS client problems under load, detected and corrected via kernel patch; this led to the use of this patch at Cray’s AWE customer site, who experienced the same problem

17 Sequential Write Performance: Varying Block Size

18 Large File Writes: 8K Blocks

19 Large File, Random Writes, Variable Sized Blocks: Performance Approaches 500 MB/second for Single Head

20 Titan Performance at SC06

21 Summary of Results Performance generally uniform for given load Very small block size combined with random access performed poorly with SLES9 client –Much improved performance with SLES10 client Like cluster file systems, NFS performance sensitive to client behavior –SLES9 Linux NFS client failed under Bringsel load –Tests completed with SLES10 client Cisco link aggregation reduces performance by 30% at low node counts –Static assignment of nodes to Ethernet links increases performance –This effect goes away for 100s of NFS clients

22 Summary of Results BlueArc SAN backend provides performance baseline The Titan NAS heads cannot deliver more performance than these storage arrays make available –Need sufficient storage (spindles, array controllers) to meet IOPS and bandwidth goals –Stripe storage for each Titan head across multiple controllers to achieve best performance Test your NFS client with your anticipated workload against your NFS server infrastructure to set baseline performance

23 Summary BlueArc NAS storage meets Cray goals for CASA Performance tuning is a continual effort Next big push: efficient protocols and transfers between NFS server tier and Cray platforms iSCSI deployments for providing network disks for login, network, and SIO nodes Export SAN infrastructure from BlueArc to rest of data center Storage Tiers: fast FC block storage, BlueArc FC and SATA, MAID

24 Phase 0: Cluster File System Only Large Cray Supercomputer Lustre or Other Cluster File System All data lands and stays In the cluster file system Backup, HSM, other data management tasks all handled here Data sharing via file transfers

25 Phase 1: Cluster File System and Shared NAS Fast NAS 10G/1G Ethernet Network Fast NAS MAID Disk Archive Fast Nas Large Cray Supercomputer Lustre or Other Cluster File System Add NAS storage for data sharing between Cray and other machines NAS backup and archive support Long-term, managed data MAID for backup Separate storage networks for NAS and CFS stores GridFTP, other software, for sharing and data migration SGI Sun IBM X

26 Phase 2: Integrate NAS with Cray Platform 10G/1G Ethernet Network X Fast NAS MAID Disk Archive Fast Nas Large Cray Supercomputer Lustre or Other Cluster File System SGI Sun IBM Integrated Fast NAS Integrate fast NAS with Cray network: reduced NFS overhead, compute node access to shared NFS store Single file system name space: all NFS blades share same name space — internal and external MAID for backup and storage tier underneath NAS: FC versus ATA Separate storage networks for NAS and CFS

27 Phase 3: Integrated SANs 10G/1G Ethernet Network X Fast NAS MAID Disk Archive Fast Nas Large Cray Supercomputer Lustre or Other Cluster File System Storage Partition SGI Sun IBM Integrated Fast NAS Single, integrated Storage Area Network for improved efficiency and RAS Volume mirroring, snapshots, LUN management Partition storage freely between shared NAS store and the cluster file System Further integration of MAID storage tier into shared storage hierarchy SAN Director Switch

28 CASA 2.0 Hardware (Potential) Servers, Clusters, and Workstations 10G/1G Ethernet Network Fast NAS Fast NAS Fast NAS Fast NAS HSM/Backup Tape Robot/ Archive HSM/ Backup Server HSM/ Backup Server Virtual Tape Archive OST FC fabric SSD

May 15Cray Technical Review 29 Questions? Comments?

10G/1G Ethernet Network COPAN MAID Disk Tier of Commodity Cluster Hardware SAN Switch Automated Tape Library Tape Drives Tape Slots Tape Import/Export

CARAVITA: MAID for Virtual Vaulting (Primary Site) 10G/1G Ethernet Network Local Disk Array Server Tier SAN Switch Automated Tape Library Tape Import/Export Server pool includes: Local Media Server Application Servers Database Servers Other Servers