Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,

Slides:



Advertisements
Similar presentations
Tom Hamilton – America’s Channel Database CSE
Advertisements

© 2006 DataCore Software Corp SANmotion New: Simple and Painless Data Migration for Windows Systems Note: Must be displayed using PowerPoint Slideshow.
Ddn.com ©2012 DataDirect Networks. All Rights Reserved. GridScaler™ Overview Vic Cornell Application Support Consultant.
NAS vs. SAN 10/2010 Palestinian Land Authority IT Department By Nahreen Ameen 1.
PNFS, 61 th IETF, DC1 pNFS: Requirements 61 th IETF – DC November 10, 2004.
Performance, Reliability, and Operational Issues for High Performance NAS Matthew O’Keefe, Alvarri CMG Meeting February 2008.
Scale-out Central Store. Conventional Storage Verses Scale Out Clustered Storage Conventional Storage Scale Out Clustered Storage Faster……………………………………………….
Managing storage requirements in VMware Environments October 2009.
2 June 2015 © Enterprise Storage Group, Inc. 1 The Case for File Server Consolidation using NAS Nancy Marrone Senior Analyst The Enterprise Storage Group,
Chapter 9 Designing Systems for Diverse Environments.
Symantec De-Duplication Solutions Complete Protection for your Information Driven Enterprise Richard Hobkirk Sr. Pre-Sales Consultant.
An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool.
1 Principles of Reliable Distributed Systems Tutorial 12: Frangipani Spring 2009 Alex Shraer.
Comparison and Performance Evaluation of SAN File System Yubing Wang & Qun Cai.
High Performance Computing Course Notes High Performance Storage.
IBM® Spectrum Storage Virtualize™ V V7000 Unified in a nutshell
STORAGE Virtualization
How to Cluster both Servers and Storage W. Curtis Preston President The Storage Group.
Module – 7 network-attached storage (NAS)
Data Storage Willis Kim 14 May Types of storages Direct Attached Storage – storage hardware that connects to a single server Direct Attached Storage.
Simplifying Storage and Accelerating Results
Frangipani: A Scalable Distributed File System C. A. Thekkath, T. Mann, and E. K. Lee Systems Research Center Digital Equipment Corporation.
Enterprise Storage Our Journey Thus Far John D. Halamka MD CIO, Harvard Medical School and Beth Israel Deaconess Medical Center.
Windows ® Powered NAS. Agenda Windows Powered NAS Windows Powered NAS Key Technologies in Windows Powered NAS Key Technologies in Windows Powered NAS.
Object-based Storage Long Liu Outline Why do we need object based storage? What is object based storage? How to take advantage of it? What's.
SANPoint Foundation Suite HA Robert Soderbery Sr. Director, Product Management VERITAS Software Corporation.
COMPANY AND PRODUCT OVERVIEW Russ Taddiken Director of Principal Storage Architecture.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
Interposed Request Routing for Scalable Network Storage Darrell Anderson, Jeff Chase, and Amin Vahdat Department of Computer Science Duke University.
pNFS extension for NFSv4 IETF 61 November, 2004
1 Advanced Storage Technologies for High Performance Computing Sorin, Faibish EMC NAS Senior Technologist IDC HPC User Forum, April 14-16, Norfolk, VA.
Oracle Storage Overview Tomáš Vencelík – Storage sales leader.
October 2, 2015 pNFS extension for NFSv4 IETF-62 March 2005 Brent Welch
HEAnet Centralised NAS Storage Justin Hourigan, Senior Network Engineer, HEAnet Limited.
Chapter © 2006 The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/ Irwin Chapter 7 IT INFRASTRUCTURES Business-Driven Technologies 7.
FlashSystem family 2014 © 2014 IBM Corporation IBM® FlashSystem™ V840 Product Overview.
Protect Your Business-Critical Data in the Cloud with SoftNAS, a Full-Featured, Highly Available Solution for the Agile Microsoft Azure Platform MICROSOFT.
Storage Tank in Data Grid Shin, SangYong(syshin, #6468) IBM Grid Computing August 23, 2003.
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Large Scale Parallel File System and Cluster Management ICT, CAS.
Using NAS as a Gateway to SAN Dave Rosenberg Hewlett-Packard Company th Street SW Loveland, CO 80537
PNFS BOF FAST Sorin Faibish, EMC Mike Eisler, NetApp Brent Welch, Panasas Piyush Shivam, Sun Microsystems.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Terascala – Lustre for the Rest of Us  Delivering high performance, Lustre-based parallel storage appliances  Simplifies deployment, management and tuning.
VMware vSphere Configuration and Management v6
High Performance Storage Solutions April 2010 Larry Jones VP, Product Marketing.
ONStor Pantera 3110 ONStor NAS. Copyright 2008 · ONStor Confidential Pantera 3110 – An Integrated Channel only NAS  Integrated standalone NAS system.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
Oct 24, 2002 Michael Ernst, Fermilab DRM for Tier1 and Tier2 centers Michael Ernst Fermilab February 3, 2003.
Highest performance parallel storage for HPC environments Garth Gibson CTO & Founder IDC HPC User Forum, I/O and Storage Panel April 21, 2009.
PNFS Birds-of-Feather FAST 2010: February 24 Sorin Faibish, EMC and pNFS friends.
Padova, 5 October StoRM Service view Riccardo Zappi INFN-CNAF Bologna.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
3/12/2013Computer Engg, IIT(BHU)1 CLOUD COMPUTING-1.
Tackling I/O Issues 1 David Race 16 March 2010.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
An Introduction to GPFS
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
Reducing Risk with Cloud Storage
Large Scale Test of a storage solution based on an Industry Standard
Storage Virtualization
Module – 7 network-attached storage (NAS)
Scalable SoftNAS Cloud Protects Customers’ Mission-Critical Data in the Cloud with a Highly Available, Flexible Solution for Microsoft Azure MICROSOFT.
Presenter name goes here Presenter title goes here
Enterprise Class Virtual Tape Libraries
Presentation transcript:

Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2, 2011

Motivation Though dCache at BNL supports distributed analysis well (up to 5000 concurrent analysis jobs) we are looking into ways to improve the usability of our environment, e.g. include large-scale interactive analysis. We want to explore to what extent we can build on industrial products as far as the facility part is concerned w/o having to rely on community extensions and interfaces.  NFS 4.1 (pNFS) appealing because of performance, simplicity, level of integration w/ OS  BlueArc successfully used by PHENIX and STAR

Areas of Interest

Whys is NFS4.1(pNFS) attractive?

BlueArc System Performance

Transparent Data Mobility Tiered Storage for Data SAN Network Storage Cluster  Automatic and transparent data migration between tiers  Rules-based policy engine reduces manual intervention  Third-party or external storage devices as an integrated tier  Reduced dependence on high performance tier for peak demands The seamless migration of data across storage tiers within a single namespace Ease data management and reduce costs Back-end Storage SAN Tier 1 High Performance Tier 2 High Capacity Tier 0 Solid State Cache Tier 3 De-duplication Encryption Compression Existing NAS Users

BlueArc Platform Comparison Mercury 50Mercury 100Titan 3100 Product ClassLower Mid-rangeMid-range Cluster Nodes2 Up to 4Up to 8 Max Storage Capacity4 PB8 PB NFS Throughput700 MB/s1100 MB/s1200 MB/s Performance (specSFS IOPS)60,000100,000 Software / File ServicesAll software and filesystem options (NFS, CIFS, iSCSI) available Titan 3200 High End Up to 8 16 PB 200, MB/s Storage OptionsAll BlueArc storage array options are available with each platform © 2010 BlueArc Corporation

Traditional Network File System Proven architecture Enterprise features Open, standard protocols Open storage philosophy No Throughput Aggregation Metadata and data together BUT

How Does pNFS Change This? pNFS adds parallel I/O to the NFS protocol Eliminates the file server bottleneck Provides parallel data paths, even for a single file pNFS is part of the NFSv4.1 standard Approved by IETF Dec, 2008 RFCs completed editorial review Oct, 2009 RFC numbers issued Jan, 2010 Multiple implementations are in development Client software is expected to be embedded in leading OS distributions The Only Industry Standard Parallel File System

Key Elements of a pNFS System Separation of metadata and data –Metadata server with metadata file systems –Multiple data movers with data file systems –Different requirements for these Clients query metadata server for data location, then perform I/O directly and in parallel to/from data movers Configure for aggregated performance and policy driven value-based data placement 10

pNFS Architecture

BlueArc pNFS Based NAS Platform Highly scalable single metadata server –Clustered for HA –Architecture supports multiple clustered metadata servers Support for heterogeneous data servers Built on Enterprise Class Platform –Reliability –Full Featured NAS – Quota, Snapshots etc. 12

pNFS for HPC Performance Scaling Get Data Location Meta-Data File System Data files automatically spread across multiple data FS for load balancing and performance Individual data files optionally striped across multiple data FS for performance Extreme Flexibility: A single server can run the entire cluster (acting as both metadata server and data mover) Can scale performance by adding data movers (and relocate the FS) Can scale capacity by adding data file systems 13 Separate Data File Systems File A File B Direct IO Clients

Connectivity Initial release will use 10GbE connection to compute clusters BA exploring both 40GbE and 100GbE for next generation of servers, in

Product Availability and Test Bed at BNL BA’s pNFS – Mid 2011 –Up to 40 data movers with up to 8PB capacity each – growing to 128 data movers –Mercury - High performance metadata server –BlueArc Mercury or BlueArc Linux-based data mover appliances Performance Goals for Test Bed at BNL –Will become DDM endpoint w/ SRM, GridFTP, Xrootd –Up to 2k concurrent Analysis Jobs –Up to 4 GB/s throughput between storage system and clients –Aiming at having ~500 TB of disk space 15

Test Bed at BNL WA Network SRM GridFTP Xrootd