A quick introduction to SANs and Panasas ActivStor

Slides:



Advertisements
Similar presentations
Topics Networking Concepts Storage Concepts Presentation by Anand Mewalal Vmware VI3 Concepts On networking and Storage.
Advertisements

Archive Task Team (ATT) Disk Storage Stuart Doescher, USGS (Ken Gacke) WGISS-18 September 2004 Beijing, China.
SAN Storage for Database Systems Future storage subsystem options for database systems.
Data Storage Solutions Module 1.2. Data Storage Solutions Upon completion of this module, you will be able to: List the common storage media and solutions.
Virtualization and Cloud Computing Data center hardware David Bednárek, Jakub Yaghob, Filip Zavoral.
Storage RAID Types of RAID Protocols SAN Microsoft Clustering (MSCS) What is clustering Terminology How we are configured at Grey Grey File Services.
The Linux Storage People Simple Fast Massively Scalable Network Storage Coraid EtherDrive ® Storage.
Introduction to Storage Area Network (SAN) Jie Feng Winter 2001.
NAS vs. SAN 10/2010 Palestinian Land Authority IT Department By Nahreen Ameen 1.
Beyond NAS and SAN: The Evolution of Storage Marc Farley Author Building Storage Networks.
Chapter 5: Server Hardware and Availability. Hardware Reliability and LAN The more reliable a component, the more expensive it is. Server hardware is.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
Smart Storage and Linux An EMC Perspective Ric Wheeler
Vorlesung Speichernetzwerke Teil 2 Dipl. – Ing. (BA) Ingo Fuchs 2003.
Storage area Network(SANs) Topics of presentation
Modern Distributed Systems Design – Security and High Availability 1.Measuring Availability 2.Highly Available Data Management 3.Redundant System Design.
Fibre Channel Erin Keith CPE 401 Spring, Fibre Channel Storage Area Networks Overview Functionality Format Applications References.
High Performance Computing Course Notes High Performance Storage.
1 OASIS: Overlays and Active Services for Internetworked Storage Tal Lavian, Yin Li, George Porter, Mel Tsai, Randy H. Katz Computer Science Division Electrical.
COEN 180 NAS / SAN. NAS Network Attached Storage (NAS) Each storage device has its own network interface. Filers: storage device that interfaces at the.
SAN Design Considerations Hylton Leigh Senior Consultant Novell Consulting, UK Stuart Thompson Senior Consultant Novell Consulting, UK.
Storage Area Network (SAN)
© 2006 EMC Corporation. All rights reserved. Managing the Data Center Section 5.2.
Agenda CS C446 Data Storage Technologies & Networks
Copyright © 2009 EMC Corporation. Do not Copy - All Rights Reserved.
COEN 180 NAS / SAN. Storage Trends Storage Trends: Money is spend on administration Morris, Truskowski: The evolution of storage systems, IBM Systems.
Session 3 Windows Platform Dina Alkhoudari. Learning Objectives Understanding Server Storage Technologies Direct Attached Storage DAS Network-Attached.
Data Storage Willis Kim 14 May Types of storages Direct Attached Storage – storage hardware that connects to a single server Direct Attached Storage.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
National Energy Research Scientific Computing Center (NERSC) The GUPFS Project at NERSC GUPFS Team NERSC Center Division, LBNL November 2003.
Storage Cheap or Fast, Pick One. Storage Great--you can do a lot of computation. But this often generates a lot of data. Where are you going to put it?
RAID-x: A New Distributed Disk Array for I/O-Centric Cluster Computing Kai Hwang, Hai Jin, and Roy Ho.
1© Copyright 2013 EMC Corporation. All rights reserved. EMC XtremSW Cache Performance. Intelligence. Protection.
Storage Area Networks The Basics. Storage Area Networks SANS are designed to give you: More disk space Multiple server access to a single disk pool Better.
Managing Storage Lesson 3.
MaxTronic International Co., Ltd.
© 2006 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice Rajesh Goel Storage Presales Lead.
Module – 4 Intelligent storage system
 International Computers Limited, 2001 AXiS 20/3/2001 AXiS Trimetra User Group Glenn Fitzgerald Manager, Storage Solutions ICL High Performance Systems.
2014 LENOVO. ALL RIGHTS RESERVED. SAN Fundamentals.
Storage Tank in Data Grid Shin, SangYong(syshin, #6468) IBM Grid Computing August 23, 2003.
1 U.S. Department of the Interior U.S. Geological Survey Contractor for the USGS at the EROS Data Center EDC CR1 Storage Architecture August 2003 Ken Gacke.
Rafaella Luque 8vo “a”. A computer network is a group of computers that are connected to each other for the purpose of communication.
Using NAS as a Gateway to SAN Dave Rosenberg Hewlett-Packard Company th Street SW Loveland, CO 80537
Clustering In A SAN For High Availability Steve Dalton, President and CEO Gadzoox Networks September 2002.
Hosted by Minimizing the Impact of Storage on Your Network W. Curtis Preston President The Storage Group.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
Click to add text Introduction to the new mainframe: Large-Scale Commercial Computing © Copyright IBM Corp., All rights reserved. Chapter 6: Accessing.
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
EMC Proven Professional. Copyright © 2012 EMC Corporation. All Rights Reserved. NAS versus SAN NAS – Architecture to provide dedicated file level access.
ClinicalSoftwareSolutions Patient focused.Business minded. Slide 1 Opus Server Architecture Fritz Feltner Sept 7, 2007 Director, IT and Systems Integration.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
S AN D IEGO S UPERCOMPUTER C ENTER N ATIONAL P ARTNERSHIP FOR A DVANCED C OMPUTATIONAL I NFRASTRUCTURE 6 TB SSA Disk StorageTek Tape Libraries 830 GB MaxStrat.
Storage Virtualization
1 CEG 2400 Fall 2012 Network Servers. 2 Network Servers Critical Network servers – Contain redundant components Power supplies Fans Memory CPU Hard Drives.
The Network Aware IoT Service at Edge Guoxi Wang.
Managing Storage Module 3.
Oracle & HPE 3PAR.
Video Security Design Workshop:
Making an Ethernet Cable
Christof Hanke, HEPIX Spring Meeting 2008, CERN
Introduction to Networks
Acutelearn Technologies Tivoli Storage Manager(TSM) Training Tivoli Storage Manager Basics: Tivoli Storage Manager Overview Tivoli Storage Manager concepts.
ASM-based storage to scale out the Database Services for Physics
JDAT Production Hardware
Storage Networks and Storage Devices
Virtualization and Cloud Computing
Building continuously available systems with Hyper-V
Cloud Computing Architecture
Presentation transcript:

A quick introduction to SANs and Panasas ActivStor My initial experience with SANs was learning just enough to get the first NGS cluster operational; always had nagging doubt that I was missing a trick or two bought SANs for dummies and found, no, there really isn’t that much more to them. Original motivation for much of our SAN purchases was to support a RH GFS project to provide high performance access to storage for SCARF cluster users. GFS was troublesome to say the least, so we went back to good old ext3 + fileserver access model. Over time, more of the SAN-based storage is being used by other projects, such as VMWare, Oracle database + other NFS based file servers. Storage Networking Kevin Haines, eScience Centre, STFC, RAL

WikiPedia defines a SAN: A storage area network (SAN) is an architecture to attach remote computer storage devices (such as disk arrays, tape libraries, and optical jukeboxes) to servers in such a way that the devices appear as locally attached to the operating system. Contrast to Direct Attached Storage, usually found inside or just outside of the server, and Network Attached Storage using file based protocols such as NFS.

SAN Topology It’s like a LAN... Host/HBA 1 Storage 1 Host/HBA 2 SAN Switch

SAN Topology It’s like a LAN... it has a switch... Host/HBA 1 Storage 1 Host/HBA 2 Storage 2 Qlogic SANBox 5200 16 ports, two pairs of stacking ports (10Gb/s?) for expansion Host/HBA 3 SAN Switch

SAN Topology It’s like a LAN... it has a switch, and network cards. Host/HBA 1 Storage 1 Host/HBA 2 Storage 2 Called HBAs – host bus adaptors Host/HBA 3 SAN Switch

SAN Topology It’s like a LAN... it has a switch, and network cards. Storage 1 SAN Switch Storage 2 Host/HBA 1 Host/HBA 2 Host/HBA 3 Storage Arrays, may have one or two integral HBAs.

SAN Topology It’s like a LAN...but a little different. Fibre Connectors 2/4/8Gb/s Host/HBA 1 Storage 1 Host/HBA 2 Storage 2 Fibre connectors, not copper 2/4/8 Gb/s WWNs instead of MAC addresses Initiators and Targets HBAs don’t talk to each other (usually?) FC Protocol not Ethernet/IP SCSI over FCP Ultimate result is to present a SCSI device to the Host OS Host/HBA 3 SAN Switch Initiators Targets WWNs

Controlling Access Zoning implemented by the switch Host/HBA 1 Storage 1 Host/HBA 2 Storage 2 Ensure that HBAs only ‘see’ the storage we want them to. Not necessary, but a good precaution: NGS site incident, no zoning or LUN masking used, wrong device formatted wiping out home file system Fine on its own, as long as we’re dealing with entire storage arrays/devices as the targets, but if storage is partitioned for use by different HBAs, then Zoning on its own is not adequate. Host/HBA 3 SAN Switch Zone 1 = HBA1 + Storage1 Zone 2 = HBA2 + HBA3 + Storage 2

Controlling Access LUN masking on the storage array Host/HBA 1 LD1 HBA2 If partition your storage array, then LUN masking, implemented on storage array, Usually a list of HBAs WWNs that are allowed to read and write to the Logical Drive sometimes, can be a binary masking that includes/excludes a range of WWNs (limited use) LUN masking can be used on its own, or in combination with Zoning, but using both gives an extra level of protection from yourself. LD2 Host/HBA 3 HBA2 LD3 HBA3 SAN Switch Zone 1 = HBA1 + Storage1 Zone 2 = HBA2 + HBA3 + Storage 2

Controlling Access LUN masking on the storage array Host/HBA 1 LD1 HBA2 LD2 Host/HBA 3 HBA2 LD3 HBA3 SAN Switch Zone 1 = HBA1 + Storage1 Zone 2 = HBA2 + HBA3 + Storage 2

Increasing Resilience Multipath HOST HBA 1 SAN Switch 1 Storage 1 HBA 2 SAN Switch 2 Multipath adds resilience to failures in the HBA, SAN switch, and storage array channels, at cost of money and complexity. Two HBAs in the host, two SAN switches interconnected, and two channels to the storage array, there are 4 possible paths to the same storage in this diagram Host ‘sees’ 4 instances of the same disk, requires extra software to manage this complexity. In RHL this is devicemapper-multipath (0.4.5). Use dm-managed devices instead of /dev/sdb /dev/sdd etc. No personal experience of implementing this, nor of requiring it. SAN has been very reliable system, ~2-3 system failures in 5 years, and only 1 component failure.

Questions about SANs?

Panasas ActivStor Our goal (again) to provide high performance and resilient storage solution to the SCARF users. Wanted something easy to manage, scalable and affordable Bad experience with GFS, even if fixed - can’t afford to expand to > 600 SAN ports – HBAs, switches cabling etc. Evaluated Isilon & Panasas, tenders from both and selected Panasas.

Panasas ActivStor Sold in 4U shelf format, with 11 ‘blades’ per shelf, network switch (or two), redundant power and battery backup. All hot swappable. 2 types of blades. Director blades provide NFS access and meta-data + volume management. Up to 3 DBs per shelf. Storage blades provide 2 disks (up two 2TB per blade) and 2xGb/s ethernet (failover). Can have 11 SBs in shelf, but must be managed by a DB somewhere else.

Panasas ActivStor Director Blades Sold in 4U shelf format, with 11 ‘blades’ per shelf, network switch (or two), redundant power and battery backup. All hot swappable. 2 types of blades. Director blades provide NFS access and meta-data + volume management. Up to 3 DBs per shelf. Storage blades provide 2 disks (up two 2TB per blade) and 2xGb/s ethernet (failover). Can have 11 SBs in shelf, but must be managed by a DB somewhere else. Director Blades Meta-data and volume management services NFS access gateway

Panasas ActivStor Storage Blades Two disks (up to 2TB per blade) Sold in 4U shelf format, with 11 ‘blades’ per shelf, network switch (or two), redundant power and battery backup. All hot swappable. 2 types of blades. Director blades provide NFS access and meta-data + volume management. Up to 3 DBs per shelf. Storage blades provide 2 disks (up two 2TB per blade) and 2xGb/s ethernet (failover). Can have 11 SBs in shelf, but must be managed by a DB somewhere else. Storage Blades Two disks (up to 2TB per blade) 2xGb/s ethernet (failover mode) 11 per shelf (must be managed by DB)

Panasas ActivStor Resilience Resilience to failure is achieved by striping the data across the SBs, with one or more SBs worth of storage allocated as spare capacity. Can configure steadily decreasing number of SBs. E.g. Start with 3 then reduce number as necessary. It is not the case that one entire SB stands idle –every SB is involved in serving the clients, therefore getting full performance capability of system. Comprehensive System monitoring, detect early signs of blade failure and initiate blade drain for pre-emptive replacement. DBs provide volume management/metadata services to rest of system, as well as NFS, and can be configured for failover. Data is striped across all Storage Blades (RAID5 or RAID1) One or more Storage Blades’ worth of space reserved for failures System monitoring for pre-emptive actions (Blade Drain) Two or more Director Blades can provide failover for each other

Panasas ActivStor Performance DirectFLOW clients Available for most major Linux distributions Directly communicates with Storage Blades RAID computations performed by client ~5000 supported (12000 option) To get best performance, need to use the DirectFLOW client, implemented as a Linux Kernel module

Panasas ActivStor Performance DirectFLOW clients Available for most major Linux distributions Directly communicates with Storage Blades RAID computations performed by client ~5000 supported (12000 option) Results show scalable claim is true RAL: 1.2 GB/s from 22 nodes (2 shelves) RoadRunner: 60GB/s (103 shelves) Summary: Feel we’ve got a good balance between SAN attached storage, where fast attached storage needed, and high-performance network-based global filespace offered by Panasas.