Peter Stefan, NIIF 29 June, 2007, Amsterdam, The Netherlands NIIF Storage Services Collaboration on Storage Services.

Slides:



Advertisements
Similar presentations
Archive Task Team (ATT) Disk Storage Stuart Doescher, USGS (Ken Gacke) WGISS-18 September 2004 Beijing, China.
Advertisements

Copyright 2005, Coraid, Inc Low Cost Networked Storage Using ATA-over-Ethernet Brantley Coile Coraid, Inc
The Linux Storage People Simple Fast Massively Scalable Network Storage Coraid EtherDrive ® Storage.
System Area Network Abhiram Shandilya 12/06/01. Overview Introduction to System Area Networks SAN Design and Examples SAN Applications.
NAS vs. SAN 10/2010 Palestinian Land Authority IT Department By Nahreen Ameen 1.
© 2009 IBM Corporation Statements of IBM future plans and directions are provided for information purposes only. Plans and direction are subject to change.
Network-Attached Storage
Server Platforms Week 11- Lecture 1. Server Market $ 46,100,000,000 ($ 46.1 Billion) Gartner.
NWfs A ubiquitous, scalable content management system with grid enabled cross site data replication and active storage. R. Scott Studham.
© 2008 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice HP StorageWorks LeftHand update Marcus.
Storage Area Network (SAN)
Storage Networking Technologies and Virtualization Section 2 DAS and Introduction to SCSI1.
Emerging Storage Options for Server Blade Architectures Server Blade Summit 2005.
Module – 7 network-attached storage (NAS)
© 2009 IBM Corporation Statements of IBM future plans and directions are provided for information purposes only. Plans and direction are subject to change.
Servers Redundant Array of Inexpensive Disks (RAID) –A group of hard disks is called a disk array FIGURE Server with redundant NICs.
Storage Networking. Storage Trends Storage growth Need for storage flexibility Simplify and automate management Continuous availability is required.
BACKUP/MASTER: Immediate Relief with Disk Backup Presented by W. Curtis Preston VP, Service Development GlassHouse Technologies, Inc.
Storage Area Networks The Basics. Storage Area Networks SANS are designed to give you: More disk space Multiple server access to a single disk pool Better.
Storage Survey and Recent Acquisition at LAL Michel Jouvin LAL / IN2P3
Object-based Storage Long Liu Outline Why do we need object based storage? What is object based storage? How to take advantage of it? What's.
Module 10 Configuring and Managing Storage Technologies.
Oracle 10g Database Storage Demystified Jeff Browning, O.C.P, R.H.C.A. Senior Manager Network Appliance, Inc. OracleWorld 2003 San Francisco.
1 An Introduction to the Storage Bridge Bay Specification.
High Performance Computing G Burton – ICG – Oct12 – v1.1 1.
Buffalo Data Storage Expansion June As organizations grow the amount of data storage capacity required to support it grows as well Increased data.
Protocol Architectures. Simple Protocol Architecture Not an actual architecture, but a model for how they work Similar to “pseudocode,” used for teaching.
Small File File Systems USC Jim Pepin. Level Setting  Small files are ‘normal’ for lots of people Metadata substitute (lots of image data are done this.
Chapter 5 Section 2 : Storage Networking Technologies and Virtualization.
School of EECS, Peking University Microsoft Research Asia UStore: A Low Cost Cold and Archival Data Storage System for Data Centers Quanlu Zhang †, Yafei.
Storage Systems Market Analysis Dec 04. Storage Market & Technologies.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
The University of Bolton School of Games Computing & Creative Technologies LCT2516 Network Architecture CCNA Exploration LAN Switching and Wireless Chapter.
1 U.S. Department of the Interior U.S. Geological Survey Contractor for the USGS at the EROS Data Center EDC CR1 Storage Architecture August 2003 Ken Gacke.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Storage Networking Evolution Jim Morin VP Strategic Planning June 2001.
Microsoft Virtual Academy Module 8 Managing the Infrastructure with VMM.
IST Storage & Backup Group 2011 Jack Shnell Supervisor Joe Silva Senior Storage Administrator Dennis Leong.
© 2008 Cisco Systems, Inc. All rights reserved.Cisco ConfidentialPresentation_ID 1 Chapter 1: Introduction to Scaling Networks Scaling Networks.
Hosted by Minimizing the Impact of Storage on Your Network W. Curtis Preston President The Storage Group.
CLASS Information Management Presented at NOAATECH Conference 2006 Presented by Pat Schafer (CLASS-WV Development Lead)
Production Grid Challenges in Hungary Péter Stefán Ferenc Szalai Gábor Vitéz NIIF/HUNGARNET.
AoE and HyperSCSI on Linux PDA Prepared by They Yu Shu.
VMware vSphere Configuration and Management v6
 The End to the Means › (According to IBM ) › 03.ibm.com/innovation/us/thesmartercity/in dex_flash.html?cmp=blank&cm=v&csr=chap ter_edu&cr=youtube&ct=usbrv111&cn=agus.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Highest performance parallel storage for HPC environments Garth Gibson CTO & Founder IDC HPC User Forum, I/O and Storage Panel April 21, 2009.
EMC Proven Professional. Copyright © 2012 EMC Corporation. All Rights Reserved. NAS versus SAN NAS – Architecture to provide dedicated file level access.
STORAGE ARCHITECTURE/ MASTER): Disk Storage: What Are Your Options? Randy Kerns Senior Partner The Evaluator Group.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Internet Protocol Storage Area Networks (IP SAN)
STORAGE ARCHITECTURE/ MASTER): Where IP and FC Storage Fit in Your Enterprise Randy Kerns Senior Partner The Evaluator Group.
Mass Storage at SARA Peter Michielse (NCF) Mark van de Sanden, Ron Trompert (SARA) GDB – CERN – January 12, 2005.
© 2007 EMC Corporation. All rights reserved. Internet Protocol Storage Area Networks (IP SAN) Module 3.4.
Storage System Optimization. Introduction Storage Types-DAS/NAS/SAN The purposes of different RAID types. How to calculate the storage size for video.
Copyright © 2010 Hitachi Data Systems. All rights reserved. Confidential – NDA Strictly Required Hitachi Storage Solutions Hitachi HDD Directions HDD Actual.
July 30, 2009opsarea meeting, IETF Stockholm1 Operational Deployment and Management of Storage over the Internet David L. Black, EMC IETF opsarea meeting.
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
Open-E Data Storage Software (DSS V6)
Storage Area Networks The Basics.
Integrating Disk into Backup for Faster Restores
Video Security Design Workshop:
Direct Attached Storage and Introduction to SCSI
Storage Networking.
Direct Attached Storage Overview
Direct Attached Storage and Introduction to SCSI
Storage Networking.
JDAT Production Hardware
Storage Networks and Storage Devices
Cost Effective Network Storage Solutions
Presentation transcript:

Peter Stefan, NIIF 29 June, 2007, Amsterdam, The Netherlands NIIF Storage Services Collaboration on Storage Services

 Objective of our storage development  Brief NIIF introduction  Driving force of storage development: users need it!  Storage building blocks:  FC,  AoE,  „Design and develop your own” box.  The two storage node solutions at NIIF  How can users access storage volumes?  Storage node integration  Conclusions Agenda NIIF –

 NIIF is the Hungarian NREN, connecting about 450 academic institutions and users.  Our service portfolio is as follows:  backbone data network (HBONE),  videoconferencing services,  web, mail, AAI, FTP, DNS services,  supercomputers, grids, and  data storage facilities.  Financed mostly by the Hungarian Ministry of Economics and Transportation. Who we are? NIIF –

Who we are? NIIF –

 There is a continuous need for large volume, moderate speed and high availability storage solutions.  Key usage areas that we know about are:  library archives,  on-line network monitoring (netflow),  videoconferencing,  grids,  off-site back up of large-volume servers. Why large storage volumes? NIIF –

 Fibre Channel (FC) SAN/NAS solutions dominate the market.  Experience with StorageTek (15000 USD/TB), T3, CX500 (6000 USD/TB), SFS (45000 USD/TB).  Generic experience:  FC works well  high price/{capacity,performance} figures   rather large operational costs (disk replacement, support)   not optimal choice for disk-based back ups.  Bulding blocks - FC NIIF –

 Interesting idea: use only Ethernet as storage interconnection!  ATA-over-Ethernet protocol by Coraid Inc.  Key idea: ATA commands are wrapped into Ethernet frames; the AoE driver strips Ethernet frames, then extracts ATA commands and conveys them as block devices. Bulding blocks - AoE NIIF –

 Two conceptually different AoE implementations:  Parallel ATA (older, 5 MB/s),  Serial ATA (newer, 60 MB/s).  Generic experience with the SATA boxes:  moderate performance figures in certain cases   very good price/performance value (1200 USD/TB), and prices keep dropping  low operational costs (standard disks, very simple architecture)  to build a PByte-scale system further virtualization layer is required  Bulding blocks - AoE NIIF –

 Use a multi-disk PC chassis with Linux!  SATA or SAS disks,  standard RAID cards,  PCI-express bus system,  iSCSI and AoE protocols,  universal storage box: GE, Infiniband, 10 GE.  Generic experience with ours:  modular  can be driven up to the hardware boundaries  our record is: 700 MB/s using RAID0 and 450 MB/s using RAID10  no vendor support  „Design and develop your own box” NIIF –

 The purpose is to provide a single, large disk volume for grid users.  Furbrished from 7 PATA boxes, 2 Cisco Catalyst switches, and 2 64-bit servers.  Involves 7x10x400 MBs of disks. Sum: 28 TBs.  Two storage processors are in master-slave layout.  Volumes are organized in ten vertical RAID6 arrays concatenated by LVS. This provides box redundancy.  Available via grid middleware and SCP-only.  Weak points: low IO (50-55 MB/s) , lack of switch-level redundancy , not scalable . Node solution 1 - PATA storage NIIF –

Node solution 1 - PATA storage NIIF –

 The purpose is to provide a single, large disk volume that can be split to smaller, yet scalable units for generic usage.  Made up of 4 SATA boxes, 2 Cisco Catalyst GE switches, and 2 servers.  GE is used as storage interconnection.  Involves 4x15x500 GBs of disks. Sum: 30 TBs.  Two Linux storage processors with HA.  Volumes are organized in 15 vertical RAID5 arrays concatenated by LVS.  RAID + volume management is performed on storage proeccors. Node solution 2 - SATA storage NIIF –

 Full box-, switch-, and storage processor-level redundancy.  Available via iSCSI for local users.  Has virtualization capability, i.e. any boxes and any storage protocols can be used (FC, AoE, iSCSI, Infiniband, Tape).  Can be used for building up hierarchical storage systems.  Not-yet FC-equivalent in terms of e.g. cache coherency .  IO is about MB/s. Node solution 2 - SATA storage NIIF –

Solution #2 - SATA storage NIIF –

Solution #2 - SATA storage NIIF –

 Storage nodes can provide block-level, file-level, and application-level access to the large disk volumes.  Block-level:  AoE,  iSCSI (particularly kind to NRENs).  File-level:  NFS, SMBFS,  Cluster file systems (GFS, Lustre).  Application-level:  SSH, SCP,  storage-capable grid middleware. How to give access to users? NIIF –

 Current status: 2 nodes, lot of plans for the future.  Create multiple redundant storage nodes at different parts of our network, i.e. at regional centers, or at large customers.  Integrate them at application level, by using storage management software, like SRM, or Grid Underground (GUG) storage management modules.  The target system is a geographically distributed storage system providing distributed, replicated, yet safely protected data sets, and standard interfaces.  First inter-node service: distributed and encrypted backup service. Storage node integration NIIF –

 During our storage development activities we revealed that building cost-efficient, reasonable- speed and reliable storage nodes is possible based on cheap building blocks ands free software.  To handle complexity we believe that application- integrated hierarchical storage systems are needed.  NRENs can play integrating role due to their positions.  QoS, monitoring and, in general, strict operational disciplines are necessary compared to, say, grid systems.  There is still a lot to develop: management, advanced monitoring, cache coherency. Conclusions NIIF –

? Thanks & Questions