General Parallel File System

Slides:



Advertisements
Similar presentations
Tivoli SANergy. SANs are Powerful, but... Most SANs today offer limited value One system, multiple storage devices Multiple systems, isolated zones of.
Advertisements

Case Study: Photo.net March 20, What is photo.net? An online learning community for amateur and professional photographers 90,000 registered users.
Andrew Hanushevsky7-Feb Andrew Hanushevsky Stanford Linear Accelerator Center Produced under contract DE-AC03-76SF00515 between Stanford University.
Introduction to cloud computing Jiaheng Lu Department of Computer Science Renmin University of China
Ddn.com ©2012 DataDirect Networks. All Rights Reserved. GridScaler™ Overview Vic Cornell Application Support Consultant.
NAS vs. SAN 10/2010 Palestinian Land Authority IT Department By Nahreen Ameen 1.
Introduction to DBA.
The Zebra Striped Network Filesystem. Approach Increase throughput, reliability by striping file data across multiple servers Data from each client is.
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
Parallel I/O A. Patra MAE 609/CE What is Parallel I/O ? zParallel processes need parallel input/output zIdeal: Processor consuming/producing data.
Study of Hurricane and Tornado Operating Systems By Shubhanan Bakre.
Chapter 13 (Web): Distributed Databases
1 Principles of Reliable Distributed Systems Tutorial 12: Frangipani Spring 2009 Alex Shraer.
Lecture 6 – Google File System (GFS) CSE 490h – Introduction to Distributed Computing, Winter 2008 Except as otherwise noted, the content of this presentation.
City University London
IBM Research Lab in Haifa Architectural and Design Issues in the General Parallel File System Benny Mandler - May 12, 2002.
Comparison and Performance Evaluation of SAN File System Yubing Wang & Qun Cai.
PRASHANTHI NARAYAN NETTEM.
Module – 7 network-attached storage (NAS)
Session 3 Windows Platform Dina Alkhoudari. Learning Objectives Understanding Server Storage Technologies Direct Attached Storage DAS Network-Attached.
RAID-x: A New Distributed Disk Array for I/O-Centric Cluster Computing Kai Hwang, Hai Jin, and Roy Ho.
Frangipani: A Scalable Distributed File System C. A. Thekkath, T. Mann, and E. K. Lee Systems Research Center Digital Equipment Corporation.
Copyright © 2012 Cleversafe, Inc. All rights reserved. 1 Combining the Power of Hadoop with Object-Based Dispersed Storage.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
File Systems and N/W attached storage (NAS) | VTU NOTES | QUESTION PAPERS | NEWS | VTU RESULTS | FORUM | BOOKSPAR ANDROID APP.
Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung Google∗
Windows ® Powered NAS. Agenda Windows Powered NAS Windows Powered NAS Key Technologies in Windows Powered NAS Key Technologies in Windows Powered NAS.
© 2011 IBM Corporation 11 April 2011 IDS Architecture.
Object-based Storage Long Liu Outline Why do we need object based storage? What is object based storage? How to take advantage of it? What's.
SANPoint Foundation Suite HA Robert Soderbery Sr. Director, Product Management VERITAS Software Corporation.
1 The Google File System Reporter: You-Wei Zhang.
Page 1 of 30 NFS Industry Conference October 22-23, 2002 NFSNFS INDUSTRYINDUSTRY CONFERENCECONFERENCE Spinnaker Networks, Inc
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
Pooja Shetty Usha B Gowda.  Network File Systems (NFS)  Drawbacks of NFS  Parallel Virtual File Systems (PVFS)  PVFS components  PVFS application.
INSTALLING MICROSOFT EXCHANGE SERVER 2003 CLUSTERS AND FRONT-END AND BACK ‑ END SERVERS Chapter 4.
July 2003 Sorrento: A Self-Organizing Distributed File System on Large-scale Clusters Hong Tang, Aziz Gulbeden and Tao Yang Department of Computer Science,
IMPROUVEMENT OF COMPUTER NETWORKS SECURITY BY USING FAULT TOLERANT CLUSTERS Prof. S ERB AUREL Ph. D. Prof. PATRICIU VICTOR-VALERIU Ph. D. Military Technical.
Latest Relevant Techniques and Applications for Distributed File Systems Ela Sharda
High Performance Storage System Harry Hulen
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
Chapter 20 Distributed File Systems Copyright © 2008.
Storage Tank in Data Grid Shin, SangYong(syshin, #6468) IBM Grid Computing August 23, 2003.
MapReduce and GFS. Introduction r To understand Google’s file system let us look at the sort of processing that needs to be done r We will look at MapReduce.
Presenters: Rezan Amiri Sahar Delroshan
Fast Crash Recovery in RAMCloud. Motivation The role of DRAM has been increasing – Facebook used 150TB of DRAM For 200TB of disk storage However, there.
GFS. Google r Servers are a mix of commodity machines and machines specifically designed for Google m Not necessarily the fastest m Purchases are based.
VMware vSphere Configuration and Management v6
Problem-solving on large-scale clusters: theory and applications Lecture 4: GFS & Course Wrap-up.
Infrastructure for Data Warehouses. Basics Of Data Access Data Store Machine Memory Buffer Memory Cache Data Store Buffer Bus Structure.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
 Introduction  Architecture NameNode, DataNodes, HDFS Client, CheckpointNode, BackupNode, Snapshots  File I/O Operations and Replica Management File.
Parallel IO for Cluster Computing Tran, Van Hoai.
GPFS: A Shared-Disk File System for Large Computing Clusters Frank Schmuck & Roger Haskin IBM Almaden Research Center.
Tackling I/O Issues 1 David Race 16 March 2010.
1 CEG 2400 Fall 2012 Network Servers. 2 Network Servers Critical Network servers – Contain redundant components Power supplies Fans Memory CPU Hard Drives.
Unit 1: IBM Tivoli Storage Manager 5.1 Overview. 2 Objectives Upon the completion of this unit, you will be able to: Identify the purpose of IBM Tivoli.
IT 5433 LM1. Learning Objectives Understand key terms in database Explain file processing systems List parts of a database environment Explain types of.
Seminar On Rain Technology
An Introduction to GPFS
Cofax Scalability Document Version Scaling Cofax in General The scalability of Cofax is directly related to the system software, hardware and network.
GPFS Parallel File System
Unit OS10: Fault Tolerance
Gregory Kesden, CSE-291 (Cloud Computing) Fall 2016
Storage Virtualization
The Google File System Sanjay Ghemawat, Howard Gobioff and Shun-Tak Leung Google Presented by Jiamin Huang EECS 582 – W16.
Chapter 17: Database System Architectures
Hadoop Technopoints.
CLUSTER COMPUTING.
Database System Architectures
Microsoft Virtual Academy
Presentation transcript:

General Parallel File System Presentation by: Lokesh Pradhan

Introduction File System Way to organize data which is expected to be retained after the program terminates by providing procedures to store, retrieve and update data as well as manage the available space on the device which contains it. As we are juniors and senior standing in comp science, we all know what is File System . However, I want repeat what it means Way to organize data which is expected to be retained after the program terminates by providing procedures to store, retrieve and update data as well as manage the available space on the device which contains it

Types of File System Types Examples Disk file system FAT, exFAT, NTFS… Optical discs CD, DVD, Blu-ray Tape file system IBM’s Linear tape Database file system DB2 Transactional file system TxF, Valor, Amino, TFFS Flat file system Amazon’s S3 Cluster file system Distributed file system Shared file system San file system Parallel file system NFS, CIFS, AFS, SMB, GFS, GPFS, LUSTRE, PAS Here the list of types. They can be classified into many types, but here are only few you might know or will from now from this slide.

In HPC world Equally large applications Large input data set (e.g. astronomy data) Parallel execution on large clusters Use parallel file systems for scalable I/O e.g. IBM’s GPFS, Sun’s Lustre FS, PanFS, and Parallel Virtual File System (PVFS) High performance computing is another style of computing at a comparable scale which process large input data set in parallel In HPC world, parallel file systems are used for highly scalable storage I/O Examples of parallel file systems are GPFS, Lustre, PanFS and PVFS ??? What do you think we can propose?

General Parallel File System Cluster: 512 nodes today, fast reliable communication Shared disk: all data and metadata on disk accessible from any node through disk I/O interface (i.e., "any to any" connectivity) Parallel: data and metadata flows from all of the nodes to all of the disks in parallel RAS: reliability, accessibility, serviceability The General Parallel File System (GPFS) is a high-performance shared disk clustered file system developed by IBM. As it is based on the characteristics of shared disk file system, it has provides concurrent high speed file access to the applications executing on multiple nodes of cluster. It can be used with AIX 5L clusters, Linux clusters, on Microsoft Windows Server, or a heterogeneous cluster of AIX, Linux and Windows nodes. Moreover, GPFS offers tools to manage and administrate the GPFS cluster and hence, allows the shared access to file systems from remote GPFS clusters. there are the characteristics of GPFS Cluster: 512 nodes today, fast reliable communication, common admin domain Shared disk: all data and metadata on disk accessible from any node through disk I/O interface (i.e., "any to any" connectivity) Parallel: data and metadata flows from all of the nodes to all of the disks in parallel RAS: reliability, accessibility, serviceability GPFS is the filesystem of the ASC Purple Supercomputer[2] which is composed of more than 12,000 processors and has 2 petabytes of total disk storage spanning more than 11,000 disks.

History of GPFS Shark video server Video streaming from single RS/6000 Complete system, included file system, network driver, control server Large data blocks, admission control, deadline scheduling Bell Atlantic video-on-demand trial (1993-94) Tiger Shark multimedia file system Multimedia file system for RS/6000 SP Data striped across multiple disks, accessible from all nodes Hong Kong and Tokyo video trials, Austin video server products GPFS parallel file system General purpose file system for commercial and technical computing on RS/6000 SP, AIX and Linux clusters. Recovery, online system management, byte-range locking, fast pre- fetch, parallel allocation, scalable directory, small-block random access. Released as a product 1.1 - 05/98. History Shark video server Video streaming from single RS/6000 Complete system, included file system, network driver, control server Large data blocks, admission control, deadline scheduling Bell Atlantic video-on-demand trial (1993-94) Tiger Shark multimedia file system Multimedia file system for RS/6000 SP Data striped across multiple disks, accessible from all nodes Hong Kong and Tokyo video trials, Austin video server products GPFS parallel file system General purpose file system for commercial and technical computing on RS/6000 SP, AIX and Linux clusters. Recovery, online system management, byte-range locking, fast prefetch, parallel allocation, scalable directory, small-block random access, ... Released as a product 1.1 - 05/98, 1.2 - 12/98, 1.3 - 04/00, …

What is Parallel I/O? Multiple processes (possibly on multiple nodes) participate in the I/O Application level parallelism “File” is stored on multiple disks on a parallel file system

What does Parallel System support? A parallel file system must support Parallel I/O Consistent global name space across all nodes of the cluster Including maintaining a consistent view across all nodes for the same file Programming model allowing programs to access file data Distributed over multiple nodes From multiple tasks running on multiple nodes Physical distribution of data across disks and network entities eliminates bottlenecks both at the disk interface and the network, providing more effective bandwidth to the I/O resources

Why use general parallel file systems? Native AIX File System No file sharing - application can only access files on its own node Applications must do their own data partitioning Distributed File System Application nodes (DCE clients) share files on server node Switch is used as a fast LAN Coarse-grained (file or segment level) parallelism Server node : performance and capacity bottleneck GPFS Parallel File System GPFS file systems are striped across multiple disks on multiple storage nodes Independent GPFS instances run on each application node GPFS instances use storage nodes as "block servers" - all instances can access all disks Native AIX File System (JFS) No file sharing - application can only access files on its own node Applications must do their own data partitioning DCE Distributed File System Application nodes (DCE clients) share files on server node Switch is used as a fast LAN Coarse-grained (file or segment level) parallelism Server node is performance and capacity bottleneck GPFS Parallel File System GPFS file systems are striped across multiple disks on multiple storage nodes Independent GPFS instances run on each application node GPFS instances use storage nodes as "block servers" - all instances can access all disks Parallel file systems offer numerous advantages and address some key issues by providing: Concurrent access to files by multiple nodes of a cluster. This prevents users from having to utilize the local disk of each node and then reassemble the output into either a coherent single file or a collection of multiple files (sometimes referred to as post-mortem reassembly). Scalable performance. Parallel file systems are designed with scalability in mind. As clusters grow, more disk and more network connections need to be incorporated into the fabric of a file system. A single disk space where serial files and files created by parallel applications can coexist and be manipulated.

Performance advantages with GPFS file system Allowing multiple processes or applications on all nodes in the cluster simultaneously Access to the same file using standard file system calls. Increasing aggregate bandwidth of your file system by spreading reads and writes across multiple disks. Balancing the load evenly across all disks to maximize their combined throughput. One disk is no more active than another. Performance advantages with GPFS file system Using GPFS to store and retrieve your files can improve system performance by: Allowing multiple processes or applications on all nodes in the cluster simultaneous access to the same file using standard file system calls. Increasing aggregate bandwidth of your file system by spreading reads and writes across multiple disks. Balancing the load evenly across all disks to maximize their combined throughput. One disk is no more active than another. Supporting very large file and file system sizes. Allowing concurrent reads and writes from multiple nodes. Allowing for distributed token (lock) management. Distributing token management reduces system delays associated with a lockable object waiting to obtaining a token. Allowing for the specification of other networks for GPFS daemon communication and for GPFS administration command usage within your cluster.

Performance advantages with GPFS file system (cont.) Supporting very large file and file system sizes. Allowing concurrent reads and writes from multiple nodes. Allowing for distributed token (lock) management. Distributing token management reduces system delays associated with a lockable object waiting to obtaining a token. Allowing for the specification of other networks for GPFS daemon communication and for GPFS administration command usage within your cluster.

GPFS Architecture Overview Implications of Shared Disk Model All data and metadata on globally accessible disks (VSD) All access to permanent data through disk I/O interface Distributed protocols, e.g., distributed locking, coordinate disk access from multiple nodes Fine-grained locking allows parallel access by multiple clients Logging and Shadowing restore consistency after node failures Implications of Shared Disk Model All data and metadata on globally accessible disks (VSD) All access to permanent data through disk I/O interface Distributed protocols, e.g., distributed locking, coordinate disk access from multiple nodes Fine-grained locking allows parallel access by multiple clients Logging and Shadowing restore consistency after node failures

GPFS Architecture Overview (cont.) Implications of Large Scale Support up to 4096 disks of up to 1 TB each (4 Petabytes) The largest system in production is 75 TB Failure detection and recovery protocols to handle node failures Replication and/or RAID protect against disk / storage node failure On-line dynamic reconfiguration (add, delete, replace disks and nodes; rebalance file system) Implications of Large Scale Support up to 4096 disks of up to 1 TB each (4 Petabytes) The largest system in production is 75 TB Failure detection and recovery protocols to handle node failures Replication and/or RAID protect against disk / storage node failure On-line dynamic reconfiguration (add, delete, replace disks and nodes; rebalance file system)

GPFS Architecture - Special Node Roles Three types of nodes: File system nodes Manager nodes Storage nodes Three types of nodes: file system, storage, and manager File system nodes Run user programs, read/write data to/from storage nodes Implement virtual file system interface Cooperate with manager nodes to perform metadata operations Manager nodes Global lock manager File system configuration: recovery, adding disks, … Disk space allocation manager Quota manager File metadata manager - maintains file metadata integrity Storage nodes Implement block I/O interface Shared access to file system and manager nodes Interact with manager nodes for recovery (e.g. fencing) Data and metadata striped across multiple disks - multiple storage nodes

Disk Data Structures: Large block size allows efficient use of disk bandwidth Fragments reduce space overhead for small files No designated "mirror", no fixed placement function: Flexible replication (e.g., replicate only metadata, or only important files) Dynamic reconfiguration: data can migrate block-by-block Multi level indirect blocks Each disk address: List of pointers to replicas Each pointer: Disk id + sector no.

Availability and Reliability Eliminate single point of failures Designed to transparently fail over token (lock) operations. Supports data replications to increase availability in the vent of a storage media failure. Offers time-tested reliability and has been installed on thousands of nodes across industries Basis of many cloud storage offerings For optimal reliability, GPFS can be configured to eliminate single points of failure. The file system can be configured to remain available automatically in the event of a disk or server failure. A GPFS file is designed to transparently fail over token (lock) operations and other GPFS cluster services, which can be distributed throughout the entire cluster to eliminate the need for dedicated metadata servers. GPFS can be configured to automatically recover from node, storage and other infrastructure failures. GPFS provides this functionality by supporting data replication to increase availability in the event of a storage media failure; multiple paths to the data in the event of a communications or server failure; and file system activity logging, enabling consistent fast recovery after system failures. In addition, GPFS supports snapshots to provide a space-efficient image of a file system at a specified time, which allows online backup and can help protect against user error. GPFS offers time-tested reliability and has been installed on thousands of nodes across industries, from weather research to broadcasting, retail, financial industry analytics and web service providers. GPFS also is the basis of many cloud storage offerings.

GPFS’s Achievement Used on six of the ten most powerful supercomputers in the world, including the largest (ASCI white) Installed at several hundred customer sites, on clusters ranging from a few nodes with less than a TB of disk, up to 512 nodes with 140 TB of disk in 2 file systems 20 filed patents ASC Purple Supercomputer which is composed of more than 12,000 processors and has 2 PB of total disk storage spanning more than 11,000 disks.

Conclusion Efficient for managing data volumes Provides world-class performance, scalability and availability for your file data Designed to optimize the use of storage Provide highly available platform for data-intensive applications Delivering real business needs by streamline data workflows, improvised services reducing cost and managing the risks. GPFS already is managing data volumes that most companies will not need to support for five years or more. You may not need a multipetabyte file system today, but with GPFS you know you will have room to expand as your data volume increases.

References "File System." Wikipedia, the Free Encyclopedia. Web. 20 Jan. 2012. <http://en.wikipedia.org/wiki/File_system>. "IBM General Parallel File System for AIX: Administration and Programming Reference - Contents." IBM General Parallel File System for AIX. IBM. Web. 20 Jan. 2012. <https://support.iap.ac.cn/hpc/ibm/ibm/gpfs/am3admst02.html>. "IBM General Parallel File System." Wikipedia, the Free Encyclopedia. Web. 20 Jan. 2012. <http://en.wikipedia.org/wiki/IBM_General_Parallel_File_System>. Intelligent Storage Management with IBM General Parallel File System. Issue brief. IBM, July 2010. Web. 21 Jan. 2012. <http://www-03.ibm.com/systems/software/gpfs/>. Mandler, Benny. Architectural and Design Issues in the General Parallel File System. IBM Haita Research Lab, May 2005. Web. 21 Jan. 2012. <Architectural and Design Issues in the General Parallel File System>. "NCSA Parallel File Systems." National Center for Supercomputing Applications at the University of Illinois. University of Illinois, 20 Mar. 2008. Web. 21 Jan. 2012. <http://www.ncsa.illinois.edu/UserInfo/Data/filesystems/>. Parallel File System. Rep. Dell Inc., May 2005. Web. 21 Jan. 2012. <www.dell.com/powersolutions>. Welch, Brent. "What Is a Cluster Filesystem?" Brent B Welch. Web. 21 Jan. 2012. <http://www.beedub.com/clusterfs.html>. Here are the references.

Questions?