2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Dealing.

Slides:



Advertisements
Similar presentations
SAN DIEGO SUPERCOMPUTER CENTER Niches, Long Tails, and Condos Effectively Supporting Modest-Scale HPC Users 21st High Performance Computing Symposia (HPC'13)
Advertisements

HPC USER FORUM I/O PANEL April 2009 Roanoke, VA Panel questions: 1 response per question Limit length to 1 slide.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO SDSC RP Update October 21, 2010.
Katie Antypas NERSC User Services Lawrence Berkeley National Lab NUG Meeting 1 February 2012 Best Practices for Reading and Writing Data on HPC Systems.
© 2004 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice InfiniBand Storage: Luster™ Technology.
An Approach to Secure Cloud Computing Architectures By Y. Serge Joseph FAU security Group February 24th, 2011.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO IEEE Symposium of Massive Storage Systems, May 3-5, 2010 Data-Intensive Solutions.
1 Recap (RAID and Storage Architectures). 2 RAID To increase the availability and the performance (bandwidth) of a storage system, instead of a single.
NWfs A ubiquitous, scalable content management system with grid enabled cross site data replication and active storage. R. Scott Studham.
Connecting HPIO Capabilities with Domain Specific Needs Rob Ross MCS Division Argonne National Laboratory
Seafile - Scalable Cloud Storage System
Reliable PVFS. High Performance I/O ? Three Categories of applications demand good I/O performance  Database management systems (DBMSs) Reading or writing.
NPACI: National Partnership for Advanced Computational Infrastructure August 17-21, 1998 NPACI Parallel Computing Institute 1 Cluster Archtectures and.
National Energy Research Scientific Computing Center (NERSC) The GUPFS Project at NERSC GUPFS Team NERSC Center Division, LBNL November 2003.
Lustre at Dell Overview Jeffrey B. Layton, Ph.D. Dell HPC Solutions |
Operating System. Architecture of Computer System Hardware Operating System (OS) Programming Language (e.g. PASCAL) Application Programs (e.g. WORD, EXCEL)
1 - Q Copyright © 2006, Cluster File Systems, Inc. Lustre Networking with OFED Andreas Dilger Principal System Software Engineer
Optimizing Performance of HPC Storage Systems
SAN DIEGO SUPERCOMPUTER CENTER HDF5/SRB Integration August 28, 2006 Mike Wan SRB, SDSC Peter Cao
Small File File Systems USC Jim Pepin. Level Setting  Small files are ‘normal’ for lots of people Metadata substitute (lots of image data are done this.
Ceph Storage in OpenStack Part 2 openstack-ch,
HPCS Lab. High Throughput, Low latency and Reliable Remote File Access Hiroki Ohtsuji and Osamu Tatebe University of Tsukuba, Japan / JST CREST.
Hadoop Hardware Infrastructure considerations ©2013 OpalSoft Big Data.
Workload-driven Analysis of File Systems in Shared Multi-Tier Data-Centers over InfiniBand K. Vaidyanathan P. Balaji H. –W. Jin D.K. Panda Network-Based.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
1 Moshe Shadmon ScaleDB Scaling MySQL in the Cloud.
Scientific Computing Experimental Physics Lattice QCD Sandy Philpott May 20, 2011 IT Internal Review 12GeV Readiness.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
Storage Tank in Data Grid Shin, SangYong(syshin, #6468) IBM Grid Computing August 23, 2003.
OSG Storage Architectures Tuesday Afternoon Brian Bockelman, OSG Staff University of Nebraska-Lincoln.
N. GSU Slide 1 Chapter 05 Clustered Systems for Massive Parallelism N. Xiong Georgia State University.
Large Scale Parallel File System and Cluster Management ICT, CAS.
SAN DIEGO SUPERCOMPUTER CENTER SDSC's Data Oasis Balanced performance and cost-effective Lustre file systems. Lustre User Group 2013 (LUG13) Rick Wagner.
Distributed Object Storage Rebuild Analysis via Simulation with GOBS Justin M Wozniak, Seung Woo Son, and Robert Ross Argonne National Laboratory Presented.
I/O on Clusters Rajeev Thakur Argonne National Laboratory.
Storage and Storage Access 1 Rainer Többicke CERN/IT.
BOINC: Progress and Plans David P. Anderson Space Sciences Lab University of California, Berkeley BOINC:FAST August 2013.
Ultimate Integration Joseph Lappa Pittsburgh Supercomputing Center ESCC/Internet2 Joint Techs Workshop.
Active Storage Processing in Parallel File Systems Jarek Nieplocha Evan Felix Juan Piernas-Canovas SDM CENTER.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
Highest performance parallel storage for HPC environments Garth Gibson CTO & Founder IDC HPC User Forum, I/O and Storage Panel April 21, 2009.
Slide 1 UCSC 100 Gbps Science DMZ – 1 year 9 month Update Brad Smith & Mary Doyle.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
Lecture 1: Network Operating Systems (NOS) An Introduction.
Accelerating High Performance Cluster Computing Through the Reduction of File System Latency David Fellinger Chief Scientist, DDN Storage ©2015 Dartadirect.
ROM AND RAM By Georgia Harris. WHAT DOES IT MEAN?  RAM: random access memory  ROM: read only memory.
SDM Center Parallel I/O Storage Efficient Access Team.
Data Evolution: 101. Parallel Filesystem vs Object Stores Amazon S3 CIFS NFS.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
A Scalable Distributed Datastore for BioImaging R. Cai, J. Curnutt, E. Gomez, G. Kaymaz, T. Kleffel, K. Schubert, J. Tafas {jcurnutt, egomez, keith,
1 TCS Confidential. 2 Objective : In this session we will be able to learn:  What is Cloud Computing?  Characteristics  Cloud Flavors  Cloud Deployment.
Centre de Calcul de l’Institut National de Physique Nucléaire et de Physique des Particules Apache Spark Osman AIDEL.
Unit 2 VIRTUALISATION. Unit 2 - Syllabus Basics of Virtualization Types of Virtualization Implementation Levels of Virtualization Virtualization Structures.
The Evolution of the Italian HPC Infrastructure Carlo Cavazzoni CINECA – Supercomputing Application & Innovation 31 Marzo 2015.
Architecture of a platform for innovation and research Erik Deumens – University of Florida SC15 – Austin – Nov 17, 2015.
Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb
Dr Andrew Peter Hammersley ESRF ESRF MX COMPUTIONAL AND NETWORK CAPABILITIES (Note: This is not my field, so detailed questions will have to be relayed.
Predrag Buncic CERN Data management in Run3. Roles of Tiers in Run 3 Predrag Buncic 2 ALICEALICE ALICE Offline Week, 01/04/2016 Reconstruction Calibration.
GPFS Parallel File System
Canadian Bioinformatics Workshops
Canadian Bioinformatics Workshops
WP18, High-speed data recording Krzysztof Wrona, European XFEL
The demonstration of Lustre in EAST data system
Operating System.
Status and Challenges: January 2017
CS122B: Projects in Databases and Web Applications Winter 2017
Network Operating Systems (NOS)
Recap: introduction to e-science
NSF : CIF21 DIBBs: Middleware and High Performance Analytics Libraries for Scalable Data Science PI: Geoffrey C. Fox Software: MIDAS HPC-ABDS.
Hadoop Technopoints.
Presentation transcript:

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Dealing with Data: Choosing a Good Storage Technology for Your Application Rick Wagner HPC Systems Manager July 1st, 2014

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Application Focus Storage choices should be driven by application need, not just what’s available. But, applications need to adapt as they scale. Writing a few small files to an NFS server is fine… writing 1000’s simultaneously will wipe out the server. If you use binary files, don’t invent your own format. Consider HDF5.

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Storage Technologies File Systems DevicesServices memory block Cloud MySQL CouchDB ext4 NFS Lustre PVFS FUSE

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Storage Technologies File Systems DevicesServices memory block Cloud MySQL CouchDB ext4 NFS Lustre PVFS FUSE Each has its own performance characteristics Not all are available everywhere

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO File Systems Classic access, POSIX, Windows Most relevant: Local Remote NFS, CIFS Parallel (Lustre, GPFS) Local file systems are good for small and temporary files Network file systems very convenient for sharing data between systems

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Parallel File Systems

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Parallel File Systems OSS72TBOSS72TB 32 OSS (Object Storage Servers) Provide 100GB/s Performance and >4PB Raw Capacity Arista G 10G 10G 10G Redundant Switches for Reliability and Performance 3 Distinct Network Architectures OSS72TBOSS72TBOSS72TBOSS72TBOSS72TBOSS72TB 64 Lustre LNET Routers 100 GB/s 64 Lustre LNET Routers 100 GB/s Mellanox 5020 Bridge 12 GB/s Mellanox 5020 Bridge 12 GB/s MDSMDS MDSMDS Myrinet 10G Switch 25 GB/s Myrinet 10G Switch 25 GB/s MDSMDS GORDON IB cluster GORDON TRITON Myrinet cluster TRITON TRESTLES IB cluster TRESTLES Metadata Servers

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO A Cautionary Tale

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Devices Raw block device (/dev/sdb) or RAM FS (/dev/shm) Useful in specific cases, like fast scratch Can be very good for small I/O

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Services Things accessed programmatically Frequents the last thought for HPC applications: A MISTAKE Databases Cloud storage (Amazon S3) Document storage (MongoDB, CouchDB)

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Know What You Need

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Order of Magnitude Guide Storagefile/directoryfile sizesBWIOPs Local HDD1000sGB100 MB/s100 Local SSD1000sGBGB/s10000 RAM FS10000sGBGB/s10000 NFS100sGB100 MB/s100 Lustre/GPFS100sTB100 GB/s1000 CloudInfiniteTB10 GB/s0 DBN/A 10000

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Choosing My application needs to: I should consider: Write a checkpoint dump from memory from a large parallel simulation. A parallel file system and a binary file format like HDF5.

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Choosing My application needs to: I should consider: Run analysis on remote systems and return the results to a web portal for users. Cloud storage for results and input, and local scratch space for the job.

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Choosing My application needs to: I should consider: Randomly access many small files, or read and write small blocks from large files. A database, RAM FS, or local scratch space.

2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Many Boxes Make a Sad Panda Database logos courtesy of RRZEiconsRRZEicons