Kirsten Fagnan NERSC User Services Februar 12, 2013 Getting Started at NERSC.

Slides:



Advertisements
Similar presentations
Computing Infrastructure
Advertisements

1 Computational models of the physical world Cortical bone Trabecular bone.
SAN DIEGO SUPERCOMPUTER CENTER Niches, Long Tails, and Condos Effectively Supporting Modest-Scale HPC Users 21st High Performance Computing Symposia (HPC'13)
GPU System Architecture Alan Gray EPCC The University of Edinburgh.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO SDSC RP Update October 21, 2010.
Katie Antypas NERSC User Services Lawrence Berkeley National Lab NUG Meeting 1 February 2012 Best Practices for Reading and Writing Data on HPC Systems.
HPCC Mid-Morning Break High Performance Computing on a GPU cluster Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
Kjiersten Fagnan JGI/NERSC Consultant
Chapter1 Fundamental of Computer Design Dr. Bernard Chen Ph.D. University of Central Arkansas.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
ORNL is managed by UT-Battelle for the US Department of Energy Tools Available for Transferring Large Data Sets Over the WAN Suzanne Parete-Koon Chris.
Lecture 1: Introduction to High Performance Computing.
N ATIONAL E NERGY R ESEARCH S CIENTIFIC C OMPUTING C ENTER 1 Comparison of Communication and I/O of the Cray T3E and IBM SP Jonathan Carter NERSC User.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
MATE-EC2: A Middleware for Processing Data with Amazon Web Services Tekin Bicer David Chiu* and Gagan Agrawal Department of Compute Science and Engineering.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
HPCC Mid-Morning Break Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery Introduction to the new GPU (GFX) cluster.
Accelerating SQL Database Operations on a GPU with CUDA Peter Bakkum & Kevin Skadron The University of Virginia GPGPU-3 Presentation March 14, 2010.
JGI/NERSC New Hardware Training Kirsten Fagnan, Seung-Jin Sul January 10, 2013.
© 2011 IBM Corporation 11 April 2011 IDS Architecture.
NERCS Users’ Group, Oct. 3, 2005 NUG Training 10/3/2005 Logistics –Morning only coffee and snacks –Additional drinks $0.50 in refrigerator in small kitchen.
Data oriented job submission scheme for the PHENIX user analysis in CCJ Tomoaki Nakamura, Hideto En’yo, Takashi Ichihara, Yasushi Watanabe and Satoshi.
GPU Programming with CUDA – Accelerated Architectures Mike Griffiths
The NERSC Global File System (NGF) Jason Hick Storage Systems Group Lead CAS2K11 September 11-14,
Introduction to HPC resources for BCB 660 Nirav Merchant
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Introduction to the HPCC Jim Leikert System Administrator High Performance Computing Center.
Introduction to the HPCC Dirk Colbry Research Specialist Institute for Cyber Enabled Research.
Katie Antypas User Services Group Lawrence Berkeley National Lab 17 February 2012 JGI Training Series.
PDSF at NERSC Site Report HEPiX April 2010 Jay Srinivasan (w/contributions from I. Sakrejda, C. Whitney, and B. Draney) (Presented by Sandy.
1 Application Scalability and High Productivity Computing Nicholas J Wright John Shalf Harvey Wasserman Advanced Technologies Group NERSC/LBNL.
Htar Hpss Tape Archiver Client API-based interface written by Mike Gleicher Originally commissioned for LLNL in 2000 Now available as part of the HPSS.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
CS 240A Applied Parallel Computing John R. Gilbert Thanks to Kathy Yelick and Jim Demmel at UCB for.
Template This is a template to help, not constrain, you. Modify as appropriate. Move bullet points to additional slides as needed. Don’t cram onto a single.
CENTER FOR HIGH PERFORMANCE COMPUTING Introduction to I/O in the HPC Environment Brian Haymore, Sam Liston,
CS 240A Applied Parallel Computing John R. Gilbert Thanks to Kathy Yelick and Jim Demmel at UCB for.
1 Lattice QCD Clusters Amitoj Singh Fermi National Accelerator Laboratory.
Hepix LAL April 2001 An alternative to ftp : bbftp Gilles Farrache In2p3 Computing Center
Running Parallel Jobs Cray XE6 Workshop February 7, 2011 David Turner NERSC User Services Group.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Parallel IO for Cluster Computing Tran, Van Hoai.
Introduction to Hartree Centre Resources: IBM iDataPlex Cluster and Training Workstations Rob Allan Scientific Computing Department STFC Daresbury Laboratory.
Tackling I/O Issues 1 David Race 16 March 2010.
Parallel Computers Today Oak Ridge / Cray Jaguar > 1.75 PFLOPS Two Nvidia 8800 GPUs > 1 TFLOPS Intel 80- core chip > 1 TFLOPS  TFLOPS = floating.
CCJ introduction RIKEN Nishina Center Kohei Shoji.
Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb
Predrag Buncic CERN Data management in Run3. Roles of Tiers in Run 3 Predrag Buncic 2 ALICEALICE ALICE Offline Week, 01/04/2016 Reconstruction Calibration.
ORNL is managed by UT-Battelle for the US Department of Energy OLCF HPSS Performance Then and Now Jason Hill HPC Operations Storage Team Lead
An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
Advanced Computing Facility Introduction
Compute and Storage For the Farm at Jlab
Getting the Most out of Scientific Computing Resources
A Brief Introduction to NERSC Resources and Allocations
Getting the Most out of Scientific Computing Resources
HPC usage and software packages
Architecture & System Overview
Architecture & Organization 1
Is System X for Me? Cal Ribbens Computer Science Department
Welcome to our Nuclear Physics Computing System
湖南大学-信息科学与工程学院-计算机与科学系
Architecture & Organization 1
Design Unit 26 Design a small or home office network
Welcome to our Nuclear Physics Computing System
High Performance Computing in Bioinformatics
Introduction to High Performance Computing Using Sapelo2 at GACRC
Presentation transcript:

Kirsten Fagnan NERSC User Services Februar 12, 2013 Getting Started at NERSC

Purpose This presentation will help you get familiar with NERSC and its facilities – Practical information – Introduction to terms and acronyms This is not a programming tutorial – But you will learn how to get help and what kind of help is available – We can give presentations on programming languages and parallel libraries – just ask

Outline What is NERSC? Computing Resources Storage Resources How to Get Help 3

What is NERSC? 4

NERSC Leads DOE in Scientific Computing Productivity 5 NERSC computing for science 5500 users, 600 projects From 48 states; 65% from universities Hundreds of users each day 1,500 publications per year Systems designed for science 1.3PF Petaflop Cray system, Hopper -Among Top 20 fastest in world -Fastest open Cray XE6 system -Additional clusters, data storage

We support a diverse workload Many codes (600+) and algorithms Computing at Scale and at High Volume 2012 Job Size Breakdown on Hopper Jan 2012 Dec 2012 Fraction of hours

Our operational priority is providing highly available HPC resources backed by exceptional user support We maintain a very high availability of resources (>90%) – One large HPC system is available at all times to run large-scale simulations and solve high throughput problems Our goal is to maximize the productivity of our users – One-on-one consulting – Training (e.g., webinars) – Extensive use of web pages – We solve or have a path to solve 80% of user tickets within 3 business days - 7 -

Computing Resources 8

Current NERSC Systems 9 Large-Scale Computing Systems Hopper (NERSC-6): Cray XE6 6,384 compute nodes, 153,216 cores 144 Tflop/s on applications; 1.3 Pflop/s peak Edison (NERSC-7): Cray Cascade To be operational in 2013 Over 200 Tflop/s on applications, 2 Pflop/s peak HPSS Archival Storage 240 PB capacity 5 Tape libraries 200 TB disk cache NERSC Global Filesystem (NGF) Uses IBM’s GPFS 8.5 PB capacity 15GB/s of bandwidth Midrange 140 Tflops total Carver IBM iDataplex cluster 9884 cores; 106TF PDSF (HEP/NP) ~1K core cluster GenePool (JGI) ~5K core cluster 2.1 PB Isilon File System Analytics & Testbeds Dirac 48 Fermi GPU nodes

Structure of the Genepool System ssh genepool.nersc.gov User Access Command Line Scheduler Service

Compute Node Hardware CountCoresSlotsScheduleable Memory Memory/ Slot Interconnects G5.25G1Gb Ethernet G7.5G14x FDR Infiniband G10.5G1G Ethernet G15.625G4 have 10G Eth 5 have Infiniband G31.25G1 has 10G Eth 2 have Infiniband G31.25G10G Ethernet The 42G nodes are scheduled “by slot” Multiple jobs can run on the same node at the same time Higher memory nodes are exclusively scheduled Only one job can run at a time on a node

Hopper - Cray XE6 1.2 GB memory / core (2.5 GB / core on "fat" nodes) for applications /scratch disk quota of 5 TB 2 PB of /scratch disk Choice of full Linux operating system or optimized Linux OS (Cray Linux) PGI, Cray, Pathscale, GNU compilers ,408 cores, 6,392 nodes "Gemini” interconnect 2 12-core AMD 'MagnyCours' 2.1 GHz processors per node 24 processor cores per node 32 GB of memory per node (384 "fat" nodes with 64 GB) 216 TB of aggregate memory Use Hopper for your biggest, most computationally challenging problems.

Edison - Cray XC30 (Phase 1 / 2) 4 / TBA GB memory / core for applications 1.6 / 6.4 PB of /scratch disk CCM compatibility mode available Intel, Cray, GNU compilers 13 10K / 105K compute cores ”Aires” interconnect 2 8-core Intel ’Sandy Bridge' 2.6 GHz processors per node (Phase 2 TBA) 16 / TBA processor cores per node 64 GB of memory per node 42 / 333 TB of aggregate memory Edison Phase 1 access Feb. or March 2013 / Phase 2 in Summer or Fall 2013 Use Edison for your most computationally challenging problems.

Carver - IBM iDataPlex 3,200 compute cores 400 compute nodes 2 quad-core Intel Nehalem 2.67 GHz processors per node 8 processor cores per node 24 GB of memory per node (48 GB on 80 "fat" nodes) 2.5 GB / core for applications (5.5 GB / core on "fat" nodes) InfiniBand 4X QDR 14 NERSC global /scratch directory quota of 20 TB Full Linux operating system PGI, GNU, Intel compilers Use Carver for jobs that use up to 512 cores, need a fast CPU, need a standard Linux configuration, or need up to 48 GB of memory on a node.

Dirac – GPU Computing Testbed 50 GPUs 50 compute nodes 2 quad-core Intel Nehalem 2.67 GHz processors 24 GB DRAM memory 44 nodes: 1 NVIDIA Tesla C2050 (Fermi) GPU with 3GB of memory and 448 cores 1 node: 4 NVIDIA Tesla C2050 (Fermi) GPU's, each with 3GB of memory and 448 processor cores. InfiniBand 4X QDR 15 CUDA 5.0, OpenCL, PGI and HMPP directives DDT CUDA-enabled debugger PGI, GNU, Intel compilers Use Dirac for developing and testing GPU codes.

How to Get Help 16

NERSC Services NERSC’s emphasis is on enabling scientific discovery User-oriented systems and services – We think this is what sets NERSC apart from other centers Help Desk / Consulting – Immediate direct access to consulting staff that includes 7 Ph.Ds User group (NUG) has tremendous influence – Monthly teleconferences & yearly meetings Requirement-gathering workshops with top scientists – One each for the six DOE Program Offices in the Office of Science – Ask, and we’ll do whatever we can to fulfill your request

Your JGI Consultants 18 Seung-Jin Sul, Ph.D. Bioinformatics Doug Jacobsen, Ph.D. Bioinformatics Kirsten Fagnan, Ph.D. Applied Math One of us is onsite every day! Stop by if you have questions!

How to Get Help 19

Passwords and Login Failures Passwords You may call NERSC 24/7 to reset your password Change it at Answer security questions in NIM, then you can reset it yourself at 20 Login Failures 3 or more consecutive login failures on a machine will disable your ability to log in Send to or call NERSC to reset your failure count

Data Resources 21

Data Storage Types “Spinning Disk” – Interactive access – I/O from compute jobs – “Home”, “Project”, “Scratch”, “Projectb”, “Global Scratch” – No on-node direct-attach disk at NERSC (except PDSF,Genepool) Archival Storage – Permanent, long-term storage – Tapes, fronted by disk cache – “HPSS” (High Performance Storage System) 22

Home Directory When you log in you are in your "Home" directory. Permanent storage – No automatic backups The full UNIX pathname is stored in the environment variable $HOME – hopper04% echo $HOME – /global/homes/r/ragerber $HOME is a global file system – You see all the same directories and files when you log in to any NERSC computer. Your quota in $HOME is 40 GB and 1M inodes (files and directories). – Use “myquota” command to check your usage and quota 23 genepool04% echo $HOME /global/homes/k/kmfagnan

Scratch Directories “Scratch” file systems are large, high-performance ”scratch” file systems. Significant I/O from your compute jobs should be directed to $SCRATCH Each user has a personal directory referenced by $SCRATCH – on Genepool this points to /global/projectb/scratch/ Data in $SCRATCH is purged (12 weeks from last access) Always save data you want to keep to HPSS (see below) $SCRATCH is local on Edison and Hopper, but Carver and other systems use a global scratch file system. ($GSCRATCH points to global scratch on Hopper and Edison, as well as Genepool) Data in $SCRATCH is not backed up and could be lost if a file system fails. 24

Project Directories All NERSC systems mount the NERSC global "Project” file systems. Projectb is specific to the JGI, but is also accessible on Hopper and Carver. "Project directories” are created upon request for projects (groups of researchers) to store and share data. The default quota in /projectb is 5 TB. Data in /projectb/projectdirs is not purged and only 5TB is backed up. This may change in the future, but for long term storage, you should use the archive. 25

File System Access Summary 26

File System Summary

IO Tips Use $SCRATCH for good IO performance Write large chunks of data (MBs or more) at a time Use a parallel IO library (e.g. HDF5) Read/write to as few files as practical from your code (try to avoid 1 file per MPI task) Use $HOME to compile unless you have too many source files or intermediate (*.o) files Do not put more than a few 1,000s of files in a single directory Save any and everything important to HPSS

Archival Storage (HPSS) For permanent, archival storage Permanent storage is magnetic tape, disk cache is transient – 24PB data in >100M files written to 32k cartridges – Cartridges are loaded/unloaded into tape drives by sophisticated library robotics Front-ending the tape subsystem is 150TB fast-access disk 29 Hostname: archive.nersc.gov Over 24 Petabyes of data stored Data increasing by 1.7X per year 120 M files stored 150 TB disk cache 8 STK robots 44,000 tape slots 44 PB maximum capacity today Average data xfer rate: 100 MB/sec

Authentication NERSC storage uses a token-based authentication method – User places encrypted authentication token in ~/.netrc file at the top level of the home directory on the compute platform Authentication tokens can be generated in 2 ways: – Automatic – NERSC auth service: Log into any NERSC compute platform; Type “hsi”; Enter NERSC password – Manual – websitehttps://nim.nersc.gov/ Under “Actions” dropdown, select “Generate HPSS Token” ; Copy/paste content into ~/.netrc; chmod 600 ~/.netrc Tokens are username and IP specific—must use NIM to generate a different token for use offsite

HPSS Clients Parallel, threaded, high performance: – HSI Unix shell-like interface – HTAR Like Unix tar, for aggregation of small files – PFTP Parallel FTP Non-parallel: – FTP Ubiquitous, many free scripting utilities GridFTP interface (garchive) – Connect to other grid-enabled storage systems

Archive Technologies, Continued… HPSS clients can emulate file system qualities – FTP-like interfaces can be deceiving: the archive is backed by tape, robotics, and a single SQL database instance for metadata – Operations that would be slow on a file system, e.g. lots of random IO, can be impractical on the archive – It’s important to know how to store and retrieve data efficiently. (See events/data-transfer-and-archiving/) HPSS does not stop you from making mistakes – It is possible to store data in such a way as to make it difficult to retrieve – The archive has no batch system. Inefficient use affects others.

Avoid Common Mistakes Don’t store many small files – Make a tar archive first, or use htar Don’t use recursively store or retrieve large directory trees Don’t stream data via UNIX pipes – HPSS can’t optimize transfers of unknown size Don’t pre-stage data to disk cache – May evict efficiently stored existing cache data Avoid directories with many files – Stresses HPSS database Long-running transfers – Can be error-prone – Keep to under 24 hours Use as few concurrent sessions as required – Limit of 15 in place

Hands-On Use htar to save a directory tree to HPSS %cd $SCRATCH (or where you put the NewUser directory) %htar –cvf NewUser.tar NewUser See if the command worked %hsi ls –l NewUser.tar -rw-r ragerber ccc Sep 12 16:46 NewUser.tar Retreive the files %cd $SCRATCH2 %htar –xvf NewUser.tar Retreive the tar file %hsi get NewUser.tar

Data Transfer and Archiving A NERSC Training Event Good information: 35

Aside: Why Do You Care About Parallelism? To learn more consider joining the HPC study group that meets on Friday from

Moore’s Law 38 2X transistors/Chip Every 1.5 years Called “Moore’s Law” Moore’s Law Microprocessors have become smaller, denser, and more powerful. Gordon Moore (co-founder of Intel) predicted in 1965 that the transistor density of semiconductor chips would double roughly every 18 months. Slide source: Jack Dongarra

Power Density Limits Serial Performance 39 High performance serial processors waste power -Speculation, dynamic dependence checking, etc. burn power -Implicit parallelism discovery More transistors, but not faster serial processors Concurrent systems are more power efficient –Dynamic power is proportional to V 2 fC –Increasing frequency (f) also increases supply voltage (V)  cubic effect –Increasing cores increases capacitance (C) but only linearly –Save power by lowering clock speed

Revolution in Processors 40 Chip density is continuing increase ~2x every 2 years Clock speed is not Number of processor cores may double instead Power is under control, no longer growing

Moore’s Law reinterpreted 41 Number of cores per chip will double every two years Clock speed will not increase (possibly decrease) Need to deal with systems with millions of concurrent threads Need to deal with inter-chip parallelism as well as intra-chip parallelism Your take-away: Future performance increases in computing are going to come from exploiting parallelism in applications