NERCS Users’ Group, Oct. 3, 2005 NUG Training 10/3/2005 Logistics –Morning only coffee and snacks –Additional drinks $0.50 in refrigerator in small kitchen.

Slides:



Advertisements
Similar presentations
The Development of Mellanox - NVIDIA GPUDirect over InfiniBand A New Model for GPU to GPU Communications Gilad Shainer.
Advertisements

NERCS Users’ Group, Oct. 3, 2005 Interconnect and MPI Bill Saphir.
The Jacquard Programming Environment Mike Stewart NUG User Training, 10/3/05.
♦ Commodity processor with commodity inter- processor connection Clusters Pentium, Itanium, Opteron, Alpha GigE, Infiniband, Myrinet, Quadrics, SCI NEC.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
HPCC Mid-Morning Break High Performance Computing on a GPU cluster Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
TTU High Performance Computing User Training: Part 2 Srirangam Addepalli and David Chaffin, Ph.D. Advanced Session: Outline Cluster Architecture File System.
Running Jobs on Jacquard An overview of interactive and batch computing, with comparsions to Seaborg David Turner NUG Meeting 3 Oct 2005.
Information Technology Center Introduction to High Performance Computing at KFUPM.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
Job Submission on WestGrid Feb on Access Grid.
IBM RS6000/SP Overview Advanced IBM Unix computers series Multiple different configurations Available from entry level to high-end machines. POWER (1,2,3,4)
IBM RS/6000 SP POWER3 SMP Jari Jokinen Pekka Laurila.
ORNL is managed by UT-Battelle for the US Department of Energy Tools Available for Transferring Large Data Sets Over the WAN Suzanne Parete-Koon Chris.
N ATIONAL E NERGY R ESEARCH S CIENTIFIC C OMPUTING C ENTER 1 Comparison of Communication and I/O of the Cray T3E and IBM SP Jonathan Carter NERSC User.
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
Real Parallel Computers. Modular data centers Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra,
HPCC Mid-Morning Break Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery Introduction to the new GPU (GFX) cluster.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
TPB Models Development Status Report Presentation to the Travel Forecasting Subcommittee Ron Milone National Capital Region Transportation Planning Board.
High Performance Computing G Burton – ICG – Oct12 – v1.1 1.
Tools and Utilities for parallel and serial codes in ENEA-GRID environment CRESCO Project: Salvatore Raia SubProject I.2 C.R. ENEA-Portici. 11/12/2007.
Introduction to HPC resources for BCB 660 Nirav Merchant
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Introduction to the HPCC Jim Leikert System Administrator High Performance Computing Center.
Seaborg Cerise Wuthrich CMPS Seaborg  Manufactured by IBM  Distributed Memory Parallel Supercomputer  Based on IBM’s SP RS/6000 Architecture.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Bigben Pittsburgh Supercomputing Center J. Ray Scott
Introduction to the HPCC Dirk Colbry Research Specialist Institute for Cyber Enabled Research.
Cluster Workstations. Recently the distinction between parallel and distributed computers has become blurred with the advent of the network of workstations.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
Katie Antypas User Services Group Lawrence Berkeley National Lab 17 February 2012 JGI Training Series.
Parallel Programming on the SGI Origin2000 With thanks to Igor Zacharov / Benoit Marchand, SGI Taub Computer Center Technion Moshe Goldberg,
Jacquard: Architecture and Application Performance Overview NERSC Users’ Group October 2005.
ARGONNE NATIONAL LABORATORY Climate Modeling on the Jazz Linux Cluster at ANL John Taylor Mathematics and Computer Science & Environmental Research Divisions.
ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed.
Software Overview Environment, libraries, debuggers, programming tools and applications Jonathan Carter NUG Training 3 Oct 2005.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
1 The NERSC Global File System NERSC June 12th, 2006.
1 Lattice QCD Clusters Amitoj Singh Fermi National Accelerator Laboratory.
Cluster Software Overview
1 THE EARTH SIMULATOR SYSTEM By: Shinichi HABATA, Mitsuo YOKOKAWA, Shigemune KITAWAKI Presented by: Anisha Thonour.
The LBNL Perceus Cluster Infrastructure Next Generation Cluster Provisioning and Management October 10, 2007 Internet2 Fall Conference Gary Jung, SCS Project.
1 Cluster Development at Fermilab Don Holmgren All-Hands Meeting Jefferson Lab June 1-2, 2005.
11 January 2005 High Performance Computing at NCAR Tom Bettge Deputy Director Scientific Computing Division National Center for Atmospheric Research Boulder,
Running Parallel Jobs Cray XE6 Workshop February 7, 2011 David Turner NERSC User Services Group.
2011/08/23 國家高速網路與計算中心 Advanced Large-scale Parallel Supercluster.
Proper use of CC at Lyon J. Brunner. Batch farm / bqs BQS classes Only for test jobs !! Regular Exceptional resource needs Very long jobs Large memory/scratch.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Introduction to Hartree Centre Resources: IBM iDataPlex Cluster and Training Workstations Rob Allan Scientific Computing Department STFC Daresbury Laboratory.
Advanced topics Cluster Training Center for Simulation and Modeling September 4, 2015.
Getting Started: XSEDE Comet Shahzeb Siddiqui - Software Systems Engineer Office: 222A Computer Building Institute of CyberScience May.
Cliff Addison University of Liverpool NW-GRID Training Event 26 th January 2007 SCore MPI Taking full advantage of GigE.
Intro to Distributed Systems Hank Levy. 23/20/2016 Distributed Systems Nearly all systems today are distributed in some way, e.g.: –they use –they.
Wouter Verkerke, NIKHEF 1 Using ‘stoomboot’ for NIKHEF-ATLAS batch computing What is ‘stoomboot’ – Hardware –16 machines, each 2x quad-core Pentium = 128.
BLUE GENE Sunitha M. Jenarius. What is Blue Gene A massively parallel supercomputer using tens of thousands of embedded PowerPC processors supporting.
Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb
Plans for the National NERC HPC services UM vn 6.1 installations and performance UM vn 6.6 and NEMO(?) plans.
Advanced Computing Facility Introduction
Compute and Storage For the Farm at Jlab
HPC usage and software packages
CRESCO Project: Salvatore Raia
Is System X for Me? Cal Ribbens Computer Science Department
NERSC Reliability Data
Introduction to High Performance Computing Using Sapelo2 at GACRC
Working in The IITJ HPC System
Presentation transcript:

NERCS Users’ Group, Oct. 3, 2005 NUG Training 10/3/2005 Logistics –Morning only coffee and snacks –Additional drinks $0.50 in refrigerator in small kitchen area; can easily go out to get coffee during 15-minute breaks –Parking garage vouchers at reception desk on second floor Lunch –On your own, but can go out in groups

NERCS Users’ Group, Oct. 3, 2005 Today’s Presentations Jacquard Introduction Jacquard Nodes and CPUs High Speed Interconnect and MVAPICH Compiling Running Jobs Software overview Hands-on Machine room tour

Overview of Jacquard Richard Gerber NERSC User Services NERSC User’s Group October 3, 2005 Oakland, CA

NERCS Users’ Group, Oct. 3, 2005 Presentation Overview Cluster overview Connecting Nodes and processors Node interconnect Disks and file systems Compilers Operating system Message passing interface Batch system and queues Benchmarks and application performance

NERCS Users’ Group, Oct. 3, 2005 Status Status Update Jacquard has been experiencing node failures. While this problem is being worked on we are making Jacquard available to users in a degraded mode. About 200 computational nodes are available, one login node, and about half of the storage nodes that support the GPFS file system. Expect lower than usual I/O performance. Because we may still experience some instability, users will not be charged until Jacquard is returned to full production

NERCS Users’ Group, Oct. 3, 2005 Introduction to Jacquard Named in honor of inventor Joseph Marie Jacquard, whose loom was the first machine to use punch cards to control a sequence of operations. Jacquard is a 640-CPU Opteron cluster running a Linux operating system. Integrated, delivered, and supported by Linux Networx Jacquard has 320 dual-processor nodes available for scientific calculations. (Not dual-core processors.) The nodes are interconnected with a high-speed InfiniBand network. Global shared file storage is provided by a GPFS file system.

NERCS Users’ Group, Oct. 3, 2005 Jacquard

NERCS Users’ Group, Oct. 3, 2005 Jacquard Characteristics Processor typeOpteron 2.2 GHz Processor theoretical peak4.4 GFlops/sec Processors per node2 Number of application nodes/processors 320 / 640 System theoretical peak (computational nodes) 2.8 TFlops/sec Physical memory per node (usable)6 (3-5) GBytes Number of spare application nodes4 Number of login nodes4 Node interconnectInfiniBand Global shared diskGPFS: 30 TBytes usable Batch systemPBS Pro

NERCS Users’ Group, Oct. 3, 2005 Jacquard’s Role Jacquard is meant to be for codes that do not scale well on Seaborg. Hope to relieve Seaborg backlog. Typical job expected to be in the concurrency range of nodes. Applications typically run 4X Seaborg speed. Jobs that cannot scale to large parallel concurrency should benefit from faster CPUs.

NERCS Users’ Group, Oct. 3, 2005 Connecting to Jacquard Interactive shell access is via SSH. ssh [–l login_name] jacquard.nersc.gov Four login nodes for compiling and launching parallel jobs. Parallel jobs do not run on login nodes. Globus file transfer utilities can be used. Outbound network services are open (e.g., ftp). Use hsi for interfacing with HPSS mass storage.

NERCS Users’ Group, Oct. 3, 2005 Nodes and processors Each jacquard node has 2 processors that share 6 GB of memory. OS/network/GPFS uses ~1 (?) GB of that. Each processor is a 2.2 GHz AMD Opteron Processor theoretical peak: 4.4 GFlops/sec Opteron offers advanced 64-bit processor, becoming widely used in HPC.

NERCS Users’ Group, Oct. 3, 2005 Node Interconnect Nodes are connected by an InfiniBand high speed network from Mellanox. Adapters and switches from Mellanox Low latency: ~7µs vs. ~25 µs on Seaborg Bandwidth ~ 2X Seaborg “Fat tree”

NERCS Users’ Group, Oct. 3, 2005 Disks and file systems Homes, scratch, and project directories are in global file system from IBM, GFPS. $SCRATCH environment variable is defined to contain path to a user’s personal scratch space. 30 TBytes total usable disk –5 GByte space, 15,000 inode quota in $HOME per user –50 GByte space, 50,000 inode quota in $SCRATCH per user $SCRATCH gives better performance, but may be purged if space is needed

NERCS Users’ Group, Oct. 3, 2005 Project directories Project directories are coming (some are already here). Designed to facilitate group sharing of code and data. Can be repo- or arbitrary group-based /home/projects/group –For sharing group code /scratch/projects/group –For sharing group data and binaries Quotas TBD

NERCS Users’ Group, Oct. 3, 2005 Compilers High performance Fortran/C/C++ compilers from Pathscale. Fortran compiler: pathf90 C/C++ compiler: pathcc, pathCC MPI compiler scripts use Pathscale compilers “underneath” and have all MPI –I, -L, -l options already defined: –mpif90 –mpicc –mpicxx

NERCS Users’ Group, Oct. 3, 2005 Operating system Jacquard is running Novell SUSE Linux Enterprise Linux 9 Has all the “usual” Linux tools and utilities (gcc, GNU utilities, etc.) It was the first “enterprise-ready” Linux for Opteron. Novell (indirectly) provides support and product lifetime assurances (5 yrs).

NERCS Users’ Group, Oct. 3, 2005 Message passing interface MPI implementation is known as “MVAPICH.” Based on MPICH from Argonne with additions and modifications from LBNL for InfiniBand. Developed and supported ultimately by Mellanox/Ohio State group. Provides standard MPI and MPI/IO functionality.

NERCS Users’ Group, Oct. 3, 2005 Batch system Batch scheduler is PBS Pro from Altair Scripts not much different from LoadLeveler: -> #PBS Queues for interactive, debug, premium charge, regular charge, low charge. Configured to run jobs using nodes (1-256 CPUs).

NERCS Users’ Group, Oct. 3, 2005 Performance and benchmarks Applications run 4x Seaborg, some more, some less NAS Parallel Benchmarks (64-way) are ~ times seaborg Three applications the author has examined: (“-O3 out of the box”): –CAM 3.0 (climate): 3.5 x Seaborg –GTC (fusion): 4.1 x Seaborg –Paratec (materials): 2.9 x Seaborg

NERCS Users’ Group, Oct. 3, 2005 User Experiences Positives –Shorter wait in the queues –Linux; many codes already run under Linux –Good performance for node jobs; some codes scale better than on Seaborg –Opteron is fast

NERCS Users’ Group, Oct. 3, 2005 User Experiences Negatives –Fortran compiler is not common, so some porting issues. –Small disk quotas. –Unstable at times. –Job launch doesn’t work well (can’t pass ENV variables). –Charge factor. –Big endian I/O.

NERCS Users’ Group, Oct. 3, 2005 Today’s Presentations Jacquard Introduction Jacquard Nodes and CPUs High Speed Interconnect and MVAPICH Compiling Running Jobs Software overview Hands-on Machine room tour

NERCS Users’ Group, Oct. 3, 2005 Hands On We have a special queue “blah” with 64 nodes reserved. You may work on your own code. Try building and running test code –Copy to your directory and untar /scratch/scratchdirs/ragerber/NUG.tar –3 NPB parallel benchmarks: ft, mg, sp –Configure in config/make.def –make ft CLASS=C NPROCS=16 –Sample PBS scripts in run/ –Try new MPI version, opt levels, -g, IPM