MIGRATING TO THE SHARED COMPUTING CLUSTER (SCC) SCV Staff Boston University Scientific Computing and Visualization.

Slides:



Advertisements
Similar presentations
Introduction to Openmp & openACC
Advertisements

Tutorial1: NEMO5 Technical Overview
Profiling your application with Intel VTune at NERSC
Rhea Analysis & Post-processing Cluster Robert D. French NCCS User Assistance.
Using the Argo Cluster Paul Sexton CS 566 February 6, 2006.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
HPCC Mid-Morning Break High Performance Computing on a GPU cluster Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
Software Tools Using PBS. Software tools Portland compilers pgf77 pgf90 pghpf pgcc pgCC Portland debugger GNU compilers g77 gcc Intel ifort icc.
PBS Job Management and Taskfarming Joachim Wagner
Using Kure and Killdevil
Job Submission on WestGrid Feb on Access Grid.
Sun Grid Engine Grid Computing Assignment – Fall 2005 James Ruff Senior Department of Mathematics and Computer Science Western Carolina University.
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
HPCC Mid-Morning Break Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery Introduction to the new GPU (GFX) cluster.
Critical Flags, Variables, and Other Important ALCF Minutiae Jini Ramprakash Technical Support Specialist Argonne Leadership Computing Facility.
TPB Models Development Status Report Presentation to the Travel Forecasting Subcommittee Ron Milone National Capital Region Transportation Planning Board.
Introduction to Boston University’s Scientific Computing Facility Doug Sondak Scientific Computing and Visualization.
Introduction to HPC resources for BCB 660 Nirav Merchant
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Introduction to the HPCC Jim Leikert System Administrator High Performance Computing Center.
MSc. Miriel Martín Mesa, DIC, UCLV. The idea Installing a High Performance Cluster in the UCLV, using professional servers with open source operating.
VIPBG LINUX CLUSTER By Helen Wang March 29th, 2013.
March 3rd, 2006 Chen Peng, Lilly System Biology1 Cluster and SGE.
Introduction to the HPCC Dirk Colbry Research Specialist Institute for Cyber Enabled Research.
:: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: :: GridKA School 2009 MPI on Grids 1 MPI On Grids September 3 rd, GridKA School 2009.
17-April-2007 High Performance Computing Basics April 17, 2007 Dr. David J. Haglin.
Lab System Environment
Common Practices for Managing Small HPC Clusters Supercomputing 12
Advanced SCC Usage Research Computing Services Katia Oleinik
Parallel Programming on the SGI Origin2000 With thanks to Igor Zacharov / Benoit Marchand, SGI Taub Computer Center Technion Moshe Goldberg,
HPC for Statistics Grad Students. A Cluster Not just a bunch of computers Linked CPUs managed by queuing software – Cluster – Node – CPU.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Software Tools Using PBS. Software tools Portland compilers pgf77 pgf90 pghpf pgcc pgCC Portland debugger GNU compilers g77 gcc Intel ifort icc.
2011/08/23 國家高速網路與計算中心 Advanced Large-scale Parallel Supercluster.
ARCHER Advanced Research Computing High End Resource
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
More Unix Naomi Altman. Directories Directory = folder mkdir - makes a new directory rmdir - removes an empty directory cd mydirectory - moves you into.
Portable Batch System – Definition and 3 Primary Roles Definition: PBS is a distributed workload management system. It handles the management and monitoring.
How to use HybriLIT Matveev M. A., Zuev M.I. Heterogeneous Computations team HybriLIT Laboratory of Information Technologies (LIT), Joint Institute for.
Advanced topics Cluster Training Center for Simulation and Modeling September 4, 2015.
Cliff Addison University of Liverpool NW-GRID Training Event 26 th January 2007 SCore MPI Taking full advantage of GigE.
NREL is a national laboratory of the U.S. Department of Energy, Office of Energy Efficiency and Renewable Energy, operated by the Alliance for Sustainable.
Advanced SCC Usage Research Computing Services Katia Oleinik
Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb
Using ROSSMANN to Run GOSET Studies Omar Laldin ( using materials from Jonathan Crider, Harish Suryanarayana ) Feb. 3, 2014.
CFI 2004 UW A quick overview with lots of time for Q&A and exploration.
© 2011 Pittsburgh Supercomputing Center XSEDE 2012 Intro To Our Environment John Urbanic Pittsburgh Supercomputing Center July, 2012.
An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.
Advanced Computing Facility Introduction
Intermediate SCC Usage
Compute and Storage For the Farm at Jlab
Specialized Computing Cluster An Introduction
PARADOX Cluster job management
HPC Roadshow Overview of HPC systems and software available within the LinkSCEEM project.
HPC usage and software packages
White Rose Grid Infrastructure Overview
Welcome to Indiana University Clusters
Using Paraguin to Create Parallel Programs
Hodor HPC Cluster LON MNG HPN Head Node Comp Node Comp Node Comp Node
Introduction to Boston University’s Shared Computing Cluster (SCC)
Integration of Singularity With Makeflow
Advanced TAU Commander
Welcome to our Nuclear Physics Computing System
Paul Sexton CS 566 February 6, 2006
CCR Advanced Seminar: Running CPLEX Computations on the ISE Cluster
Welcome to our Nuclear Physics Computing System
Requesting Resources on an HPC Facility
Introduction to High Performance Computing Using Sapelo2 at GACRC
Quick Tutorial on MPICH for NIC-Cluster
Working in The IITJ HPC System
Presentation transcript:

MIGRATING TO THE SHARED COMPUTING CLUSTER (SCC) SCV Staff Boston University Scientific Computing and Visualization

 Glenn Bresnahan – Director, SCV  MGHPCC  Buy-in Program  Kadin Tseng – HPC Programmer/Consultant  Code migration  Katia Oleinik – Graphics Programmer/Consultant  GPU computing on Katana and SCC  Don Johnson – Scientific Applications/Consultant  Software packages for E&E Topics Migrating to the SCC 2

 At Massachussetts Green High Performance Computing Center, Holyoke  Hardware 1024 cores (64 Intel Xeon E nodes with 16 cores each) -- more later ! 594 TB of usable disk space No GPUs at present. Katana GPUs will be moved to SCC by July 1, 2013  Software Most system and application software packages have been installed Webpage available to check status of packages (later slide)  Operating system is Linux Centos 6.4  Login nodes:scc1.bu.edu,scc2.bu.edu,scc3.bu.edu,scc4.bu.edu  Network speed between main campus and MGHPCC is 10GBit/s  Who has access to the SCC ? Users with Katana/Blue Gene access  Home directory disk storage for each user is 10GB  All projects get 50 GB each on /project and /projectnb to start.  /project is backed up daily (just like on Katana)  In addition, /project and /projectnb are backed up by snapshots ─ deleted files are retrievable by owner  PI may ask for up to 1 TB storage for free; purchase more through buy-in Highlights of Shared Computing Cluster (SCC) Migrating to the SCC 3

 For majority of users, /project and /projectnb migration is done !  If your /project files not migrated yet, you will be contacted soon  Migrated /project and /projectnb are shared by SCC and Katana  Home directories ($HOME) on SCC and Katana are distinct (no migration)  In general, don’t copy dotfiles (e.g.,.cshrc,.bash) among systems, otherwise things may break. Adding contents to them is OK  If you use.aliases on Katana, check before copying over to SCC  On SCC, to access Katana home directory, prepend file path with /katana scc1:~ % cd /katana/usr3/graduate/user scc1:~ % cp myfile /katana/usr3/graduate/....../. scc1:~ % cp -r /katana/usr3/graduate/....../myfolder. Where are my files on SCC and Katana ? Migrating to the SCC 4

To maintain a single copy of a file on both SCC and Katana home directories, you may accomplish it with:  makefile # source is on Katana, executable is local on SCC BASE = /katana/usr3/graduate/username mm_mpi_gnu: ${BASE}/matrix_multiply_mpi.f90 mpif90 -o ${BASE}/matrix_multiply_mpi.f90 -O3 # source on Katana, build exec for AMD & Intel & move back to Katana mm_mpi_katana: ${BASE}matrix_multiply_acc.f90 mpif90 -o ${BASE}matrix_multiply_acc.f90 –O3 -tp x64 mv ${BASE}/  Symbolic link abc.c in folder Misc on SCC points to a real file on Katana scc1:~ % ln –s /katana/usr4/ugrad/name/abc.c Misc/abc.c How to Keep $HOME Files In Sync Migrating to the SCC 5

 Compilers  GNU: gcc, c++, gfortran -- optimize with –O3  PGI: pgcc, pgCC, pgfortran -- optimize with –fast  MPI  MPI wrapper: mpicc, mpiCC, mpif90 – built for GNU compilers (PGI later)  MPI implementation: OpenMPI  mpif90 –show shows what compiler family, linked libs, etc.  OpenMP – shared-memory directive-based parallel computing for CPUs; turn on with -mp (PGI); -fopenmp (GNU)  OpenACC – directive-based parallel computing for GPUs; turn on with: -acc (PGI)  Cuda-based GPU computing (Katia will talk about this) Compilers... Migrating to the SCC 6

 Open Grid Scheduler (OGS) is open source version of Katana’s SGE Batch scripts for both are essentially interchangeable  Batch commands: qsub, qstat, qdel, qhost,...  Files needed by a compute node must reside on SCC, e.g., executables and I/O data files (/katana not supported on nodes)  OpenMP or thread-based jobs: -pe omp N (1 <= N <= 16)  MPI jobs: -pe mpi_X_tasks_per_node N ( X=4, 8, 16; N=i*X; i=1, 2, 3,...)  Example: X=4, N=16 => 4 nodes with 4 cores each (rest of each node shared)  Example: X=16, N=16 => exclusive use of an entire node; 2 nodes if N=32  Serial jobs: omit -pe  Cuda-based GPU jobs: available July 1; use Katana now  OpenACC-based GPU jobs: available July 1; use Katana now  Default walltime limit: 12 hours (2 hours on Katana)  Maximum walltime limit: no limit for now (24/72 hours on Katana)  Core use limit: ≤ 256 cores (fewer later) concurrently per user (128 on Katana) Batch Usages Migrating to the SCC 7

MPI Square Matrix Multiply on SCC Migrating to the SCC 8  2x8 cores case uses -pe mpi_8_tasks_per_node 16  4x8 cores case uses -pe mpi_8_tasks_per_node 32  2x16 cores case uses -pe mpi_16_tasks_per_node 32

MPI Square Matrix Multiply on Katana Migrating to the SCC 9

MPI Square Matrix Multiply on the Blue Gene Migrating to the SCC 10 For this application, wall time is shorter with 16 cores on the SCC than 128 cores on the Blue Gene. With significantly more memory and application software, the SCC is more general-purpose and versatile.

 Primarily to store temporary files during batch jobs  /scratch on login and compute nodes  Not backed up. File life is 10 days  Nodes are assigned at runtime. Node info accessible from $PE_HOSTFILE  I/O addressing /scratch points to local compute node’s scratch  After batch job completes, assigned nodes are reported in your-batch- script.poXXXX. To access your data... scc1:~ % cd /net/scc-ad8/scratch Scratch Disks Migrating to the SCC 11

 Before the end of 2013 Phasing Out of Katana & Blue Gene ? Migrating to the SCC 12

 SCC home page  Status of Software packages:  Software Updates  Snapshots / IS&T Archive / Tape Archive  Technical Summary Relevant SCC Webpages Migrating to the SCC 13

 Research Computing Home Page (  Resource Applications  Help System Web-based tutorials ( (MPI, OpenMP, MATLAB, IDL, Graphics tools) Consultations by appointment Katia Oleinik – Graphics, R, Mathematica Yann Tambouret – HPC, Python Robert Putnam – VR, C programming Kadin Tseng – HPC, MATLAB, Fortran Useful SCV Info Migrating to the SCC 14