Getting Started: XSEDE Comet Shahzeb Siddiqui - Software Systems Engineer Office: 222A Computer Building Institute of CyberScience May.

Slides:



Advertisements
Similar presentations
Rhea Analysis & Post-processing Cluster Robert D. French NCCS User Assistance.
Advertisements

SAN DIEGO SUPERCOMPUTER CENTER Using Gordon to Accelerate LHC Science Rick Wagner San Diego Supercomputer Center XSEDE 13 July 22-25, 2013 San Diego, CA.
Research CU Boulder Additional Topics in Research Computing Shelley Knuth Peter Ruprecht
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO SDSC RP Update October 21, 2010.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
HPCC Mid-Morning Break MPI on HPCC Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Research
WEST VIRGINIA UNIVERSITY HPC and Scientific Computing AN OVERVIEW OF HIGH PERFORMANCE COMPUTING RESOURCES AT WVU.
Using Kure and Killdevil
Using HPC for Ansys CFX and Fluent
Academic and Research Technology (A&RT)
Quick Tutorial on MPICH for NIC-Cluster CS 387 Class Notes.
Introduction to UNIX/Linux Exercises Dan Stanzione.
Research Computing with Newton Gerald Ragghianti Newton HPC workshop Sept. 3, 2010.
Getting Started on Topsail Mark Reed Charles Davis ITS Research Computing.
WORK ON CLUSTER HYBRILIT E. Aleksandrov 1, D. Belyakov 1, M. Matveev 1, M. Vala 1,2 1 Joint Institute for nuclear research, LIT, Russia 2 Institute for.
Introduction to HPC resources for BCB 660 Nirav Merchant
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Introduction to the HPCC Jim Leikert System Administrator High Performance Computing Center.
MSc. Miriel Martín Mesa, DIC, UCLV. The idea Installing a High Performance Cluster in the UCLV, using professional servers with open source operating.
Introduction to the HPCC Dirk Colbry Research Specialist Institute for Cyber Enabled Research.
SIMPLE DOES NOT MEAN SLOW: PERFORMANCE BY WHAT MEASURE? 1 Customer experience & profit drive growth First flight: June, minute turn at the gate.
Lab System Environment
Introduction to Using SLURM on Discover Chongxun (Doris) Pan September 24, 2013.
Using the BYU Supercomputers. Resources Basic Usage After your account is activated: – ssh You will be logged in to an interactive.
Parallel Programming on the SGI Origin2000 With thanks to Igor Zacharov / Benoit Marchand, SGI Taub Computer Center Technion Moshe Goldberg,
ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed.
2011/08/23 國家高速網路與計算中心 Advanced Large-scale Parallel Supercluster.
ARCHER Advanced Research Computing High End Resource
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
How to use HybriLIT Matveev M. A., Zuev M.I. Heterogeneous Computations team HybriLIT Laboratory of Information Technologies (LIT), Joint Institute for.
Advanced topics Cluster Training Center for Simulation and Modeling September 4, 2015.
Modules, Compiling WRF, and Running on CHPC Clusters Adam Varble WRF Users Meeting 10/26/15.
Debugging Lab Antonio Gómez-Iglesias Texas Advanced Computing Center.
Introduction to Parallel Computing Presented by The Division of Information Technology Computer Support Services Department Research Support Group.
NREL is a national laboratory of the U.S. Department of Energy, Office of Energy Efficiency and Renewable Energy, operated by the Alliance for Sustainable.
Requesting Resources on an HPC Facility Michael Griffiths and Deniz Savas Corporate Information and Computing Services The University of Sheffield
Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.
OAR : a batch scheduler Grenoble University LIG (Mescal Team)
Advanced Computing Facility Introduction
Linux & Joker – An Introduction
Compute and Storage For the Farm at Jlab
Interacting with the cluster ssh, sftp, & slurm batch scripts
Outline Introduction/Questions
Hands on training session for core skills
GRID COMPUTING.
Auburn University
PARADOX Cluster job management
Assumptions What are the prerequisites? … The hands on portion of the workshop will be on the command-line. If you are not familiar with the command.
HPC usage and software packages
ASU Saguaro 09/16/2016 Jung Hyun Kim.
Computational Physics (Lecture 17)
Joker: Getting the most out of the slurm scheduler
Hodor HPC Cluster LON MNG HPN Head Node Comp Node Comp Node Comp Node
Architecture & System Overview
Introduction to XSEDE Resources HPC Workshop 08/21/2017
Cray Announces Cray Inc.
Introduction to eXtreme Science and Engineering Discovery Environment (XSEDE): Comet and Jetstream Sharon Solis Research Computing Consultant Enterprise.
Using Dogwood Instructor: Mark Reed
Using HPC for Ansys CFX and Fluent
CCR Advanced Seminar: Running CPLEX Computations on the ISE Cluster
Advanced Computing Facility Introduction
Requesting Resources on an HPC Facility
High Performance Computing in Bioinformatics
Using Dogwood Instructor: Mark Reed
Introduction to High Performance Computing Using Sapelo2 at GACRC
Frank Mueller North Carolina State University
Working in The IITJ HPC System
ACI-Ref Virtual Residency 2019
Presentation transcript:

Getting Started: XSEDE Comet Shahzeb Siddiqui - Software Systems Engineer Office: 222A Computer Building Institute of CyberScience May x th 2015

2 System Overview Comet is one of the XSEDE cluster designed by Dell and SDSC that delivers 2.0 PFLOPS of performance. The system was available on XSEDE early 2015 Features include: next gen processor with AVX2, Mellanox FDR InfiniBand interconnect, and Aeon storage San Diego Supercomputer Center (SDSC) at University of California was awarded $12 million grant from NSF to build a petascale supercomputer Comet

3 System Specification

4 Software Environment OS: CentOS Cluster Management: Rocks File System: NFS, Lustre Scheduler: SLURM User Environment: Modules Compilers: Intel, PGI Fortran/C/C++ MPI: Intel MPI, MVAPICH, OpenMPI Debugger: DDT Performance IPM, mpiP, PAPI, TAU

5 Accessing Comet Requirement: –XSEDE account –Submit proposal through XSEDE Allocation Request System Hostname: comet.sdsc.xsede.org Comet supports Single Sign On (SSO) through XSEDE User portal XSEDE SSO is a single point of entry to all XSEDE systems To login to XSEDE portal –To access comet from XSEDE portal gsissh –p 2222 comet.sdsc.xsede.org There are four login nodes when you access Comet: comet-ln[1-4].sdsc.edu

6 Environment Module Environment Module provides a method for dynamically changing the shell environment to use different software Few default modules are loaded at login including MVAPICH implementation of MPI and Intel compiler Its recommended to use MVAPICH and Intel Compiler for running MPI-based application to achieve the best performance

7 Module Command Options

8 Account Management & Charging All jobs run on XSEDE system must be charged to a project. To view a list of authorized projects you have access to use the show_accounts command Charging: The charge unit measure for XSEDE systems is SU (Service Unit) which corresponds to one compute core for one hour. A job that requires 24 cores for 1 hour versus 1 core for 24hours will be charged the same 24 SU

9 Compiling with Intel, GCC, and PGI IntelPGIGCC

10 Running jobs on Comet Comet uses the Simple Linux Utility for Resource Management (SLURM) batch environment for running job in batch mode. sbatch : Submit Job squeue –u user1: view all jobs by user1 scancel : Cancel a job Comet has three queues to run your job:

11 Sample Job Script for MPI and OpenMP #!/bin/bash #SBATCH --job-name="hellompi" #SBATCH --output="hellompi.%j.%N.out" #SBATCH --partition=compute #SBATCH --nodes=2 #SBATCH --ntasks-per-node=24 #SBATCH --export=ALL #SBATCH -t 01:30:00 #This job runs with 2 nodes, 24 cores per node for a total of 48 cores. #ibrun in verbose mode will give binding detail ibrun -v../hello_mpi #!/bin/bash #SBATCH --job-name="hell_openmp" #SBATCH --output="hello_openmp.%j.%N.out" #SBATCH --partition=compute #SBATCH --nodes=1 #SBATCH --ntasks-per-node=24 #SBATCH --export=ALL #SBATCH -t 01:30:00 #SET the number of openmp threads export OMP_NUM_THREADS=24 #Run the job using mpirun_rsh./hello_openmp MPI Job-script OpenMP Job-script

12 Storage Overview on Comet SSD: 250GB of SSD per compute node that can be accessible during job execution in the directory local to each compute node: /scratch/$USER/$SLURM_JOB_ID /oasis/scratch/comet/$USER/temp_project Parallel Lustre Filesystem: 7PB of 200GB/sec performance storage and 6PB of 100GB/sec durable storage /oasis/projects

13 References