How to use HybriLIT Matveev M. A., Zuev M.I. Heterogeneous Computations team HybriLIT Laboratory of Information Technologies (LIT), Joint Institute for.

Slides:



Advertisements
Similar presentations
Issues of HPC software From the experience of TH-1A Lu Yutong NUDT.
Advertisements

1 Computational models of the physical world Cortical bone Trabecular bone.
Computing with Accelerators: Overview ITS Research Computing Mark Reed.
Hybrid Redux: CUDA / MPI 1. CUDA / MPI Hybrid – Why? 2  Harness more hardware  16 CUDA GPUs > 1!  You have a legacy MPI code that you’d like to accelerate.
HPCC Mid-Morning Break High Performance Computing on a GPU cluster Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
Using Kure and Killdevil
Contemporary Languages in Parallel Computing Raymond Hummel.
HPCC Mid-Morning Break Interactive High Performance Computing Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
HPCC Mid-Morning Break Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery Introduction to the new GPU (GFX) cluster.
ORIGINAL AUTHOR JAMES REINDERS, INTEL PRESENTED BY ADITYA AMBARDEKAR Overview for Intel Xeon Processors and Intel Xeon Phi coprocessors.
MIGRATING TO THE SHARED COMPUTING CLUSTER (SCC) SCV Staff Boston University Scientific Computing and Visualization.
OpenMP in a Heterogeneous World Ayodunni Aribuki Advisor: Dr. Barbara Chapman HPCTools Group University of Houston.
WORK ON CLUSTER HYBRILIT E. Aleksandrov 1, D. Belyakov 1, M. Matveev 1, M. Vala 1,2 1 Joint Institute for nuclear research, LIT, Russia 2 Institute for.
BY: ALI AJORIAN ISFAHAN UNIVERSITY OF TECHNOLOGY 2012 GPU Architecture 1.
Introduction to the HPCC Jim Leikert System Administrator High Performance Computing Center.
VIPBG LINUX CLUSTER By Helen Wang March 29th, 2013.
Introduction to the HPCC Dirk Colbry Research Specialist Institute for Cyber Enabled Research.
Sobolev Showcase Computational Mathematics and Imaging Lab.
Debugging and Profiling GMAO Models with Allinea’s DDT/MAP Georgios Britzolakis April 30, 2015.
Lab System Environment
Taking the Complexity out of Cluster Computing Vendor Update HPC User Forum Arend Dittmer Director Product Management HPC April,
Swarm on the Biowulf2 Cluster Dr. David Hoover, SCB, CIT, NIH September 24, 2015.
Introduction to Using SLURM on Discover Chongxun (Doris) Pan September 24, 2013.
Using the BYU Supercomputers. Resources Basic Usage After your account is activated: – ssh You will be logged in to an interactive.
GPU Architecture and Programming
Multi-Core Development Kyle Anderson. Overview History Pollack’s Law Moore’s Law CPU GPU OpenCL CUDA Parallelism.
ARCHES: GPU Ray Tracing I.Motivation – Emergence of Heterogeneous Systems II.Overview and Approach III.Uintah Hybrid CPU/GPU Scheduler IV.Current Uintah.
Software Tools Using PBS. Software tools Portland compilers pgf77 pgf90 pghpf pgcc pgCC Portland debugger GNU compilers g77 gcc Intel ifort icc.
Belgrade, 24 September 2014 George S. Markomanolis, Oriol Jorba, Kim Serradell Introduction to an HPC environment.
Contemporary Languages in Parallel Computing Raymond Hummel.
HPC at HCC Jun Wang Outline of Workshop2 Familiar with Linux file system Familiar with Shell environment Familiar with module command Familiar with queuing.
Getting Started: XSEDE Comet Shahzeb Siddiqui - Software Systems Engineer Office: 222A Computer Building Institute of CyberScience May.
Modules, Compiling WRF, and Running on CHPC Clusters Adam Varble WRF Users Meeting 10/26/15.
Debugging Lab Antonio Gómez-Iglesias Texas Advanced Computing Center.
Introduction to Parallel Computing Presented by The Division of Information Technology Computer Support Services Department Research Support Group.
Requesting Resources on an HPC Facility Michael Griffiths and Deniz Savas Corporate Information and Computing Services The University of Sheffield
Multicore Applications in Physics and Biochemical Research Hristo Iliev Faculty of Physics Sofia University “St. Kliment Ohridski” 3 rd Balkan Conference.
11 Brian Van Straalen Portable Performance Discussion August 7, FASTMath SciDAC Institute.
Sobolev(+Node 6, 7) Showcase +K20m GPU Accelerator.
Martin Kruliš by Martin Kruliš (v1.1)1.
J.J. Keijser Nikhef Amsterdam Grid Group MyFirstMic experience Jan Just Keijser 26 November 2013.
Advanced Computing Facility Introduction
Outline Introduction/Questions
Hands on training session for core skills
GPGPU use cases from the MoBrain community
Auburn University
NFV Compute Acceleration APIs and Evaluation
HPC Roadshow Overview of HPC systems and software available within the LinkSCEEM project.
Modern supercomputers, Georgian supercomputer project and usage areas
Assumptions What are the prerequisites? … The hands on portion of the workshop will be on the command-line. If you are not familiar with the command.
HPC usage and software packages
Volunteer Computing for Science Gateways
Using Longleaf ITS Research Computing
GPU Computing Jan Just Keijser Nikhef Jamboree, Utrecht
Heterogeneous Computation Team HybriLIT
Joker: Getting the most out of the slurm scheduler
Hodor HPC Cluster LON MNG HPN Head Node Comp Node Comp Node Comp Node
High-Performance Computing System
Using Dogwood Instructor: Mark Reed
CCR Advanced Seminar: Running CPLEX Computations on the ISE Cluster
Advanced Computing Facility Introduction
Overview of HPC systems and software available within
Requesting Resources on an HPC Facility
High Performance Computing in Bioinformatics
Using OpenMP offloading in Charm++
Using Dogwood Instructor: Mark Reed
MPI MPI = Message Passing Interface
Frank Mueller North Carolina State University
Working in The IITJ HPC System
GPU and Large Memory Jobs on Ibex
Presentation transcript:

How to use HybriLIT Matveev M. A., Zuev M.I. Heterogeneous Computations team HybriLIT Laboratory of Information Technologies (LIT), Joint Institute for Nuclear Research

Computation nodes Processor Number Num. of cores TDP (W) Base Frequency(GHz) Memory Size (Gb) Intel Xeon CPUE v NVIDIA TESLAK NVIDIA TESLAK20X Intel Xeon Phi5110P Intel Xeon Phi7120P Hardware Single precision76.8 Tf Double precision :29.1 Tf TDP:6.4 kW

Lang GNU Intel OpenMPI 1.6.5, Intel MPI CUDA 5.5, 6.0, 7.0 PGI 15.3 Cgcciccmpiccmpiiccnvccpgcc C++g++icpcmpicxxmpiicpcnvccpgc++ Fortrangfortranifort mpif77/ mpif90 mpiifortpgf77/pgf90 Software OpenMP: GNU 4.4.7, Intel OpenCL: CUDA 6.0, 7.0, Intel OpenCL 4.6.0

SLURM To run applications by batch mode; MODULES To set environment variables; System software Scientific Linux 6.7. Parallel technologies for C/C++, Fortran: MPI, OpenMP, CUDA, OpenCL I II III

Module add Compile application Run application Run task 3 steps

Run task Step 1: m odule add $ module avail $ module add openmpi/1.6.5 $ module add cuda-6.5-x86_64 $ module list Step 2: compile application $ mpicc source_mpi.c –o exec_mpi $ nvcc source_cuda.cu –o exec_cuda

Run task $ srun -p test mpiexec –n 1 exec_mpi Step 3: r un CPU application by SRUN Hello world from process 0 of 1 by host blade04.hydra.local SLURM note -p, --partition=name. Partition avail: cpu7 nodes; gpu4 nodes; phi2 nodes; test2 nodes. -n, --ntasks=num. mpiexec – to run mpi application.

Run task $ srun -p test --gres=gpu:3 exec_cuda Step 3: r un GPU application by SRUN Hello World! Detected 3 CUDA capable device(s) SLURM note -p, --partition=name. --gres=gpu:3 – number of gpu or phi devices to task.

Run task $ sbatch script.sh Submitted batch job 586 $ cat slurm-586.out Hello world from process 0 of 1 by host blade04.hydra.local Step 3: r un application by SBATCH $ cat script.sh #!/bin/sh #SBATCH –p test mpiexec –n 1./exec_mpi

SRUN or SBATCH? srun -p test -w blade06 -n 1 --gres=gpu:3 exec_cuda Hello World! Detected 3 CUDA capable device(s) OR $ sbatch script.sh Submitted batch job 587 $ cat slurm-587.out Hello World! Detected 3 CUDA capable device(s) #!/bin/sh #SBATCH –p test #SBATCH –w blade06 #SBATCH –n 1 #SBATCH --gres=gpu:3./exec_cuda SBATCH

Basic SLURM commands JOBID NAME USER STATE TIME NODE 312 script.sh user R 0:31 blade05 $ sinfo – to show status of partitions: PART AVAIL TIMELIM NODES STATE cpu up infinite 7 idle gpu up infinite 4 idle phi up infinite 2 idle test up infinite 1 idle $scancel – to cancel running task. $ srun – to run application by interactive mode; $ sbatch – to run application by queue mode; $ squeue – to show status of tasks:

Thanks for attention!