Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Usage Seminar for the 64-nodes P4-Xeon Cluster in Science Faculty March 24, 2004.

Slides:



Advertisements
Similar presentations
Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Kickstart Tutorial/Seminar on using the 64-nodes P4-Xeon Cluster in Science Faculty June 11, 2003.
Advertisements

Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Overview of Cluster Hardware and Software Class 2.
Academic Technology Services The UCLA Grid Portal - Campus Grids and the UC Grid Joan Slottow and Prakashan Korambath Research Computing Technologies UCLA.
Chapter One The Essence of UNIX.
CCPR Workshop Lexis Cluster Introduction October 19, 2007 David Ash.
Software Tools Using PBS. Software tools Portland compilers pgf77 pgf90 pghpf pgcc pgCC Portland debugger GNU compilers g77 gcc Intel ifort icc.
IT MANAGEMENT OF FME, 21 ST JULY  THE HPC FACILITY  USING PUTTY AND WINSCP TO ACCESS THE SERVER  SENDING FILES TO THE SERVER  RUNNING JOBS 
Introduction to HPC Workshop October Introduction Rob Lane HPC Support Research Computing Services CUIT.
ISG We build general capability Job Submission on the Olympus Cluster J. DePasse; S. Brown, PhD; T. Maiden Pittsburgh Supercomputing Center Public Health.
Introduction to Running CFX on U2  Introduction to the U2 Cluster  Getting Help  Hardware Resources  Software Resources  Computing Environment  Data.
IT Infrastructure: Software September 18, LEARNING GOALS Identify the different types of systems software. Explain the main functions of operating.
Sun Grid Engine Grid Computing Assignment – Fall 2005 James Ruff Senior Department of Mathematics and Computer Science Western Carolina University.
Linux clustering Morris Law, IT Coordinator, Science Faculty, Hong Kong Baptist University.
Quick Tutorial on MPICH for NIC-Cluster CS 387 Class Notes.
A crash course in njit’s Afs
Introduction to UNIX/Linux Exercises Dan Stanzione.
Local Area Networks: Software © Abdou Illia, Spring 2007 School of Business Eastern Illinois University (Week 8, Thursday 3/1/2007)
Research Computing with Newton Gerald Ragghianti Newton HPC workshop Sept. 3, 2010.
Building service testbeds on FIRE D5.2.5 Virtual Cluster on Federated Cloud Demonstration Kit August 2012 Version 1.0 Copyright © 2012 CESGA. All rights.
WORK ON CLUSTER HYBRILIT E. Aleksandrov 1, D. Belyakov 1, M. Matveev 1, M. Vala 1,2 1 Joint Institute for nuclear research, LIT, Russia 2 Institute for.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
MSc. Miriel Martín Mesa, DIC, UCLV. The idea Installing a High Performance Cluster in the UCLV, using professional servers with open source operating.
BIOSTAT LINUX CLUSTER By Helen Wang October 11, 2012.
VIPBG LINUX CLUSTER By Helen Wang March 29th, 2013.
HPC at HCC Jun Wang Outline of Workshop1 Overview of HPC Computing Resources at HCC How to obtain an account at HCC How to login a Linux cluster at HCC.
Intro to Linux/Unix (user commands) Box. What is Linux? Open Source Operating system Developed by Linus Trovaldsa the U. of Helsinki in Finland since.
Using The Cluster. What We’ll Be Doing Add users Run Linpack Compile code Compute Node Management.
March 3rd, 2006 Chen Peng, Lilly System Biology1 Cluster and SGE.
17-April-2007 High Performance Computing Basics April 17, 2007 Dr. David J. Haglin.
O.S.C.A.R. Cluster Installation. O.S.C.A.R O.S.C.A.R. Open Source Cluster Application Resource Latest Version: 2.2 ( March, 2003 )
AE6382 Secure Shell Usually referred to as ssh, the name refers to both a program and a protocol. The program ssh is one of the most useful networking.
Sharif University of technology, Parallel Processing course, MPI & ADA Server Introduction By Shervin Daneshpajouh.
CCPR Workshop Introduction to the Cluster July 13, 2006.
HPC for Statistics Grad Students. A Cluster Not just a bunch of computers Linked CPUs managed by queuing software – Cluster – Node – CPU.
Cluster Software Overview
How to for compiling and running MPI Programs. Prepared by Kiriti Venkat.
Software Tools Using PBS. Software tools Portland compilers pgf77 pgf90 pghpf pgcc pgCC Portland debugger GNU compilers g77 gcc Intel ifort icc.
Running Parallel Jobs Cray XE6 Workshop February 7, 2011 David Turner NERSC User Services Group.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Introduction to HPC Workshop October Introduction Rob Lane & The HPC Support Team Research Computing Services CUIT.
Unix Servers Used in This Class  Two Unix servers set up in CS department will be used for some programming projects  Machine name: eustis.eecs.ucf.edu.
1 Running MPI on “Gridfarm” Bryan Carpenter February, 2005.
Portable Batch System – Definition and 3 Primary Roles Definition: PBS is a distributed workload management system. It handles the management and monitoring.
Advanced topics Cluster Training Center for Simulation and Modeling September 4, 2015.
Getting Started: XSEDE Comet Shahzeb Siddiqui - Software Systems Engineer Office: 222A Computer Building Institute of CyberScience May.
Introduction to Parallel Computing Presented by The Division of Information Technology Computer Support Services Department Research Support Group.
Wouter Verkerke, NIKHEF 1 Using ‘stoomboot’ for NIKHEF-ATLAS batch computing What is ‘stoomboot’ – Hardware –16 machines, each 2x quad-core Pentium = 128.
Active-HDL Server Farm Course 11. All materials updated on: September 30, 2004 Outline 1.Introduction 2.Advantages 3.Requirements 4.Installation 5.Architecture.
Introduction to HPC Workshop March 1 st, Introduction George Garrett & The HPC Support Team Research Computing Services CUIT.
NREL is a national laboratory of the U.S. Department of Energy, Office of Energy Efficiency and Renewable Energy, operated by the Alliance for Sustainable.
An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.
Advanced Computing Facility Introduction
GRID COMPUTING.
Specialized Computing Cluster An Introduction
PARADOX Cluster job management
HPC usage and software packages
Using Paraguin to Create Parallel Programs
Architecture & System Overview
CommLab PC Cluster (Ubuntu OS version)
Postdoctoral researcher Department of Environmental Sciences, LSU
Telnet/SSH Connecting to Hosts Internet Technology.
Introduction to HPC Workshop
Compiling and Job Submission
Outline Subversion server Sandpit cluster What is it?
CUPS Print Services.
High Performance Computing in Bioinformatics
Introduction to High Performance Computing Using Sapelo2 at GACRC
Quick Tutorial on MPICH for NIC-Cluster
Working in The IITJ HPC System
Presentation transcript:

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Usage Seminar for the 64-nodes P4-Xeon Cluster in Science Faculty March 24, 2004

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Aims and Target audience Aims: –Usage review –Introducing news and events –Sharing among existing users Target audience –Existing and potential cluster users

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Today’s Outline Introduction to upgraded software and its usage Review of serial and parallel job submission Software demo Briefing of coming recurring parallel computing course Sharing and opinions from the existing users

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU System upgrade February 27, 2004 Upgrade to ROCKS SAN adapter inserted Compute nodes all reinstalled Many application software upgrade

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Hardware Configuration 1 master node + 64 compute nodes + Gigabit Interconnection Master node –Dell PE2650, P4-Xeon 2.8GHz x 2 –4GB RAM, 36GB x 2 U160 SCSI (mirror) –Gigabit ethernet ports x 2 SCSI attached storage –Dell PV220S –73GB x 10 (RAID5)

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Hardware Configuration (cont) Compute nodes –Dell PE2650, P4-Xeon 2.8GHz x 2 –2GB RAM, 36GB U160 SCSI HD –Gigabit ethernet ports x 2 Gigabit Interconnect –Extreme Blackdiamond 6816 Gigabit ethernet –256Gb backplane –72 Gigabit ports (8 ports card x 9)

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Software installed Cluster operating system –ROCKS from MPI and PVM libraries –LAM/MPI 7.0.4, MPICH , PVM beolin Compilers –GCC –PGI C/C++/f77/f90/hpf version 5.1 MATH libraries –ATLAS 3.6.0, ScaLAPACK, SPRNG 2.0a Application software –MATLAB 6.1 with MPITB –R –Gromacs 3.2, NAMD2.5, Gamess –Gaussian 03, Q-Chem 2.1 Editors –vi, pico, emacs, joe Queuing system –Torque/PBS, Maui scheduler

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Cluster O.S. – ROCKS Developed by NPACI and SDSC Based on RedHat Entreprise Linux 3.0 Useful command for users to monitor jobs in all nodes. E.g. –cluster-fork date –cluster-ps morris –cluster-kill morris Web based management and monitoring –

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Ganglia

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Hostnames Master node –tdgrocks.sci.hkbu.edu.hk Compute nodes –comp-pvfs-0-1, …, comp-pvfs-0-64 –Short names: cp0-1, cp0-2, …, cp0-64

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Network diagram Master node Compute node Gigibit ethernet switch tdgrocks.sci.hkbu.edu.hk comp-pvfs-0-1 ( ) comp-pvfs-0-2 ( ) comp-pvfs-0-64 ( )

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Login to the master node Login is allowed remotely in all HKBU networked PCs by ssh or vncviewer SSH Login (terminal login) –Using your favourite ssh client software, namely putty, SSHsecureshell on windows and openssh on Linux/UNIXputtySSHsecureshell openssh –E.g. on all SCI workstations (spc01 – spc30), type ssh tdgrocks.sci.hkbu.edu.hk

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Login to the master node VNC Login (graphical login) –Using vncviewer download from –E.g. in spc01 – spc30.sci.hkbu.edu.hk, vncviewer vnc.sci.hkbu.edu.hk:51 –E.g. in windows, run vncviewer and upon asking the server address, type vnc.sci.hkbu.edu.hk:51

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Username and password The unified password authentication has been implemented Same as that of your netware account Password authentication using NDS-AS Setup similar to net1 and net4 in ITSC

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU ssh key generation To make use of multiple nodes in the PC cluster, users are restricted to use ssh. Key generation is done once automatically during first login You may input a passphrase to protect the key pair The key pair is stored in your $HOME/.ssh/

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU User Policy Users are allowed to remote login from other networked PCs in HKBU. All users must use their own user account to login. The master node (frontend) is used only for login, simple editing of program source code, preparing the job dispatching script and dispatching of jobs to compute node. No foreground or background jobs can be run on it. Dispatching of jobs must be done via the PBS system.

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Torque/PBS system Provide a fair and efficient job dispatching and queuing system to the cluster PBS script shall be written for running job Either sequential or parallel jobs can be handled by PBS Jobs error and output are stored in different filenames according to job IDs.

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU PBS script example (sequential) PBS scripts are shell script with directives preceding with #PBS The above example request only 1 node and deliver the job named ‘prime’ in default queue. The PBS system will mail a message after the job executed. #!/bin/bash #PBS -l nodes=1 #PBS -N prime #PBS -m ae #PBS -q default # the above is the PBS directive used in batch queue echo Running on host `hostname` /u1/local/share/example/pbs/prime

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Delivering PBS job Prepare and compile executable cp /u1/local/share/example/pbs/prime.c. cc –o prime prime.c -lm Prepare and edit PBS script as previous cp /u1/local/share/example/pbs/prime.bat. Submit the job qsub prime.bat

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU PBS script example (parallel) #!/bin/sh #PBS -N cpi #PBS -r n #PBS -e cpi.err #PBS -o cpi.log #PBS -m ae #PBS -l nodes=5:ppn=2 #PBS -l walltime=01:00:00 # This job's working directory echo Working directory is $PBS_O_WORKDIR cd $PBS_O_WORKDIR echo Running on host `hostname` echo This jobs runs on the following processors: echo `cat $PBS_NODEFILE` # Define number of processors NPROCS=`wc -l < $PBS_NODEFILE` echo This job has allocated $NPROCS nodes # Run the parallel MPI executable “cpi” /u1/local/mpich-1.2.5/bin/mpirun -v -machinefile $PBS_NODEFILE -np $NPROCS /u1/local/share/example/pbs/cpi

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU Delivering parallel jobs Copy the PBS script examples cp /u1/local/share/example/pbs/runcpi. Submit the PBS job qsub runcpi Note the error and output files named cpi.e??? and cpi.o???

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU End of Part 1 Thank you!

Using the P4-Xeon cluster HPCCC, Science Faculty, HKBU MPICH and LAM’s PATH MPICH with gcc –/u1/local/mpich-1.2.5/bin MPICH with pgi –/u1/local/mpich-pgi/bin LAM –/usr/bin LAM –/u1/local/lam-7.0.4/bin