Software Tools Using PBS. Software tools Portland compilers pgf77 pgf90 pghpf pgcc pgCC Portland debugger GNU compilers g77 gcc Intel ifort icc.

Slides:



Advertisements
Similar presentations
© 2007 IBM Corporation IBM Global Engineering Solutions IBM Blue Gene/P Job Submission.
Advertisements

Chapter 3. MPI MPI = Message Passing Interface Specification of message passing libraries for developers and users –Not a library by itself, but specifies.
Network for Computational Nanotechnology (NCN) Purdue, Norfolk State, Northwestern, UC Berkeley, Univ. of Illinois, UTEP Advanced Portable Batch System.
Using the Argo Cluster Paul Sexton CS 566 February 6, 2006.
Job Submission Using PBSPro and Globus Job Commands.
Koç University High Performance Computing Labs Hattusas & Gordion.
Lecture 2 TTH 03:30AM-04:45PM Dr. Jianjun Hu CSCE569 Parallel Computing University of South Carolina Department of.
Lecture 3 Tuesday, February 10, 2015 [With the help of free online resources]
Network for Computational Nanotechnology (NCN) Purdue, Norfolk State, Northwestern, UC Berkeley, Univ. of Illinois, UTEP Basic Portable Batch System (PBS)
Southgreen HPC system Concepts Cluster : compute farm i.e. a collection of compute servers that can be shared and accessed through a single “portal”
Software Tools Using PBS. Software tools Portland compilers pgf77 pgf90 pghpf pgcc pgCC Portland debugger GNU compilers g77 gcc Intel ifort icc.
Batch Queuing Systems The Portable Batch System (PBS) and the Load Sharing Facility (LSF) queuing systems share much common functionality in running batch.
Using Clusters -User Perspective. Pre-cluster scenario So many different computers: prithvi, apah, tejas, vayu, akash, agni, aatish, falaq, narad, qasid.
Running Jobs on Jacquard An overview of interactive and batch computing, with comparsions to Seaborg David Turner NUG Meeting 3 Oct 2005.
VIPBG LINUX CLUSTER By Helen Wang Sept. 10, 2014.
Tutorial on MPI Experimental Environment for ECE5610/CSC
IT MANAGEMENT OF FME, 21 ST JULY  THE HPC FACILITY  USING PUTTY AND WINSCP TO ACCESS THE SERVER  SENDING FILES TO THE SERVER  RUNNING JOBS 
Parallel Systems Parallel Systems Tools Dr. Guy Tel-Zur.
Introduction to HPC Workshop October Introduction Rob Lane HPC Support Research Computing Services CUIT.
ISG We build general capability Job Submission on the Olympus Cluster J. DePasse; S. Brown, PhD; T. Maiden Pittsburgh Supercomputing Center Public Health.
New MPI Library on the cluster Since WSU’s Grid had an upgrade of its operating system recently, we need to use a new MPI Library to compile and run our.
Using Parallel Computing Resources at Marquette
High Performance Computing
Job Submission on WestGrid Feb on Access Grid.
Sun Grid Engine Grid Computing Assignment – Fall 2005 James Ruff Senior Department of Mathematics and Computer Science Western Carolina University.
Using the BYU Supercomputers. Resources Basic Usage After your account is activated: – ssh ssh.fsl.byu.edu You will be logged in to an interactive node.
Introduction to Scientific Computing on BU’s Linux Cluster Doug Sondak Linux Clusters and Tiled Display Walls Boston University July 30 – August 1, 2002.
Quick Tutorial on MPICH for NIC-Cluster CS 387 Class Notes.
Critical Flags, Variables, and Other Important ALCF Minutiae Jini Ramprakash Technical Support Specialist Argonne Leadership Computing Facility.
 Accessing the NCCS Systems  Setting your Initial System Environment  Moving Data onto the NCCS Systems  Storing Data on the NCCS Systems  Running.
Electronic Visualization Laboratory, University of Illinois at Chicago MPI on Argo-new Venkatram Vishwanath Electronic Visualization.
Sun Grid Engine. Grids Grids are collections of resources made available to customers. Compute grids make cycles available to customers from an access.
VIPBG LINUX CLUSTER By Helen Wang March 29th, 2013.
Intro to Linux/Unix (user commands) Box. What is Linux? Open Source Operating system Developed by Linus Trovaldsa the U. of Helsinki in Finland since.
Bigben Pittsburgh Supercomputing Center J. Ray Scott
Clusters at IIT KANPUR - 1 Brajesh Pande Computer Centre IIT Kanpur.
Carnegie MellonCarnegie Mellon University Robust Speech Group1 An Introduction to the Portable Batch System (PBS) Michael L. Seltzer (with a huge thank.
Introduction to Using SLURM on Discover Chongxun (Doris) Pan September 24, 2013.
Rochester Institute of Technology Job Submission Andrew Pangborn & Myles Maxfield 10/19/2015Service Oriented Cyberinfrastructure Lab,
Using the BYU Supercomputers. Resources Basic Usage After your account is activated: – ssh You will be logged in to an interactive.
Network Queuing System (NQS). Controls batch queues Only on Cray SV1 Presently 8 queues available for general use and one queue for the Cray analyst.
HPC for Statistics Grad Students. A Cluster Not just a bunch of computers Linked CPUs managed by queuing software – Cluster – Node – CPU.
1 High-Performance Grid Computing and Research Networking Presented by David Villegas Instructor: S. Masoud Sadjadi
How to for compiling and running MPI Programs. Prepared by Kiriti Venkat.
Cluster Computing Applications for Bioinformatics Thurs., Sept. 20, 2007 process management shell scripting Sun Grid Engine running parallel programs.
Running Parallel Jobs Cray XE6 Workshop February 7, 2011 David Turner NERSC User Services Group.
1 HPCI Presentation Kulathep Charoenpornwattana. March 12, Outline Parallel programming with MPI Running MPI applications on Azul & Itanium Running.
Introduction to HPC Workshop October Introduction Rob Lane & The HPC Support Team Research Computing Services CUIT.
Portable Batch System – Definition and 3 Primary Roles Definition: PBS is a distributed workload management system. It handles the management and monitoring.
Cliff Addison University of Liverpool NW-GRID Training Event 26 th January 2007 SCore MPI Taking full advantage of GigE.
Debugging Lab Antonio Gómez-Iglesias Texas Advanced Computing Center.
Introduction to Parallel Computing Presented by The Division of Information Technology Computer Support Services Department Research Support Group.
Wouter Verkerke, NIKHEF 1 Using ‘stoomboot’ for NIKHEF-ATLAS batch computing What is ‘stoomboot’ – Hardware –16 machines, each 2x quad-core Pentium = 128.
Introduction to HPC Workshop March 1 st, Introduction George Garrett & The HPC Support Team Research Computing Services CUIT.
Grid Computing: An Overview and Tutorial Kenny Daily BIT Presentation 22/09/2016.
GRID COMPUTING.
Auburn University
PARADOX Cluster job management
Unix Scripts and PBS on BioU
HPC usage and software packages
MPI Basics.
Using Paraguin to Create Parallel Programs
CommLab PC Cluster (Ubuntu OS version)
Postdoctoral researcher Department of Environmental Sciences, LSU
Paul Sexton CS 566 February 6, 2006
Introduction to HPC Workshop
Compiling and Job Submission
Sun Grid Engine.
Quick Tutorial on MPICH for NIC-Cluster
Working in The IITJ HPC System
Presentation transcript:

Software Tools Using PBS

Software tools Portland compilers pgf77 pgf90 pghpf pgcc pgCC Portland debugger GNU compilers g77 gcc Intel ifort icc

…software tools MPICH MPI compilers mpicc mpiCC mpif77 Mpif90 mpirun Your favorite text editor PBS

Using PBS PBS an open source job scheduler* and queue manager

PBS – Portable Batch System Others Torque Sun Grid Engine Platform LSF IBM Load Leveler PBS/Torque widely used

PBS openPBS – open source version PBSpro – commercial version

qsub [options] script Submits jobs to a queue (execution) Job specifications in the file script More on pbs scripts later Pbs will queue, schedule and launch job specified in the command/script qsub returns jobid - jobnumber.servername (9999.master)

qsub – common options -q queuename -- submit to the queue queuename -q -N name -- give the job the name name -N mybigjob1 -I – run the job in interactive mode rather than batch

qsub – common options -o path -- location of pbs output, defaults to ~owner -e path -- location of pbs error file, defaults to ~owner -a datetime --places job in a wait state until datetime datetime is [[[[CC]YY]MM]DD]hhmm[.SS] -- run a job on/after 5:00 pm on Feb. 14 -a

qsub – common options -m [a,b,e] a --mail is sent when the job is aborted by the batch system*. b --mail is sent when the job begins execution. e --mail is sent when the job terminates.

qsub – common options -M useraddress - defines mail addresses for (see –m) -M -S path define the location of the shell to interpret the script -S /bin/csh

qsub – common options -l job resource list (important) nodes=n ppn=m -l nodes=2:ppn=2

qstat - get status of jobs/queues qstat –a – get status report in alternative format

qstat –a – states E - Job is exiting after having run. H - Job is held. Q - job is queued, eligible to run or routed. R - job is running. T - job is being moved to new location. (not here) W - job is waiting for its execution time (qsub -a option) to be reached.

qstat – other options qstat –an -- lists jobs in queue/running. For running jobs shows nodes/processors assigned to the job.

qstat – options qstat –q -- list all queues and that status

qdel – deleting a job from the queue qdel jobid jobid is sequence_number[.server_name] for example – qdel 9999.master or qdel 9999

other PBS commands qhold jobid -- places job jobid on hold qrls jobid -- moves job jobid from hold to queued or executing status qalter jobid – modifies attributes of the job jobid specified in qsub after it is submitted ** not all attributes may be modified

other PBS commands qmove dest jobid -- moves job jobid from current queue to the queue dest for example – qmove short 9999

PBS qsub scripts Standard (almost) shell in whatever shell is supported on the system. Must include the line – mpirun myprog may include other statements/directives – like…

PBS qsub scripts – cont. a simple script – demtest.sh #!/bin/sh #PBS -N dem #PBS -q parallel #PBS -l nodes=2:ppn=2 mpirun -np 4 -machinefile $PBS_NODEFILE hello cat $PBS_NODEFILE

PBS scripts cont… then – qsub demtest.sh

PBS scripts cont… and your get dem.o Hello world! I'm node 1 Hello world! I'm node 2 Hello world! I'm node 3 Hello world! I'm node 0 node7 node6

Xpbs

For more information on PBS commands- System man pages “man qstat” bs/sld001.htm