An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.

Slides:



Advertisements
Similar presentations
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Advertisements

Introduction to HPC Workshop October Introduction Rob Lane HPC Support Research Computing Services CUIT.
Sun Grid Engine Grid Computing Assignment – Fall 2005 James Ruff Senior Department of Mathematics and Computer Science Western Carolina University.
Using the BYU Supercomputers. Resources Basic Usage After your account is activated: – ssh ssh.fsl.byu.edu You will be logged in to an interactive node.
Introduction to RCC for Intro to MRI 2014 July 25, 2014.
Research Computing with Newton Gerald Ragghianti Nov. 12, 2010.
Introduction to UNIX/Linux Exercises Dan Stanzione.
Types of Operating System
Research Computing with Newton Gerald Ragghianti Newton HPC workshop Sept. 3, 2010.
 Accessing the NCCS Systems  Setting your Initial System Environment  Moving Data onto the NCCS Systems  Storing Data on the NCCS Systems  Running.
Introduction to HPC resources for BCB 660 Nirav Merchant
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Introduction to the HPCC Jim Leikert System Administrator High Performance Computing Center.
VIPBG LINUX CLUSTER By Helen Wang March 29th, 2013.
HPC at HCC Jun Wang Outline of Workshop1 Overview of HPC Computing Resources at HCC How to obtain an account at HCC How to login a Linux cluster at HCC.
Bigben Pittsburgh Supercomputing Center J. Ray Scott
Introduction to the HPCC Dirk Colbry Research Specialist Institute for Cyber Enabled Research.
17-April-2007 High Performance Computing Basics April 17, 2007 Dr. David J. Haglin.
Using the BYU Supercomputers. Resources Basic Usage After your account is activated: – ssh You will be logged in to an interactive.
Looking Ahead: A New PSU Research Cloud Architecture Chuck Gilbert - Systems Architect and Systems Team Lead Research CI Coordinating Committee Meeting.
Katie Antypas User Services Group Lawrence Berkeley National Lab 17 February 2012 JGI Training Series.
HPC for Statistics Grad Students. A Cluster Not just a bunch of computers Linked CPUs managed by queuing software – Cluster – Node – CPU.
Running Parallel Jobs Cray XE6 Workshop February 7, 2011 David Turner NERSC User Services Group.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Introduction to HPC Workshop October Introduction Rob Lane & The HPC Support Team Research Computing Services CUIT.
Remote & Collaborative Visualization. TACC Remote Visualization Systems Longhorn – Dell XD Visualization Cluster –256 nodes, each with 48 GB (or 144 GB)
Portable Batch System – Definition and 3 Primary Roles Definition: PBS is a distributed workload management system. It handles the management and monitoring.
Introduction to Hartree Centre Resources: IBM iDataPlex Cluster and Training Workstations Rob Allan Scientific Computing Department STFC Daresbury Laboratory.
Advanced topics Cluster Training Center for Simulation and Modeling September 4, 2015.
Getting Started: XSEDE Comet Shahzeb Siddiqui - Software Systems Engineer Office: 222A Computer Building Institute of CyberScience May.
NREL is a national laboratory of the U.S. Department of Energy, Office of Energy Efficiency and Renewable Energy, operated by the Alliance for Sustainable.
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
Using ROSSMANN to Run GOSET Studies Omar Laldin ( using materials from Jonathan Crider, Harish Suryanarayana ) Feb. 3, 2014.
Advanced Computing Facility Introduction
Compute and Storage For the Farm at Jlab
Interacting with the cluster ssh, sftp, & slurm batch scripts
Hackinars in Bioinformatics
Accessing the VI-SEEM infrastructure
GRID COMPUTING.
Specialized Computing Cluster An Introduction
Auburn University
Welcome to Indiana University Clusters
PARADOX Cluster job management
INTRODUCTION TO VIPBG LINUX CLUSTER
What is HPC? High Performance Computing (HPC)
HPC usage and software packages
INTRODUCTION TO VIPBG LINUX CLUSTER
Welcome to Indiana University Clusters
How to use the HPCC to do stuff
Creating and running applications on the NGS
BIOSTAT LINUX CLUSTER By Helen Wang October 29, 2015.
Hodor HPC Cluster LON MNG HPN Head Node Comp Node Comp Node Comp Node
Architecture & System Overview
CommLab PC Cluster (Ubuntu OS version)
Practice #0: Introduction
Postdoctoral researcher Department of Environmental Sciences, LSU
File Transfer Olivia Irving and Cameron Foss
Welcome to our Nuclear Physics Computing System
Introduction to HPC Workshop
Introduction to TAMNUN server and basics of PBS usage
Compiling and Job Submission
Welcome to our Nuclear Physics Computing System
High-Performance Computing at the Martinos Center
Advanced Computing Facility Introduction
Requesting Resources on an HPC Facility
Unit 1: Introduction to Operating System
High Performance Computing in Bioinformatics
Introduction to High Performance Computing Using Sapelo2 at GACRC
Quick Tutorial on MPICH for NIC-Cluster
Working in The IITJ HPC System
Presentation transcript:

An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing

✦ Mission Improve opportunities for research and scholarship Improve competitiveness in securing external funding Provide high-performance computing resources and support to UF researchers UF Research Computing

✦ Funding Faculty participation (i.e. grant money) provides funds for hardware purchases  Matching grant program! ✦ Comprehensive management Hardware maintenance and 24x7 monitoring Relieve researchers of the majority of systems administration tasks UF Research Computing

✦ Shared Hardware Resources Over 5000 cores (AMD and Intel) InfiniBand interconnect (2) 230 TB Parallel Distributed Storage 192 TB NAS Storage (NFS, CIFS) NVidia Tesla (C1060) GPUs (8) Nvida Fermi (M2070) GPUs (16) Several large memory (512GB) nodes UF Research Computing

Machine room at Larson Hall UF Research Computing

✦ So what’s the point? Large-scale resources available Staff and support to help you succeed

UF Research Computing ✦ Where do you start?

✦ User Accounts Qualifications:  Current UF faculty, UF graduate student, and researchers Request at: Requirements:  GatorLink Authentication  Faculty sponsorship for graduate students and researchers UF Research Computing

✦ Account Policies Personal activities are strictly prohibited on HPC Center systems Class accounts deleted at end of semester Data are not backed up! Home directories must not be used for I/O  Use /scratch/hpc/ Storage systems may not be used to archive data from other systems Passwords expire every 6 months UF Research Computing

Cluster basics Login to head node Head node(s) Tell the scheduler what you want to do Scheduler Your job runs on the cluster Compute resources

Cluster login submit.hpc.ufl.edu ssh /home/$U SER Windows: PuTTY Mac/Linux: Terminal Login to head node Head node(s)

Logging in

Linux Command Line ✦ Lots of online resources Google: linux cheat sheet ✦ Training sessions: Feb 9 th ✦ User manuals for applications

✦ Storage Home Area: /home/$USER  For code compilation and user file management only, do not write job output here On UF-HPC: Lustre File System  /scratch/hpc/$USER, 114 TB UF Research Computing Must be used for all file I/O

Storage at HPC submit.hpc.ufl.edu ssh /scratch/hpc/$USER /home/$U SER $ cd /scratch/hpc/magitz/ Copy your data to submit using scp or a SFTP program like Cyberduck or FileZilla

Scheduling a job ✦ Need to tell scheduler what you want to do Information about how long your job will run How many CPUs you want and how you want them grouped How much RAM your job will use The commands that will be run Tell the scheduler what you want to do Scheduler

General Cluster Schematic submit.hpc.ufl.edu ssh /scratch/hpc/$USER /home/$U SER qsub

✦ Job Scheduling and Usage PBS/Torque batch system Test nodes (test01, 04, 05) available for interactive use, testing and short jobs  e.g.: ssh test01 Job scheduler selects jobs based on priority  Priority is determined by several components  Investors have higher priority  Non-investor jobs limited to 8 processor equivalents (PEs)  RAM: requests beyond a couple GB/core starts counting toward the total PE value of a job UF Research Computing

✦ Ordinary Shell Script #!/bin/bash pwd date hostname UF Research Computing Read the manual for your application of choice. Commands typed on the command line can be put in a script.

✦ Submission Script #!/bin/bash # #PBS -N My_Job_Name #PBS -M #PBS -m abe #PBS -o My_Job_Name.log #PBS -j oe #PBS -l nodes=1:ppn=1 #PBS -l pmem=900mb #PBS -l walltime=00:05:00 cd $PBS_O_WORKDIR pwd date hostname UF Research Computing Tell the scheduler what you want to do Scheduler

General Cluster Schematic submit.hpc.ufl.edu ssh /scratch/hpc/$USER /home/$U SER ssh test01 qsub test01 test04 test05 Test nodes for interactive sessions, testing scripts and short jobs Access from submit

✦ Job Management qsub : job submission qstat –u : check queue status qdel : job deletion  qdelmine : to delete all of your jobs UF Research Computing

✦ Current Job UFHPC submit: job queue for general use  investor: dedicated queue for investors  other: job queue for non-investors testq: test queue for small and short jobs tesla: queue for jobs utilizing GPUs: #PBS –q tesla UF Research Computing

✦ Using GPUs Interactively We still need to schedule them qsub -I -l nodes=1:gpus=1,walltime=01:00:00 -q tesla More information: UF Research Computing

✦ Help and Support Help Request Tickets   Not just for “bugs” but for any kind of question or help requests  Searchable database of solutions We are here to help!  UF Research Computing

✦ Help and Support (Continued)  Documents on hardware and software resources  Various user guides  Many sample submission scripts  Frequently Asked Questions  Account set up and maintenance UF Research Computing