Research Consulting & Faculty Engagement

Slides:



Advertisements
Similar presentations
Rhea Analysis & Post-processing Cluster Robert D. French NCCS User Assistance.
Advertisements

High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
Installing and running COMSOL on a Windows HPCS2008(R2) cluster
Research Computing with Newton Gerald Ragghianti Nov. 12, 2010.
HPCC Mid-Morning Break Interactive High Performance Computing Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
Cluster Components Compute Server Disk Storage Image Server.
Hands-On Microsoft Windows Server 2008 Chapter 1 Introduction to Windows Server 2008.
Introduction to UNIX/Linux Exercises Dan Stanzione.
MIGRATING TO THE SHARED COMPUTING CLUSTER (SCC) SCV Staff Boston University Scientific Computing and Visualization.
Research Computing with Newton Gerald Ragghianti Newton HPC workshop Sept. 3, 2010.
WORK ON CLUSTER HYBRILIT E. Aleksandrov 1, D. Belyakov 1, M. Matveev 1, M. Vala 1,2 1 Joint Institute for nuclear research, LIT, Russia 2 Institute for.
Research Support Services Research Support Services.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Introduction to the HPCC Jim Leikert System Administrator High Performance Computing Center.
Overview of the Computer Resource Team (CRT) Blaise Barney (LLNL) Rob Cunningham (LANL) Barbara Jennings (Sandia) PSAAP Kickoff Meeting July 8, 2008 Albuquerque,
VIPBG LINUX CLUSTER By Helen Wang March 29th, 2013.
HPC at HCC Jun Wang Outline of Workshop1 Overview of HPC Computing Resources at HCC How to obtain an account at HCC How to login a Linux cluster at HCC.
Introduction to the HPCC Dirk Colbry Research Specialist Institute for Cyber Enabled Research.
Module 1: Installing and Configuring Servers. Module Overview Installing Windows Server 2008 Managing Server Roles and Features Overview of the Server.
Getting Started on Emerald Research Computing Group.
The LBNL Perceus Cluster Infrastructure Next Generation Cluster Provisioning and Management October 10, 2007 Internet2 Fall Conference Gary Jung, SCS Project.
Running Parallel Jobs Cray XE6 Workshop February 7, 2011 David Turner NERSC User Services Group.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Unix Servers Used in This Class  Two Unix servers set up in CS department will be used for some programming projects  Machine name: eustis.eecs.ucf.edu.
Introduction to Hartree Centre Resources: IBM iDataPlex Cluster and Training Workstations Rob Allan Scientific Computing Department STFC Daresbury Laboratory.
Advanced topics Cluster Training Center for Simulation and Modeling September 4, 2015.
NREL is a national laboratory of the U.S. Department of Energy, Office of Energy Efficiency and Renewable Energy, operated by the Alliance for Sustainable.
An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.
Advanced Computing Facility Introduction
Compute and Storage For the Farm at Jlab
Workstations & Thin Clients
Hackinars in Bioinformatics
GRID COMPUTING.
Specialized Computing Cluster An Introduction
Welcome to Indiana University Clusters
PARADOX Cluster job management
BEST CLOUD COMPUTING PLATFORM Skype : mukesh.k.bansal.
HPC usage and software packages
Working With Azure Batch AI
How to use the HPCC to do stuff
Chapter 2: System Structures
Hodor HPC Cluster LON MNG HPN Head Node Comp Node Comp Node Comp Node
Creating an Oracle Database
Architecture & System Overview
OSG Connect and Connect Client
Introduction to XSEDE Resources HPC Workshop 08/21/2017
Introduction to Operating System (OS)
Assignment Preliminaries
LQCD Computing Operations
Bomgar Remote support software
Postdoctoral researcher Department of Environmental Sciences, LSU
XSEDE’s Campus Bridging Project
Welcome to our Nuclear Physics Computing System
Grid Means Business OGF-20, Manchester, May 2007
Introduction to HPC Workshop
Center for High Performance Computing
Shared Research Computing Policy Advisory Committee (SRCPAC)
20409A 7: Installing and Configuring System Center 2012 R2 Virtual Machine Manager Module 7 Installing and Configuring System Center 2012 R2 Virtual.
CCR Advanced Seminar: Running CPLEX Computations on the ISE Cluster
HC Hyper-V Module GUI Portal VPS Templates Web Console
Welcome to our Nuclear Physics Computing System
Advanced Computing Facility Introduction
High Performance Computing in Bioinformatics
Introduction to High Performance Computing Using Sapelo2 at GACRC
Quick Tutorial on MPICH for NIC-Cluster
Working in The IITJ HPC System
Division of Engineering Computing Services
Introduction to research computing using Condor
Presentation transcript:

Research Consulting & Faculty Engagement Overview of CHPC Anita Orendt Assistant Director Research Consulting & Faculty Engagement anita.orendt@utah.edu

CHPC Mission The University of Utah's Center for High Performance Computing provides large-scale computer systems, storage, networking, and the expertise to optimize the use of these high-end computer technologies. CHPC supports faculty and research groups whose main focus requires computing, storage, and advanced networking as core instruments central to their research. 9/19/2018 http://www.chpc.utah.edu Slide 2

Downtown Data Center Came online spring 2012 Shared with enterprise (academic/hospital) groups (wall between rooms) 92 racks and 1.2MW of power with upgrade path to add capacity for research computing Metro optical ring connecting campus, data center, & internet2 24/7/365 facility

Overview HPC Clusters Storage Protected Environment Software http://www.chpc.utah.edu 9/19/2018 Overview HPC Clusters Storage Protected Environment Software Other Services Accounts & Allocation Getting Started Getting Help 9/19/2018 http://www.chpc.utah.edu

CHPC Clusters Condominium Model http://www.chpc.utah.edu 9/19/2018 CHPC Clusters Condominium Model General resources – No charge for usage Allocation process for cycles on kingspeak and ember Out of allocation – freecycle allowed (with pre-emption) Owner nodes 20 core Intel IvyBridge (2.5GHz cpu speed), 64GB RAM, 1TB local hard drive, 5 year warranty, IB connectivity @ ~$5500/node Others allowed as owner-guest when not in use (pre-emption) Lonepeak – no allocation and no pre-emption Two owner clusters – ash and telluride 9/19/2018 http://www.chpc.utah.edu

CHPC Provides Core Infrastructure Physical needs (racks, power, cooling) Ethernet and core IB fabric IB to ethernet bridging Login/head/management nodes Scheduling, batch and allocation systems HPC scratch space Some application and licensing costs High speed data transfer nodes Emerging “science DMZ” network 8/25/2014 http://www.chpc.utah.edu Slide 6

NFS Ember Lonepeak 141 nodes/1692 cores Infiniband and GigE http://www.chpc.utah.edu 9/19/2018 NFS Ember 141 nodes/1692 cores Infiniband and GigE General 70 nodes/852 cores /scratch/lonepeak/serial Lonepeak 16 nodes/256 cores Large memory, GigE Kingspeak 199 nodes/3408 cores Infiniband and GigE General 48 nodes/832 cores 12 GPU nodes (6 General) Switch Ash cluster (417 nodes) Telluride cluster (89 nodes) /scratch/ibrix/chpc_gen NFS Apex Arch homerfs NFS Home Directories & Group Directories /scratch/kingspeak/serial NFS Administrative Nodes 9/19/2018 http://www.chpc.utah.edu

Useful Information Getting Started Guide CHPC policies http://www.chpc.utah.edu 9/19/2018 Useful Information Getting Started Guide https://www.chpc.utah.edu/docs/manuals/getting_started/ CHPC policies https://wiki.chpc.utah.edu/display/policy/CHPC+Policy+Manual Cluster Usage Guides https://wiki.chpc.utah.edu/display/DOCS/Kingspeak+User+Guide https://wiki.chpc.utah.edu/display/DOCS/Ember+User+Guide https://wiki.chpc.utah.edu/display/DOCS/Lonepeak+User+Guide Application Documentation https://wiki.chpc.utah.edu/display/DOCS/CHPC+Software+Documentation /scratch/parallel going away, may come back as something else. 9/19/2018 http://www.chpc.utah.du

Storage CHPC home directory space NFS mounted on all HPC platforms http://www.chpc.utah.edu 9/19/2018 Storage CHPC home directory space NFS mounted on all HPC platforms /uufs/chpc.utah.edu/common/home/<uNID> Home directories not backed up (some exceptions) Periodic Archive Option available with purchase of tapes Groups can purchase home and/or group space 1TB max home directory/group @$500/TB – 5 year warranty; tape backup (nightly incremental/weekly full) Group space @$210/TB; can get quarterly archives with purchase of tapes /scratch/parallel going away, may come back as something else. 9/19/2018 http://www.chpc.utah.du

Scratch File Systems Ember, Kingspeak, and Ash /scratch/ibrix/chpc_gen (56 TB) – also on telluride /scratch/kingspeak/serial (175 TB) Lonepeak /scratch/lonepeak/serial (33 TB) All NFS mounted on all nodes Scrubbed of files older than 60 days All nodes also have a local scratch space 9/19/2018 http://www.chpc.utah.edu

File Systems Remember NFS mounted spaces are a shared resource! Access speed based on connectivity Local disk fastest – local to each node; varies in size /scratch/local NFS /scratch/kingspeak/serial /scratch/ibrix/chpc_gen home directories and group spaces (don’t use for large i/o!) Remember NFS mounted spaces are a shared resource! To check the current status of the file systems (and clusters) -- https://www.chpc.utah.edu/chpc/systems/graphs.php 9/19/2018 http://www.chpc.utah.edu

Protected Environment http://www.chpc.utah.edu 9/19/2018 Protected Environment Dedicated protected resources for handling of data/projects with protected information Currently HIPAA, looking at FISMA & FERPA Have both general, owner and VM resources HPC – apexarch with associated homerfs file system Significant area of growth for CHPC Described in recent paper (Bradford et al.) www.ncbi.nlm.nih.gov/pubmed/23911553 Will be changing to web maintenance. 9/19/2018 http://www.chpc.utah.edu

http://www.chpc.utah.edu 9/19/2018 Software CHPC supports a wide array of software – partial list at https://wiki.chpc.utah.edu/display/DOCS/CHPC+Software+Documentation For ALL clusters (built with mpich2): /uufs/chpc.utah.edu/sys/pkg/packagename Specific build for EACH cluster (usually due to Infiniband/MPI needs: /uufs/cluster /sys/pkg/packagename We work with research groups to install needed software, included software licensed only to a given group Will be changing to web maintenance. 9/19/2018 http://www.chpc.utah.edu

Other Services Windows Statistics Servers Virtual Machine (VM) Farms http://www.chpc.utah.edu 9/19/2018 Other Services Windows Statistics Servers Two servers each with 48 core and 512 GB RAM, one for general (kachina) and the other for use in protected environment (swasey) SAS 9.3 with text miner, SPSS, R, STATA, Mathematica, Matlab Virtual Machine (VM) Farms For applications that do not fit in our normal HPC environment Allows for maximum use of infrastructure and minimizes physical servers to be managed Two farms – one for general needs and the second for needs in the protected environment Setting up MOU/SLA with VM owners prior to provisioning of VM Will be changing to web maintenance. 9/19/2018 http://www.chpc.utah.edu

Accounts CHPC uses campus uNID and password http://www.chpc.utah.edu 9/19/2018 Accounts CHPC uses campus uNID and password Account Application Procedure – Online process Complete CHPC account form http://www.chpc.utah.edu/apps/profile/account_request.php PIs must have account and also approve accounts for any members of their research group (can delegate) For collaborators outside of University of Utah must complete affiliate form with HR to get uNID https://www.hr.utah.edu/forms/affiliate.php and then use account application Will be changing to web maintenance. 9/19/2018 http://www.chpc.utah.edu

Security Policies No clear text passwords, use ssh and scp You may not share your account under any circumstances Don’t leave your terminal unattended while logged into your account Do not introduce classified or sensitive work onto CHPC systems unless on Protected Environment Do not try to break passwords, tamper with files etc. Do not distribute or copy privileged data or software Report suspicions to CHPC (security@chpc.utah.edu) See http://www.chpc.utah.edu/docs/policies/security.html for more details 9/19/2018 http://www.chpc.utah.edu

Allocation General Allocation Process Information http://www.chpc.utah.edu 9/19/2018 Allocation General Allocation Process Information https://wiki.chpc.utah.edu/display/policy/1.4+Allocation+Policies Regular allocation form https://www.chpc.utah.edu/apps/profile/allocation_form.php Requests due Sept 1, Dec 1, Mar 1, and Jun 1 Allocation in SU 1 core-hour = 1SU on ember & 1 core-hour = 1.5 SU on kingspeak Quick allocation https://www.chpc.utah.edu/apps/profile/allocation_quick_form.php Check usage -- http://www.chpc.utah.edu/docs/allocations/ Will be changing to web maintenance. 9/19/2018 http://www.chpc.utah.edu

Accessing Clusters Login or interactive nodes with each cluster http://www.chpc.utah.edu 9/19/2018 Accessing Clusters Login or interactive nodes with each cluster ssh –X kingspeak.chpc.utah.edu ssh –X ember.chpc.utah.edu ssh –X lonepeak.chpc.utah.edu ssh –X telluride.chpc.utah.edu ssh –X ash-guest.chpc.utah.edu Interactive nodes only used for short compiles, editing and very short test runs No more than 15 minutes and no jobs of any length that make heavy use of cpu or memory! Will be changing to web maintenance. 9/19/2018 http://www.chpc.utah.edu

SSH clients for windows http://www.chpc.utah.edu 9/19/2018 SSH clients for windows PuTTY http://www.chiark.greenend.org.uk/~sgtatham/putty/ XShell4 http://www.netsarang.com/download/down_xsh.html Will be changing to web maintenance. 9/19/2018 http://www.chpc.utah.edu

FastX – Tool for Remote X http://www.chpc.utah.edu 9/19/2018 FastX – Tool for Remote X https://www.starnet.com/fastx  Used to interact with remote linux systems graphically in much more efficient and effective way then simple X forwarding Graphical sessions can be detached from without being closing,  allowing users to  reattach to the session from the same or other systems Server on all interactive nodes as well as several dedicated fastx servers Clients for windows, mac and linux; can be installed on both university and personal desktops. Will be changing to web maintenance. 9/19/2018 http://www.chpc.utah.edu

http://www.chpc.utah.edu 9/19/2018 FastX For FastX – see “To Use” section of documentation at https://wiki.chpc.utah.edu/display/DOCS/FastX Download client from CHPC site at https://www.chpc.utah.edu/apps/profile/software.php Do install Start program Set host to kingspeak1.chpc.utah.edu OR kingspeak2.chpc.utah.edu Have dedicated fastX servers – frisco1-frisco6 Frisco5,6 for vgl applications Will be changing to web maintenance. 9/19/2018 http://www.chpc.utah.edu

http://www.chpc.utah.edu 9/19/2018 Login scripts CHPC provides login scripts (“dot” files) when creating account for both tcsh and bash shells These files set the environment so that applications are found, batch commands work – Do not remove or edit! Shell set to tcsh – can change to bash at www.chpc.utah.edu (choose PROFILE, login, select edit profile) Will occasionally request that users update these files At end ‘#source .aliases’ – use this for customization Login scripts (see with ls –a) .tcshrc. .bashrc, .bash_profile Will be changing to web maintenance. 9/19/2018 http://www.chpc.utah.edu

Batch System Information http://www.chpc.utah.edu 9/19/2018 Batch System Information Used to access compute nodes http://docs.adaptivecomputing.com/ Two components Torque (OpenPBS) -- Resource Manager Moab (Maui) – Scheduler – enforces policies and sets job priorities Will be changing to web maintenance. 9/19/2018 http://www.chpc.utah.edu

Some Basic Batch commands http://www.chpc.utah.edu 9/19/2018 Some Basic Batch commands qsub <script> -- submit job qstat -- lists all jobs in queue qdel $jobid -- deletes job pbsnodes –a -- lists all nodes in cluster showq -- lists all jobs in queue {-r for running, -i idle, -b blocked} showstart $jobid -- shows estimated start time checkjob $jobid -- give more detailed info about job Will be changing to web maintenance. 9/19/2018 http://www.chpc.utah.edu

http://www.chpc.utah.edu 9/19/2018 Batch Policies https://wiki.chpc.utah.edu/pages/viewpage.action?pageId=281706582 https://wiki.chpc.utah.edu/display/policy/2.1.3+Kingspeak+Job+Scheduling+Policy Walltime limit – 72 hours Accounts piname for general allocation piname-{kp,em} for owner nodes on kingspeak or ember owner-guest for guest jobs on owner nodes on kingspeak and ember smithp-guest for guest jobs on ash & cheatham-guest for guest jobs on telluride Will be changing to web maintenance. 9/19/2018 http://www.chpc.utah.edu

http://www.chpc.utah.edu 9/19/2018 Sample Batch Script #PBS -S /bin/tcsh #PBS -l nodes=2:ppn=20,walltime=1:00:00 #PBS –A account #PBS -N myjob # Create scratch directory mkdir -p /scratch/kingspeak/serial/$USER/$PBS_JOBID # Change to working directory cd /scratch/kingspeak/serial/$USER/$PBS_JOBID # Copy data files to scratch directory cp $HOME/work_dir/files /scratch/kingspeak/serial/$USER/$PBS_JOBID #Set Environment source /uufs/kingspeak.peaks/sys/pkg/mvapich2/1.9i/etc/mvapich2.sh #Execute Job mpiexec -np 40 -machinefile $PBS_NODEFILE ./executable # Copy files back home and cleanup cp * $HOME/work_dir && rm –rf /scratch/kingspeak/serial/$USER/$PBS_JOBID One time you must create the /scratch/serial/$USER directory. 9/19/2018 http://www.chpc.utah.edu

Getting Help CHPC website and wiki Jira Ticketing System http://www.chpc.utah.edu 9/19/2018 Getting Help CHPC website and wiki www.chpc.utah.edu and wiki.chpc.utah.edu Getting started guide, cluster usage guides, software manual pages, CHPC policies Jira Ticketing System Email: issues@chpc.utah.edu Help Desk: 405 INSCC, 581-6440 (9-5 M-F) We use chpc-hpc-users@lists.utah.edu for sending messages to users; also have Twitter accounts for announcements -- @CHPCOutages & @CHPCUpdates Will be changing to web maintenance. 9/19/2018 http://www.chpc.utah.edu

CHPC Fall Presentation Series http://www.chpc.utah.edu 9/19/2018 CHPC Fall Presentation Series All Presentations in INSCC Auditorium unless noted September 9, 2014 1-2 pm Overview of CHPC September 11, 2014, 2-3pm Protected Environment at CHPC **BMI Green Conference Room (421 Wakara Way, room 2002W) September 16, 2014, 1-3pm Introductory Linux for HPC Part 1 September 18, 2014, 1-3pm Introductory Linux for HPC Part 2 September 23, 2014, 1-2pm XSEDE Resource Support at CHPC September 25, 2014, 1-2pm Introduction to Scientific Visualization September 25, 2014, 2-3pm NLP and AI Services at CHPC **BMI Classroom (421 Wakara Way Room 1470) September 25, 2014, 1-2pm Scientific Visualization September 30, 2014, 1-2pm Chemistry Packages at CHPC

http://www.chpc.utah.edu 9/19/2018 October 2, 2014, 1-3pm Hands-on Introduction to Python, Part 1 October 7, 2014, 1-3pm Hands-on Introduction to Python, Part 2 October 9, 2014, 1-3pm Hands-on Introduction to Numpy & Scipy October 21, 2014, 1-2pm Introduction to Parallel Computing October 23, 2014, 1-2pm Introduction to Programming with MPI October 30, 2014, 1-2pm Introduction to Programming with OpenMP November 4, 2014, 1-2pm Hybrid MPI-OpenMP Programming November 6, 2014, 1-2pm Introduction to I/O at CHPC November 11, 2014, 1-2pm Introduction to Debugging November 13, 2014, 1-2pm Introduction to Profiling November 18, 2014, 1-2pm Introduction to GPU Programming http://www.chpc.utah.edu/docs/presentations/ If you would like training for yourself or your group, CHPC staff would be happy to accommodate your request. Please contact anita.orendt@utah.edu

Advanced Cyberinfrastructure – Research and Education Facilitator (ACI-REF)

Project Vision Advancing scientific discovery through a national network of Advanced Cyberinfrastructure (ACI) Research and Education Facilitators (ACI-REFs). For more information, please visit http://aciref.org

Project Goals Enable scientific and research outcomes that would not be possible without the project and ACI-REF community Create a community that fosters mentorship and sharing of expertise at individual (ACI-REF to researcher) and campus (institution to institution) levels to help others advance their capabilities and outcomes Develop a framework for sharing best practices and lessons learned for facilitation of interactions with researchers that can be deployed at other campuses

Project Goals Increase the breadth and depth of interactions with researchers on campuses in various disciplines Enable researchers to be more competitive for external research funding and to foster new collaborations among researchers, campuses, and other organizations CIO-level commitment from partner campuses to ensure long-term sustainability Develop a plan for enabling other institutions to join the consortium