ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.

Slides:



Advertisements
Similar presentations
including File Management
Advertisements

Overview of Wisconsin Campus Grid Dan Bradley Center for High-Throughput Computing.
Blackbird: Accelerated Course Archives Using Condor with Blackboard Sam Hoover, IT Systems Architect Matt Garrett, System Administrator.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Southgreen HPC system Concepts Cluster : compute farm i.e. a collection of compute servers that can be shared and accessed through a single “portal”
ISG We build general capability Job Submission on the Olympus Cluster J. DePasse; S. Brown, PhD; T. Maiden Pittsburgh Supercomputing Center Public Health.
Jia Yao Director: Vishwani D. Agrawal High Performance Compute Cluster April 13,
Information Technology Center Introduction to High Performance Computing at KFUPM.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
Efficiently Sharing Common Data HTCondor Week 2015 Zach Miller Center for High Throughput Computing Department of Computer Sciences.
Introduction to z/OS Basics © 2006 IBM Corporation Chapter 7: Batch processing and the Job Entry Subsystem (JES) Batch processing and JES.
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
 Contents 1.Introduction about operating system. 2. What is 32 bit and 64 bit operating system. 3. File systems. 4. Minimum requirement for Windows 7.
ISG We build general capability Purpose After this tutorial, you should: Be comfortable submitting work to the batch queuing system of olympus and be familiar.
Utilizing Condor and HTC to address archiving online courses at Clemson on a weekly basis Sam Hoover 1 Project Blackbird Computing,
JGI/NERSC New Hardware Training Kirsten Fagnan, Seung-Jin Sul January 10, 2013.
Introduction to the new mainframe © Copyright IBM Corp., All rights reserved. Chapter 5: Batch processing and the Job Entry Subsystem (JES) Batch.
Introduction to UNIX/Linux Exercises Dan Stanzione.
Types of Operating System
Introduction to HP LoadRunner Getting Familiar with LoadRunner >>>>>>>>>>>>>>>>>>>>>>
CHAPTER 2 OPERATING SYSTEM OVERVIEW 1. Operating System Operating System Definition A program that controls the execution of application programs and.
Introduction to HPC resources for BCB 660 Nirav Merchant
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Introduction to the HPCC Jim Leikert System Administrator High Performance Computing Center.
MSc. Miriel Martín Mesa, DIC, UCLV. The idea Installing a High Performance Cluster in the UCLV, using professional servers with open source operating.
VIPBG LINUX CLUSTER By Helen Wang March 29th, 2013.
Bigben Pittsburgh Supercomputing Center J. Ray Scott
Introduction to the HPCC Dirk Colbry Research Specialist Institute for Cyber Enabled Research.
MaterialsHub - A hub for computational materials science and tools.  MaterialsHub aims to provide an online platform for computational materials science.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
17-April-2007 High Performance Computing Basics April 17, 2007 Dr. David J. Haglin.
Common Practices for Managing Small HPC Clusters Supercomputing 12
Experiences with a HTCondor pool: Prepare to be underwhelmed C. J. Lingwood, Lancaster University CCB (The Condor Connection Broker) – Dan Bradley
SSS Test Results Scalability, Durability, Anomalies Todd Kordenbrock Technology Consultant Scalable Computing Division Sandia is a multiprogram.
How to get started on cees Mandy SEP Style. Resources Cees-clusters SEP-reserved disk20TB SEP reserved node35 (currently 25) Default max node149 (8 cores.
Using the BYU Supercomputers. Resources Basic Usage After your account is activated: – ssh You will be logged in to an interactive.
Types of computer operation. There a several different methods of operation. Most computers can undertake each of these simultaneously. These methods.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
HPC for Statistics Grad Students. A Cluster Not just a bunch of computers Linked CPUs managed by queuing software – Cluster – Node – CPU.
M. Schott (CERN) Page 1 CERN Group Tutorials CAT Tier-3 Tutorial October 2009.
Weekly Report By: Devin Trejo Week of June 14, 2015-> June 20, 2015.
1 Lattice QCD Clusters Amitoj Singh Fermi National Accelerator Laboratory.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Tier 3 Status at Panjab V. Bhatnagar, S. Gautam India-CMS Meeting, July 20-21, 2007 BARC, Mumbai Centre of Advanced Study in Physics, Panjab University,
Portable Batch System – Definition and 3 Primary Roles Definition: PBS is a distributed workload management system. It handles the management and monitoring.
Microsoft ® Official Course Module 6 Managing Software Distribution and Deployment by Using Packages and Programs.
Computer Performance. Hard Drive - HDD Stores your files, programs, and information. If it gets full, you can’t save any more. Measured in bytes (KB,
Wouter Verkerke, NIKHEF 1 Using ‘stoomboot’ for NIKHEF-ATLAS batch computing What is ‘stoomboot’ – Hardware –16 machines, each 2x quad-core Pentium = 128.
L ECTURE -6 Topics: Introduction to Operating system. Functions of an Operating system. Classification of Operating systems.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
CFI 2004 UW A quick overview with lots of time for Q&A and exploration.
An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.
IPPP Grid Cluster Phil Roffe David Ambrose-Griffith.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
Advanced Computing Facility Introduction
Compute and Storage For the Farm at Jlab
GRID COMPUTING.
Specialized Computing Cluster An Introduction
Welcome to Indiana University Clusters
Welcome to Indiana University Clusters
Architecture & System Overview
MaterialsHub - A hub for computational materials science and tools.
CommLab PC Cluster (Ubuntu OS version)
Processor Management Damian Gordon.
CCR Advanced Seminar: Running CPLEX Computations on the ISE Cluster
High Performance Computing in Bioinformatics
Introduction to High Performance Computing Using Sapelo2 at GACRC
LO2 – Understand Computer Software
Processor Management Damian Gordon.
Presentation transcript:

ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing Center

ISG We build general capability Questions to answer What is Olympus? Why do I need Olympus? How do I and my group get access to Olympus? How do I use Olympus? What will I learn in the hands-on training?

ISG We build general capability RAM Disk Some common terms... Core – a single processing unit capable of performing computations. Multicore Processor – a unit that contains two or more independent cores Socket – place where a multicore processor is plugged in and connects to other resources in the computer Node – collection of sockets and share a common memory space Cluster – a collection of nodes connected through an external network

ISG We build general capability What is Olympus? It is a cluster (1536 cores, 7.5 TB of RAM) 24 nodes (64 cores each) 4 sockets AMD multicore processor 16 cores (1.4 GHz) 512 GB of RAM (8 Nodes) 256 GB of RAM (16 Nodes) Local 3TB hard disk Nodes connected to each other by Gigabit Ethernet Shared file System (2TB) Upgraded Q with ~50 TB file system Many software packages currently available E.g. Python, R, gcc, FRED, FluTE New software can be installed by request

ISG We build general capability Why do I need Olympus? I can do more work, more quickly. More capacity allows for more simulation runs 1 million county level influenza simulations per day 100,000 state level influenza simulations per day I can do larger work. Large memory nodes allow bigger simultaneous simulation runs national level influenza simulations per day I can work more efficiently and effectively. Queuing system allows setting up large batches of runs through scripting rather than manual input creation Support from PSC/ISG staff to help you work more efficiently.

ISG We build general capability How do I get access to Olympus? Can I get an account? Anyone doing MIDAS related research can get an account on Olympus. How? Request an account through the MIDAS Portal I have an account, now what do I do? You can access the machine through ssh at the address olympus.psc.edu. What does an account get me? Home directory for storing files Ability to submit work to the machine Support if you run into trouble and/or need software installed Consultation for working more efficiently and effectively

ISG We build general capability How do I use Olympus? When you log onto the machine, you will be on the head node. A limited machine that allows you to interact with the full cluster. What should I do on the head node? Manage your files Setup work that you plan to submit to the machine How I do I actually run a job on Olympus? Jobs will need to be submitted to a queuing system which will put the work in line for execution when resources are available.

ISG We build general capability ssh How do I use Olympus? What is a queuing system? Software allowing users to submit work on a multiuser machine by scheduling work on machine and running when the resources are available. How do I use the queuing system? Prepare the work you want to do on the head node Create a submission script Defines the work to be accomplished Defines the resources needed to accomplish the work (e.g. how many cores do I need) Submit the work to the queue Check status of the job (queued, running, done) Work will run when the resources are available When finished, collect results on the head node I want to run simulations. I am going to create a whole lot of inputs to run. I need 32 cores, 128 GB of RAM and 2 hours of time. Head Node Olympus Cluster PBS / TORQUE Queuing System WORK Inputs Queued Running Output Wow, I got a lot of work done! I am going to go get coffee. Done

ISG We build general capability What will I learn in the hands- on? Logging on to Olympus How do I create a submission script? How do I submit and monitor work through the queuing system? Further information about the machine to help you effectively and efficiently use the resource.