Computational chemistry packages (efficient usage issues?) Jemmy Hu SHARCNET HPC Consultant Summer School June 3, 2016 /work/jemmyhu/ss2016/chemistry/

Slides:



Advertisements
Similar presentations
MCTS GUIDE TO MICROSOFT WINDOWS 7 Chapter 10 Performance Tuning.
Advertisements

Southgreen HPC system Concepts Cluster : compute farm i.e. a collection of compute servers that can be shared and accessed through a single “portal”
Chapter 14 Chapter 14: Server Monitoring and Optimization.
1.1 Installing Windows Server 2008 Windows Server 2008 Editions Windows Server 2008 Installation Requirements X64 Installation Considerations Preparing.
Asynchronous Solution Appendix Eleven. Training Manual Asynchronous Solution August 26, 2005 Inventory # A11-2 Chapter Overview In this chapter,
Installing and running COMSOL on a Windows HPCS2008(R2) cluster
Reproducible Environment for Scientific Applications (Lab session) Tak-Lon (Stephen) Wu.
Lecture 8 Configuring a Printer-using Magic Filter Introduction to IP Addressing.
Dzongkha Unicode Agenda System Requirements: Hardware & Software
Understanding the Basics of Computational Informatics Summer School, Hungary, Szeged Methos L. Müller.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment, Enhanced Chapter 12: Managing and Implementing Backups and Disaster Recovery.
A crash course in njit’s Afs
Linux Shell. 2 Linux Command-Line Interface ■ Linux shells: A shell is a command interpreter that allows you to type commands from the keyboard to interact.
Introduction to UNIX/Linux Exercises Dan Stanzione.
EUROPEAN UNION Polish Infrastructure for Supporting Computational Science in the European Research Space Ab initio grid chemical software ports – transferring.
Task Farming on HPCx David Henty HPCx Applications Support
Introduction to HP LoadRunner Getting Familiar with LoadRunner >>>>>>>>>>>>>>>>>>>>>>
 Accessing the NCCS Systems  Setting your Initial System Environment  Moving Data onto the NCCS Systems  Storing Data on the NCCS Systems  Running.
Tutorial 11 Installing, Updating, and Configuring Software
MCTS Guide to Microsoft Windows 7
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
The Pipeline Processing Framework LSST Applications Meeting IPAC Feb. 19, 2008 Raymond Plante National Center for Supercomputing Applications.
VIPBG LINUX CLUSTER By Helen Wang March 29th, 2013.
HPC at HCC Jun Wang Outline of Workshop1 Overview of HPC Computing Resources at HCC How to obtain an account at HCC How to login a Linux cluster at HCC.
Please Note: Information contained in this document is considered LENOVO CONFIDENTIAL For Lenovo Internal Use Only Do Not Copy or Distribute!! For Lenovo.
Compiled Matlab on Condor: a recipe 30 th October 2007 Clare Giacomantonio.
Installation Overview Lab#2 1Hanin Abdulrahman. Installing Ubuntu Linux is the process of copying operating system files from a CD, DVD, or USB flash.
AE6382 Secure Shell Usually referred to as ssh, the name refers to both a program and a protocol. The program ssh is one of the most useful networking.
Research Computing Environment at the University of Alberta Diego Novillo Research Computing Support Group University of Alberta April 1999.
Module 1: Installing Microsoft Windows XP Professional.
HPC for Statistics Grad Students. A Cluster Not just a bunch of computers Linked CPUs managed by queuing software – Cluster – Node – CPU.
© Janice Regan, CMPT 300, May CMPT 300 Introduction to Operating Systems Memory: Relocation.
MIPS Project -- Simics Yang Diyi Outline Introduction to Simics Simics Installation – Linux – Windows Guide to Labs – General idea Score Policy.
Faucets Queuing System Presented by, Sameer Kumar.
1 Computer Systems II Introduction to Processes. 2 First Two Major Computer System Evolution Steps Led to the idea of multiprogramming (multiple concurrent.
Cluster Computing Applications for Bioinformatics Thurs., Sept. 20, 2007 process management shell scripting Sun Grid Engine running parallel programs.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Page 1 of 38 Lenovo Confidential Lenovo Confidential Lenovo Confidential Lenovo Confidential Lenovo Confidential Please Note: Information contained in.
OPTIMIZATION OF DIESEL INJECTION USING GRID COMPUTING Miguel Caballer Universidad Politécnica de Valencia.
Portable Batch System – Definition and 3 Primary Roles Definition: PBS is a distributed workload management system. It handles the management and monitoring.
CCNA1 v3 Module 1 v3 CCNA 1 Module 1 JEOPARDY K. Martin.
Advanced topics Cluster Training Center for Simulation and Modeling September 4, 2015.
CCJ introduction RIKEN Nishina Center Kohei Shoji.
Debugging Lab Antonio Gómez-Iglesias Texas Advanced Computing Center.
Wouter Verkerke, NIKHEF 1 Using ‘stoomboot’ for NIKHEF-ATLAS batch computing What is ‘stoomboot’ – Hardware –16 machines, each 2x quad-core Pentium = 128.
Active-HDL Server Farm Course 11. All materials updated on: September 30, 2004 Outline 1.Introduction 2.Advantages 3.Requirements 4.Installation 5.Architecture.
NREL is a national laboratory of the U.S. Department of Energy, Office of Energy Efficiency and Renewable Energy, operated by the Alliance for Sustainable.
CACI Proprietary Information | Date 1 PD² v4.2 Increment 2 SR13 and FPDS Engine v3.5 Database Upgrade Name: Semarria Rosemond Title: Systems Analyst, Lead.
Advanced Computing Facility Introduction
Gaussian 09 Tutorial Ph. D. Candidate
1.1.2 OneOs Downloading Software Upgrade
GRID COMPUTING.
Welcome to Indiana University Clusters
Assumptions What are the prerequisites? … The hands on portion of the workshop will be on the command-line. If you are not familiar with the command.
HPC usage and software packages
OpenPBS – Distributed Workload Management System
Welcome to Indiana University Clusters
How to use the HPCC to do stuff
BIMSB Bioinformatics Coordination
Practice #0: Introduction
Bruce Pullig Solution Architect
College of Engineering
Haiyan Meng and Douglas Thain
CCR Advanced Seminar: Running CPLEX Computations on the ISE Cluster
High Performance Computing in Bioinformatics
Michael P. McCumber Task Force Meeting April 3, 2006
Introduction to High Performance Computing Using Sapelo2 at GACRC
Quick Tutorial on MPICH for NIC-Cluster
Working in The IITJ HPC System
Kajornsak Piyoungkorn,
Presentation transcript:

Computational chemistry packages (efficient usage issues?) Jemmy Hu SHARCNET HPC Consultant Summer School June 3, 2016 /work/jemmyhu/ss2016/chemistry/

Chemistry Software managed by SHARCNET Electronic Structure TheoryMolecular Dynamics Gaussian (09, 03) -commercialAMBER -commercial ADF/Band -commercialGromacs NWChemCPMD Quantum ExpressoNAMD ORCA Gamess/Firefly Abinit MolproVASP TurbomoleCHARMM User/group owned software (commercial)

Nature of the software SoftwareParallelPurpose GaussianThreaded (OpenMP) ADFMPI NWChemMPI sw.org/index.php/Main_Page Quantum ExpressoMPI VASPMPI vasp/59-about-vasp ORCAMPI

Open Software used by 2 or more groups – add to SHARCNET software stack with software usage instruction/example webpage online Software used by one group – user installs in his/her own space with help from SHARCNET if necessary SHARCNET owned commercial software – free Self owned software – user’s response for license issues Software are installed as module package, setup in your bash configuration file, i.e.,.bash_profile file under /home/userid/ module load gaussian/g09_D.01 General policy for software

module load packages Login to a cluster, e.g., ssh Open the shell configuration file, e.g., vi.bash_profile Add the module load lines at the end of the file module load gaussian/g09_D.01 module load adf/ module load nwchem/6.5 module load espresso/5.1.1 module load orca/3.0 Save the file Exit Re-login Goto your /work directory: cd /work/userID Copy the files cp –rf /work/jemmyhu/ss2016/chemistry.

System Info Parallel Computer Memory Architectures Hybrid Distributed-Shared Memory (Cluster solution)

Size of Threaded Jobs on specific system SystemNodesCPU/Node (threads) Memory GB (Max for full cpus usage) orcaOpteron16 or 2430 sawXeon815 brownXeon816 Job uses less than the max threads and memory on a node will be preferable

Gaussian 09 issues Latest version: g09-D.01 Input sample File location Memory Disk Restart Utilities

#p b3lyp/ g(2df,p) opt freq > 0 1 c o 1 co co= (leave an blank line at the end) G09 input file: simplest case File name: CO-b3lyp-1.com

Real jobs should be submitted to the threaded queue using sqsub command First check the files under your /work and /scratch directory open two windows, one for /work, one to view files in /scratch For testing purpose today, we run jobs from command line on the dev. node on saw (saw-dev1, …, saw-dev6) ssh saw-dev4 run g09 CO-b3lyp-1.com >& CO-b3lyp-1.log & Output file CO-b3lyp-1.log Runtime files (5 of them,.rwf is big) /scratch/userID Job uses defaults [1 cpu, minimum memory (256MB?)] Only.log output file will be generated in the run directory Runtime files will be removed automatically once finished successfully, otherwise they will stay. Run job, file structure

Specify checkpoint file: name.chk %chk=CO-b3lyp-1chk #p b3lyp/ g(2df,p) opt freq > 0 1 c o 1 co co= Input file: CO-b3lyp-1chk.com Run on dev: g09 CO-b3lyp-1chk.com >& CO-b3lyp-1chk.log & Files: CO-b3lyp-1chk.log, CO-b3lyp-1chk.chk

G09 input file: CO-b3lyp-4.com, use 4 cpus %nosave %mem=500MB %chk=CO-b3lyp-4 %nproc=4 #p b3lyp/ g(2df,p) opt freq > 0 1 c o 1 co co=1.1281

Run Job in the queue %mem=500MB (--mpp=2x%mem, memory per job, for all cpus) %nproc=4 (-n 4) Submit command is sqsub -q threaded -n 4 -r 1.0d –mpp=2g -o CO out g09 CO com What’s happening afterwards? -sqjobs to see job status, jobID -If runs, where are the files?.log and.chk files are generated in the run directory runtime files,.rwf, etc will be in /scratch/yourID/jobID/ -Failed? remove the runtime files, rm –rf /scratch/yourID/jobID. correct input file, re-run

%rwf=CO-b3lyp-4rwf %nosave %mem=500MB %chk=CO-b3lyp-4rwf %nproc=4 #p b3lyp/ g(2df,p) opt freq > 0 1 c o 1 co co= The named.rwf file will be generated in the current directory Input file with named.rwf: CO-b3lyp-4rwf.com

Checkpointing 1..chk can be used to restart an Optimization job, or used for property (freq, etc) calculation afterwards. 2. In G09,.rwf can be used to restart a frequency calculation.

Restart (see examples) Restart opt job from.chk Restart freq from.rwf %rwf=/scratch/jemmyhu/Cu_test.rwf %NoSave %chk=Cu_test.chk %mem=500mb %nproc=4 #p restart (one blank line) %NoSave %chk=Cu_test.chk %mem=500mb %nproc=4 #B3LYP/DGDZVP Opt Geom=AllCheck Guess=Read (one blank line)

Gaussian Utilities Utilities formchk cubegen freqchk …… use the same version direct run from command line NBO

Review Input file, memory (%mem) /cpu (%nproc) sqsub (-n, --mpp), sqjobs, sqkill Use job info from scheduler File location Failed jobs File cleanup

ADF ADF is frequently used for studying transition metal complexes and molecules with heavy atoms. ADF offers unique capabilities to predict molecular properties of organic electronics materials.transition metal complexesheavy atomsorganic electronics ADF, BAND, DFTB

Use ADF on SHARCNET Available on major clusters, viz-stations Versions: , Setup, add module load line to.bash_profile file, i.e., module load adf/ Examples /opt/sharcnet/adf/ /examples

module show adf/ /opt/sharcnet/modules/adf/ : module-whatis Provides ADF binary for x86_64 Centos 6 systems conflict adf setenv ADFROOT /opt/sharcnet/adf setenv ADFHOME /opt/sharcnet/adf/ setenv ADFBIN /opt/sharcnet/adf/ /bin setenv ADFRESOURCES /opt/sharcnet/adf/ /atomicdata setenv SCMLICENSE /opt/sharcnet/adf/ /license setenv SCM_TMPDIR /scratch/jemmyhu setenv SCM_USETMPDIR yes prepend-path PATH /opt/sharcnet/adf/ /bin Default settings

Two input options name.inp (one step a time) use a runscript ‘ADF’ or ‘BAND’, to bypass the runtime files to /scratch/username/jobid/ sqsub -q mpi --nompirun -n 8 --mpp=2g -r 1.0d -o GO_H2O-adf.log ADF GO_H2O-adf.inp sqsub -q mpi --nompirun -n 8 --ppn=4 --mpp=2g -r 1.0d -o GO_H2O-adf.log ADF GO_H2O-adf.inp sqsub -q mpi --nompirun -n 8 --mpp=2g -r 1.0d -o name_band.log BAND name_band.inp

/home/jemmyhu/work/ss2016/chemistry/examples/adf GO_H2O-adf.inp

sqsub -q mpi --nompirun -n 8 --ppn=4 --mpp=2g -r 1.0d -o NMR_B3LYP.log./NMR_B3LYP.run Input for multi-steps: name.run runtime files are under /scratch/username/

Restart TAPE13 It serves as a checkpoint file. TAPE13 can be used to restart the calculation at a point not too far before the fatal condition occurred TAPE13 is upgraded during the calculation but discarded upon normal termination. To keep TAPE13, -use the SAVE key TAPE21 Upon normal (i.e. program-controlled) termination of a calculation, the TAPE21 result file can be used for restart purposes When a crash occurs, however, chances are that TAPE21 has not correctly been closed and that its data structure is inconsistent

ADF GUI GUI software to run on viz-workstations vdi-centos6.user.sharcnet.ca viz-uwo10.sharcnet.ca viz-uwo11.sharcnet.ca Turn on X11 locally, then login to he workstation adfview

pwd /home/jemmyhu/work/ss2016/chemistry/examples/nwchem ls h2o_hu.nw sqsub -q mpi -n 4 -r 3.0d --mpp=4g -o h2o_hu.log nwchem h2o_hu.nw RUN nwchem on SHARCNET

Research group owns the license Installation guideline

What else? Question? Discussion?