“High-performance computational GPU-stand for teaching undergraduate and graduate students the basics of quantum-mechanical calculations“ “Komsomolsk-on-Amur.

Slides:



Advertisements
Similar presentations
Optimization on Kepler Zehuan Wang
Advertisements

Molecular dynamics modeling of thermal and mechanical properties Alejandro Strachan School of Materials Engineering Purdue University
Photoreflectance of Semiconductors Tyler A. Niebuhr.
Timothy Blattner and Shujia Zhou May 18, This project is sponsored by Lockheed Martin We would like to thank Joseph Swartz, Sara Hritz, Michael.
Intro to GPU’s for Parallel Computing. Goals for Rest of Course Learn how to program massively parallel processors and achieve – high performance – functionality.
Software Demonstration and Course Description P.N. Lai.
HPCC Mid-Morning Break High Performance Computing on a GPU cluster Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
AMBER. AMBER 7 What is AMBER? –A collective name for a suite of programs that allow users to carry out molecular dynamic simulations. –And a set of molecular.
Case Studies Class 5. Computational Chemistry Structure of molecules and their reactivities Two major areas –molecular mechanics –electronic structure.
MOLDEN: a pre- and post- processing program
Computing lattice constant, bulk modulus and equilibrium energies of solids Bulk Si Diamond structure.
11 Desktop Grids for International Scientific Collaboration International Desktop Grid Federation APPLICATION OF DESKTOP GRID TECHNOLOGY IN MATERIAL SCIENCE.
Large-Scale Density Functional Calculations James E. Raynolds, College of Nanoscale Science and Engineering Lenore R. Mullin, College of Computing and.
Using the Trigger Test Stand at CDF for Benchmarking CPU (and eventually GPU) Performance Wesley Ketchum (University of Chicago)
Debunking the 100X GPU vs. CPU Myth: An Evaluation of Throughput Computing on CPU and GPU Presented by: Ahmad Lashgar ECE Department, University of Tehran.
HPCC Mid-Morning Break Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery Introduction to the new GPU (GFX) cluster.
2012/06/22 Contents  GPU (Graphic Processing Unit)  CUDA Programming  Target: Clustering with Kmeans  How to use.
Lectures Introduction to computational modelling and statistics1 Potential models2 Density Functional.
National Center for Supercomputing Applications University of Illinois at Urbana-Champaign Direct Self-Consistent Field Computations on GPU Clusters Guochun.
Training Program on GPU Programming with CUDA 31 st July, 7 th Aug, 14 th Aug 2011 CUDA Teaching UoM.
Processing of a CAD/CAE Jobs in grid environment using Elmer Electronics Group, Physics Department, Faculty of Science, Ain Shams University, Mohamed Hussein.
Chapter 2 Computer Clusters Lecture 2.3 GPU Clusters for Massive Paralelism.
Shared memory systems. What is a shared memory system Single memory space accessible to the programmer Processor communicate through the network to the.
Thermoacoustics in random fibrous materials Seminar Carl Jensen Tuesday, March
Computationally Efficient Histopathological Image Analysis: Use of GPUs for Classification of Stromal Development Olcay Sertel 1,2, Antonio Ruiz 3, Umit.
GPU Performance Prediction GreenLight Education & Outreach Summer Workshop UCSD. La Jolla, California. July 1 – 2, Javier Delgado Gabriel Gazolla.
Computer Graphics Graphics Hardware
1 Chapter 04 Authors: John Hennessy & David Patterson.
1 MacSpartan A Tutorial on intended as a general reference to use of the MacSpartan. This tutorial provides a brief overview describes in detail the various.
MS Thesis Defense “IMPROVING GPU PERFORMANCE BY REGROUPING CPU-MEMORY DATA” by Deepthi Gummadi CoE EECS Department April 21, 2014.
Accelerating MATLAB with CUDA
Network for Computational Nanotechnology (NCN) Purdue, Norfolk State, Northwestern, MIT, Molecular Foundry, UC Berkeley, Univ. of Illinois, UTEP Multi.
Molecular Dynamics Simulations of Diffusion in Polymers Zach Eldridge Department of Mechanical Engineering University of Arkansas Fayetteville, AR
Computación algebraica dispersa con GPUs y su aplicación en tomografía electrónica Non-linear iterative optimization method for locating particles using.
YOU LI SUPERVISOR: DR. CHU XIAOWEN CO-SUPERVISOR: PROF. LIU JIMING THURSDAY, MARCH 11, 2010 Speeding up k-Means by GPUs 1.
Accelerating Statistical Static Timing Analysis Using Graphics Processing Units Kanupriya Gulati and Sunil P. Khatri Department of ECE, Texas A&M University,
Fast Support Vector Machine Training and Classification on Graphics Processors Bryan Catanzaro Narayanan Sundaram Kurt Keutzer Parallel Computing Laboratory,
EEW508 Structure of Surfaces Surface structure Rice terrace.
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
Geant4 based simulation of radiotherapy in CUDA
Computational Aspects of Multi-scale Modeling Ahmed Sameh, Ananth Grama Computing Research Institute Purdue University.
Jie Chen. 30 Multi-Processors each contains 8 cores at 1.4 GHz 4GB GDDR3 memory offers ~100GB/s memory bandwidth.
QCAdesigner – CUDA HPPS project
GPU-based Computing. Tesla C870 GPU 8 KB / multiprocessor 1.5 GB per GPU 16 KB up to 768 threads () up to 768 threads ( 21 bytes of shared memory and.
1)Leverage raw computational power of GPU  Magnitude performance gains possible.
Physics “Advanced Electronic Structure” Lecture 1. Theoretical Background Contents: 1. Historical Overview. 2. Basic Equations for Interacting Electrons.
1 Workshop 9: General purpose computing using GPUs: Developing a hands-on undergraduate course on CUDA programming SIGCSE The 42 nd ACM Technical.
GPU Programming Shirley Moore CPS 5401 Fall 2013
CUDA Basics. Overview What is CUDA? Data Parallelism Host-Device model Thread execution Matrix-multiplication.
Sunpyo Hong, Hyesoon Kim
S. Pardi Frascati, 2012 March GPGPU Evaluation – First experiences in Napoli Silvio Pardi.
Rappture GUI for Carbon Nano Tube Arrays’ mechanical and thermal property simulation By Yide Wang Professor Tim Fisher Sridhar Sadasivam.
Multicore Applications in Physics and Biochemical Research Hristo Iliev Faculty of Physics Sofia University “St. Kliment Ohridski” 3 rd Balkan Conference.
Grid computing simulation of superconducting vortex lattice in superconducting magnetic nanostructures M. Rodríguez-Pascual 1, D. Pérez de Lara 2, E.M.
Computing lattice constant, bulk modulus and equilibrium energies of bulk cubic SrTiO 3 Bulk SrTiO 3 Cubic structure.
GPAW Setup Optimization Center for Atomic-scale Materials Design Technical University of Denmark Ask Hjorth Larsen.
Molecular Dynamics Analysis Toolkit Karl Debiec and Nick Rego Chong Group Department of Chemistry August 30 th 2013.
Computer Graphics Graphics Hardware
GPGPU use cases from the MoBrain community
BY SELLAVEL E (CA15M006) Guided By Prof.B.Viswanathan
Heterogeneous Computation Team HybriLIT
ECE408 Fall 2015 Applied Parallel Programming Lecture 21: Application Case Study – Molecular Dynamics.
ECE408 / CS483 Applied Parallel Programming Lecture 23: Application Case Study – Electrostatic Potential Calculation.
Atomistic simulations of contact physics Alejandro Strachan Materials Engineering PRISM, Fall 2007.
Atomistic materials simulations at The DoE NNSA/PSAAP PRISM Center
© 2012 Elsevier, Inc. All rights reserved.
Computer Graphics Graphics Hardware
CSE 502: Computer Architecture
Multicore and GPU Programming
Presentation transcript:

“High-performance computational GPU-stand for teaching undergraduate and graduate students the basics of quantum-mechanical calculations“ “Komsomolsk-on-Amur State Technical University” , Russia, Komsomolsk-on-Amur city, Lenina Singapore-2013 Sergey Seriy

Overview  Teaching of classical and DFT “ab-initio” calculations( samples of bulk, slab, particle structures, thin-film, alloys, etc)  Using Atomistic Simulation Environment (ASE) and GPU-based software for calculations  ASE universality: electronic structure codes + LAMMPS or ABINIT or CP2K or GPAW, with GPU- support  ASE exercises: tests of defect energies, heats of formation, elastic constants, etc.  Conclusion

Main Goal: “Teaching for calculate basic structural properties of single crystals, formation energies of defects, and structural and elastic properties of simple nanostructures, using GPU (NVidia, ATI and both)” Need to learn formats of input parameter files, atomic conguration files, and output files of ab-initio code (CP2K, Abinit, GPAW) classical code (LAMMPS) Need to create atomic congurations for single crystal structures, crystallic compounds, point defects (vacancies, interstitials, substitutions), planar defects (varying surfaces, stacking faults), and strained structures

Need a tool applicable to quickly evaluate basic properties from classical potentials and ab-initio methods. Ideally, a single universal tool would be able to create basic atomic congurations and manipulate them serve these atomistic congurations as inputs to a variety of methods/simulation codes and obtain energies Anything like that available? Atomistic Simulation Environment (ASE)

Atomistic Simulation Environment (ASE) v. 3.0 universal Python interface to many DFT codes (calculators), with visualization, simple GUI, documentation, and tutorials creates molecules, crystal structures, surfaces, nanotubes, analyzes symmetry and spacegroups provides support for Equation of state, structure optimization,dissociation, diusion, constrains, NEB, vibration analysis, phonon calculations, infrared intensities, molecular dynamics, STM, electron transport, …

-- and CP2K – DFT-based molecular dynamics code (consist in old version ASE 2.0)

Conclusions  ASE provides a universal interface to many electronic- structure codes  ASE interface for LAMMPS, ABINIT and CP2K on GPU was utilized in learning students  Following the LAMMPS example, ASE can provide support to other classical and “ab-initio” codes  ASE simplies and increases eciency of atomistic simulation learning

16

17

18

Multicore, CPU & GPU SpecificationsCore i7 960GTX285 Processing Elements 4 cores, 4 way GHz 30 cores, 8 way GHz Resident Strands/Threads (max) 4 cores, 2 threads, 4 way SIMD: 32 strands 30 cores, 32 SIMD vectors, 32 way SIMD: threads SP GFLOP/s Memory Bandwidth25.6 GB/s159 GB/s Register File MB Local Store-480 kB Core i7 (45nm) GTX285 (55nm)

20

21

22

23

24

25

26

Mathematical modeling: nano- coatings for cutting tools

VMD – “Visual Molecular Dynamics”  Visualization and analysis of molecular dynamics simulations, sequence data, volumetric data, quantum chemistry simulations, particle systems, …  User extensible with scripting and plugins

Molecular orbital calculation and display: factor of 120x faster Imaging of gas migration pathways in proteins with implicit ligand sampling: factor of 20x to 30x faster GPU Acceleration in VMD Electrostatic field calculation, ion placement: factor of 20x to 44x faster

Mesoscale modelling on CUDA: Fluid Dynamics Double precision 384 x 384 x 192 grid (max that fits in 4GB) Vertical slice of temperature at y=0 Transition from stratified (left) to turbulent (right) Regime depends on Rayleigh number: Ra = gαΔT/κν 8.5x speedup versus Fortran code running on 8- core 2.5 GHz Xeon

Thank you for your attention