PuReMD: Purdue Reactive Molecular Dynamics Package Hasan Metin Aktulga and Ananth Grama Purdue University TST Meeting,May 13-14, 2010.

Slides:



Advertisements
Similar presentations
Simulazione di Biomolecole: metodi e applicazioni giorgio colombo
Advertisements

PERFORMANCE EVALUATION OF USER-REAXC PACKAGE Hasan Metin Aktulga Postdoctoral Researcher Scientific Computing Group Lawrence Berkeley National Laboratory.
Presented by Dealing with the Scale Problem Innovative Computing Laboratory MPI Team.
Scalable Multi-Cache Simulation Using GPUs Michael Moeng Sangyeun Cho Rami Melhem University of Pittsburgh.
Navigation Set Hierarchy Tom Gianos Chapter 2.2. Mike Dickheiser Works (worked?) for Red Storm Entertainment Works (worked?) for Red Storm Entertainment.
Formulation of an algorithm to implement Lowe-Andersen thermostat in parallel molecular simulation package, LAMMPS Prathyusha K. R. and P. B. Sunil Kumar.
LINCS: A Linear Constraint Solver for Molecular Simulations Berk Hess, Hemk Bekker, Herman J.C.Berendsen, Johannes G.E.M.Fraaije Journal of Computational.
Ionization of the Hydrogen Molecular Ion by Ultrashort Intense Elliptically Polarized Laser Radiation Ryan DuToit Xiaoxu Guan (Mentor) Klaus Bartschat.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Performance of Applications Using Dual-Rail InfiniBand 3D Torus Network on the.
1 A component mode synthesis method for 3D cell by cell calculation using the mixed dual finite element solver MINOS P. Guérin, A.M. Baudron, J.J. Lautard.
OpenFOAM on a GPU-based Heterogeneous Cluster
PRISM Mid-Year Review Reactive Atomistics, Molecular Dynamics.
Science Advisory Committee Meeting - 20 September 3, 2010 Stanford University 1 04_Parallel Processing Parallel Processing Majid AlMeshari John W. Conklin.
© David Kirk/NVIDIA and Wen-mei W. Hwu, ECE 498AL, University of Illinois, Urbana-Champaign 1 ECE 498AL Programming Massively Parallel Processors.
1 Parallel Simulations of Underground Flow in Porous and Fractured Media H. Mustapha 1,2, A. Beaudoin 1, J. Erhel 1 and J.R. De Dreuzy IRISA – INRIA.
PuReMD: A Reactive (ReaxFF) Molecular Dynamics Package Hassan Metin Akgulta, Sagar Pandit, Joseph Fogarty, Ananth Grama Acknowledgements:
1 Scalable Distributed Fast Multipole Methods Qi Hu, Nail A. Gumerov, Ramani Duraiswami Institute for Advanced Computer Studies Department of Computer.
1 ProActive performance evaluation with NAS benchmarks and optimization of OO SPMD Brian AmedroVladimir Bodnartchouk.
Efficient Parallel Implementation of Molecular Dynamics with Embedded Atom Method on Multi-core Platforms Reporter: Jilin Zhang Authors:Changjun Hu, Yali.
Performance Evaluation of Hybrid MPI/OpenMP Implementation of a Lattice Boltzmann Application on Multicore Systems Department of Computer Science and Engineering,
Parallel Performance of Hierarchical Multipole Algorithms for Inductance Extraction Ananth Grama, Purdue University Vivek Sarin, Texas A&M University Hemant.
Algorithms and Software for Large-Scale Simulation of Reactive Systems _______________________________ Ananth Grama Coordinated Systems Lab Purdue University.
Stratified Magnetohydrodynamics Accelerated Using GPUs:SMAUG.
Designing and Evaluating Parallel Programs Anda Iamnitchi Federated Distributed Systems Fall 2006 Textbook (on line): Designing and Building Parallel Programs.
Martin Berzins (Steve Parker) What are the hard apps problems? How do the solutions get shared? What non-apps work is needed? Thanks to DOE for funding.
Reactive Molecular Dynamics: Progress Report Hassan Metin Aktulga 1, Joseph Fogarty 2, Sagar Pandit 2, and Ananth Grama 3 1 Lawrence Berkeley Lab 2 University.
Computational issues in Carbon nanotube simulation Ashok Srinivasan Department of Computer Science Florida State University.
Algorithmic and Numerical Techniques for Atomistic Modeling Hasan Metin Aktulga PhD Thesis Defense June 23 rd, 2010.
Chicago, July 22-23, 2002DARPA Simbiosys Review 1 Monte Carlo Particle Simulation of Ionic Channels Trudy van der Straaten Umberto Ravaioli Beckman Institute.
Scheduling Many-Body Short Range MD Simulations on a Cluster of Workstations and Custom VLSI Hardware Sumanth J.V, David R. Swanson and Hong Jiang University.
ParCFD Parallel computation of pollutant dispersion in industrial sites Julien Montagnier Marc Buffat David Guibert.
PuReMD: A Reactive (ReaxFF) Molecular Dynamics Package Ananth Grama and Metin Aktulga
Acurate determination of parameters for coarse grain model.
Scaling Area Under a Curve. Why do parallelism? Speedup – solve a problem faster. Accuracy – solve a problem better. Scaling – solve a bigger problem.
Performance Model for Parallel Matrix Multiplication with Dryad: Dataflow Graph Runtime Hui Li School of Informatics and Computing Indiana University 11/1/2012.
Overcoming Scaling Challenges in Bio-molecular Simulations Abhinav Bhatelé Sameer Kumar Chao Mei James C. Phillips Gengbin Zheng Laxmikant V. Kalé.
Molecular simulation methods Ab-initio methods (Few approximations but slow) DFT CPMD Electron and nuclei treated explicitly. Classical atomistic methods.
Algorithms, Numerical Techniques, and Software for Atomistic Modeling.
Molecular Simulation of Reactive Systems. _______________________________ Sagar Pandit, Hasan Aktulga, Ananth Grama Coordinated Systems Lab Purdue University.
Computational Aspects of Multi-scale Modeling Ahmed Sameh, Ananth Grama Computing Research Institute Purdue University.
Algorithms and Software for Large-Scale Simulation of Reactive Systems _______________________________ Metin Aktulga, Sagar Pandit, Alejandro Strachan,
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
Reactive Molecular Dynamics: Algorithms, Software, and Applications. Ananth Grama Computer Science, Purdue University
Domain Decomposition in High-Level Parallelizaton of PDE codes Xing Cai University of Oslo.
Distributed simulation with MPI in ns-3 Joshua Pelkey and Dr. George Riley Wns3 March 25, 2011.
Review Session BS123A/MB223 UC-Irvine Ray Luo, MBB, BS.
© David Kirk/NVIDIA and Wen-mei W. Hwu, ECE 498AL, University of Illinois, Urbana-Champaign 1 ECE 498AL Spring 2010 Lecture 13: Basic Parallel.
MSc in High Performance Computing Computational Chemistry Module Parallel Molecular Dynamics (i) Bill Smith CCLRC Daresbury Laboratory
A Pattern Language for Parallel Programming Beverly Sanders University of Florida.
Structured ALE Solver. Overview Structured ALE mesh automatically generated Smaller input deck; Easier modifications to the mesh; Less I/O time. Shorter.
Scaling Conway’s Game of Life. Why do parallelism? Speedup – solve a problem faster. Accuracy – solve a problem better. Scaling – solve a bigger problem.
Parallel Programming in Chess Simulations Part 2 Tyler Patton.
PuReMD Design Initialization – neighbor-list, bond-list, hydrogenbond-list and Coefficients of QEq matrix Bonded interactions – Bond-order, bond-energy,
SIMD Implementation of Discrete Wavelet Transform Jake Adriaens Diana Palsetia.
Parallel Molecular Dynamics A case study : Programming for performance Laxmikant Kale
Hybrid Parallel Implementation of The DG Method Advanced Computing Department/ CAAM 03/03/2016 N. Chaabane, B. Riviere, H. Calandra, M. Sekachev, S. Hamlaoui.
PuReMD: Purdue Reactive Molecular Dynamics Software Ananth Grama Center for Science of Information Computational Science and Engineering Department of.
1 Nanoscale Modeling and Computational Infrastructure ___________________________ Ananth Grama Professor of Computer Science, Associate Director, PRISM.
Evolution at CERN E. Da Riva1 CFD team supports CERN development 19 May 2011.
David Gleich, Ahmed Sameh, Ananth Grama, et al.
Comparison to LAMMPS-REAX
Algorithms and Software for Large-Scale Simulation of Reactive Systems
PuReMD: Purdue Reactive Molecular Dynamics Software Ananth Grama Center for Science of Information Computational Science and Engineering Department of.
Integrated tool set for electro-thermal evaluation of ICs
Mid-Year Review Template March 2, 2010 Purdue University
Large Time Scale Molecular Paths Using Least Action.
Immersed Boundary Method Simulation in Titanium Objectives
Molecular simulation methods
Tahsin Reza Matei Ripeanu Nicolas Tripoul
Algorithms and Software for Large-Scale Simulation of Reactive Systems
Presentation transcript:

PuReMD: Purdue Reactive Molecular Dynamics Package Hasan Metin Aktulga and Ananth Grama Purdue University TST Meeting,May 13-14, 2010

2 Outline (Progress Report) ReaxFF Parallelization Algorithms and Numerical Techniques PuReMD Performance

Overview of ReaxFF 3 ps ns ss ms nm mm mm Statistical & Continuum methods Classical MD methods Ab-initio methods ReaxFF

Overview of ReaxFF Key Challenges: dynamic bonds modeled through bond orders  dynamic 3-body & 4-body interactions complex formulations of bond, angle, dihedral, van der Waals interactions additional interactions: lone pair, over/under-coordination, 3-body & 4- body conjugation dynamic charges using the QEq method 4

Parallelization: Decomposition Domain decomposition: 3D torus 5

Parallelization: Outer-Shell 6 b r b r b r/2 b r full shellhalf shell midpoint-shell tower-plate shell

7 Parallelization: Outer-Shell b r b r b r/2 b r full shellhalf shell midpoint-shell tower-plate shell choose full- shell due to dynamic bonding despite the comm. overhead

Parallelization: Boundary Intrs. 8

Parallelization: Messaging 9

10 Parallelization: Messaging

Very efficient generation of neighbors lists Elimination of bond order derivative lists Truncate bond related computations at the outer-shell Lookup tables for fast computation of non- bonded interactions Highly optimized parallel solver for QEq 11 Algorithmic & Numerical Tech.

Mathematical formulation: Diagonally scaled parallel CG Decreased iteration count through extrapolation from solutions to prev. steps Decreased communication overhead: redundant computations to avoid reverse comm iterate both systems together 12 QEq Solver

PuReMD Performance Weak scaling test Strong scaling test Comparison to LAMMPS-REAX Hera cluster at LLNL 4 AMD Opterons/node cores/node 800 batch nodes – cores, 127 TFLOPS/sec 32 GB memory / node Infiniband interconnect 13

Bulk Water: 6540 atoms in a 40x40x40 A 3 box / core 14 Performance: Weak Scaling

QEq scaling  Efficiency Weak Scaling 15

16 Bulk Water: atoms in a 80x80x80 A 3 box Performance: Strong Scaling

17 Efficiency and throughput Strong Scaling

18 Weak Scaling: 4-5 times faster, much lower memory footprint Comparison to Lammps-Reax

19 Strong Scaling: 3-4 times faster, much lower memory footprint Comparison to Lammps-Reax

Conclusions Verified accuracy against the original ReaxFF code through intensive test Efficient and scalable parallel implementation for ReaxFF in C using MPI 3-5 faster, much smaller memory usage Modular and extensible design allows easy improvements and enhancements Ready for PRISM device simulations Open-source code, to be released with GPL 20