Computational Techniques for Efficient Carbon Nanotube Simulation

Slides:



Advertisements
Similar presentations
PHYS466 Project Kyoungmin Min, Namjung Kim and Ravi Bhadauria.
Advertisements

Formulation of an algorithm to implement Lowe-Andersen thermostat in parallel molecular simulation package, LAMMPS Prathyusha K. R. and P. B. Sunil Kumar.
The Charm++ Programming Model and NAMD Abhinav S Bhatele Department of Computer Science University of Illinois at Urbana-Champaign
Dynamic Load Balancing for VORPAL Viktor Przebinda Center for Integrated Plasma Studies.
Course Outline Introduction in algorithms and applications Parallel machines and architectures Overview of parallel machines, trends in top-500 Cluster.
Joo Chul Yoon with Prof. Scott T. Dunham Electrical Engineering University of Washington Molecular Dynamics Simulations.
© David Kirk/NVIDIA and Wen-mei W. Hwu, ECE 498AL, University of Illinois, Urbana-Champaign 1 ECE 498AL Programming Massively Parallel Processors.
Domain decomposition in parallel computing Ashok Srinivasan Florida State University COT 5410 – Spring 2004.
Efficient Parallel Implementation of Molecular Dynamics with Embedded Atom Method on Multi-core Platforms Reporter: Jilin Zhang Authors:Changjun Hu, Yali.
Algorithms and Software for Large-Scale Simulation of Reactive Systems _______________________________ Ananth Grama Coordinated Systems Lab Purdue University.
Designing and Evaluating Parallel Programs Anda Iamnitchi Federated Distributed Systems Fall 2006 Textbook (on line): Designing and Building Parallel Programs.
Molecular Dynamics Sathish Vadhiyar Courtesy: Dr. David Walker, Cardiff University.
Computational issues in Carbon nanotube simulation Ashok Srinivasan Department of Computer Science Florida State University.
Chapter 3 Parallel Algorithm Design. Outline Task/channel model Task/channel model Algorithm design methodology Algorithm design methodology Case studies.
Molecular Dynamics Collection of [charged] atoms, with bonds – Newtonian mechanics – Relatively small #of atoms (100K – 10M) At each time-step – Calculate.
Scheduling Many-Body Short Range MD Simulations on a Cluster of Workstations and Custom VLSI Hardware Sumanth J.V, David R. Swanson and Hong Jiang University.
Computational Nanotechnology A preliminary proposal N. Chandra Department of Mechanical Engineering Florida A&M and Florida State University Proposed Areas.
Protein Folding Programs By Asım OKUR CSE 549 November 14, 2002.
Long-Time Molecular Dynamics Simulations through Parallelization of the Time Domain Ashok Srinivasan Florida State University
Computer Science and Engineering Parallelizing Defect Detection and Categorization Using FREERIDE Leonid Glimcher P. 1 ipdps’05 Scaling and Parallelizing.
Computational Aspects of Multi-scale Modeling Ahmed Sameh, Ananth Grama Computing Research Institute Purdue University.
Multi-scale Modeling of Nanocrystalline Materials
Mechanics of defects in Carbon nanotubes S Namilae, C Shet and N Chandra.
Laser-Assisted Deposition of Polymer-Matrix Nanocomposites: A Combined Experimental and Computational Study DMI # , Manufacturing Processes & Equipment.
Lecture 4 TTH 03:30AM-04:45PM Dr. Jianjun Hu CSCE569 Parallel Computing University of South Carolina Department of.
Motivation: Sorting is among the fundamental problems of computer science. Sorting of different datasets is present in most applications, ranging from.
MS 15: Data-Aware Parallel Computing Data-Driven Parallelization in Multi-Scale Applications – Ashok Srinivasan, Florida State University Dynamic Data.
Data Structures and Algorithms in Parallel Computing Lecture 3.
Deepak Srivastava Computational Nanotechnology at CSC/NAS NASA Ames Research Center Moffett Field, CA Collaborators: M. Menon – University of Kentucky.
Data Structures and Algorithms in Parallel Computing Lecture 7.
Introduction to Research 2007 Introduction to Research 2007 Ashok Srinivasan Florida State University Recent collaborators V.
Parallelization Strategies Laxmikant Kale. Overview OpenMP Strategies Need for adaptive strategies –Object migration based dynamic load balancing –Minimal.
© David Kirk/NVIDIA and Wen-mei W. Hwu, ECE 498AL, University of Illinois, Urbana-Champaign 1 ECE 498AL Spring 2010 Lecture 13: Basic Parallel.
A Pattern Language for Parallel Programming Beverly Sanders University of Florida.
Slide 1 NEMOVAR-LEFE Workshop 22/ Slide 1 Current status of NEMOVAR Kristian Mogensen.
Monte Carlo Linear Algebra Techniques and Their Parallelization Ashok Srinivasan Computer Science Florida State University
Scalable Time-Parallelization of Molecular Dynamics Simulations in Nano Mechanics Y. Yu, Ashok Srinivasan, and N. Chandra Florida State University
Case Study 5: Molecular Dynamics (MD) Simulation of a set of bodies under the influence of physical laws. Atoms, molecules, forces acting on them... Have.
Data-Driven Time-Parallelization in the AFM Simulation of Proteins L. Ji, H. Nymeyer, A. Srinivasan, and Y. Yu Florida State University
Namas Chandra and Sirish Namilae
Parallel Molecular Dynamics A case study : Programming for performance Laxmikant Kale
Computer Science and Engineering Parallelizing Feature Mining Using FREERIDE Leonid Glimcher P. 1 ipdps’04 Scaling and Parallelizing a Scientific Feature.
Monte Carlo Linear Algebra Techniques and Their Parallelization Ashok Srinivasan Computer Science Florida State University
Auburn University
Lingva Technical English Projects Graphene
36th Dayton-Cincinnati Aerospace Sciences Symposium
Computational Techniques for Efficient Carbon Nanotube Simulation
Parallel Graph Algorithms
Auburn University COMP7330/7336 Advanced Parallel and Distributed Computing Mapping Techniques Dr. Xiao Qin Auburn University.
M. Mahmoodi1, M. Arjmand2, U. Sundararaj2 and S. S. Park1
Implementing Simplified Molecular Dynamics Simulation in Different Parallel Paradigms Chao Mei April 27th, 2006 CS498LVK.
Department of Computer Science University of California,Santa Barbara
Computational Nanotechnology
L21: Putting it together: Tree Search (Ch. 6)
Long-Time Molecular Dynamics Simulations in Nano-Mechanics through Parallelization of the Time Domain Ashok Srinivasan Florida State University
Milind A. Bhandarkar Adaptive MPI Milind A. Bhandarkar
Algorithms and Software for Large-Scale Simulation of Reactive Systems
Carbon Nanotube Diode Design
Course Outline Introduction in algorithms and applications
CS 501: Software Engineering Fall 1999
Sathish Vadhiyar Courtesy: Dr. David Walker, Cardiff University
Gary M. Zoppetti Gagan Agrawal
CSC4005 – Distributed and Parallel Computing
Algorithms and Software for Large-Scale Simulation of Reactive Systems
IXPUG, SC’16 Lightning Talk Kavitha Chandrasekar*, Laxmikant V. Kale
Parallel Programming in C with MPI and OpenMP
Gary M. Zoppetti Gagan Agrawal Rishi Kumar
COMPUTER ORGANIZATION AND ARCHITECTURE
Computational issues Issues Solutions Large time scale
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

Computational Techniques for Efficient Carbon Nanotube Simulation Ashok Srinivasan Computer Science Florida State University Namas Chandra Mechanical Engineering Florida State University

Outline Background Parallel nanotube simulation Nanocomposites Parallelization Conclusions and future work

Background Uses of Carbon nanotubes CNT properties Materials NEMS Transistors Displays Etc CNT properties Can span 23,000 miles without failing due to its own weight 100 times stronger than steel Many times stiffer than any known material Conducts heat better than diamond Can be a conductor or insulator without any doping Lighter than feather

Sequential computation Molecular dynamics, using Brenner’s potential Short-range interactions Neighbors can change dynamically during the course of the simulation Computational scheme Find force on each particle due to interactions with “close” neighbors Update position and velocity of each atom

Force computations Pair interactions Bond angles Four body

Profile of execution time 1: Force 2: Neighbor list 3: Predictor/corrector 4: Thermostating 5: Miscellaneous

Profile for force computations

Parallel nanotube simulation Shared memory Message passing Load balancing

Shared memory parallelization Do each of the following loops in parallel For each atom Update forces due to atom i If neighboring atoms are owned by other threads, update an auxiliary array For each thread Collect force terms for atoms it owns Srivastava, et al, SC-97 and CSE 2001 Simulated 105 to 107 atoms Speedup around 16 on 32 processors Include long-range forces too Lexical decomposition

Message passing parallelization Decompose domain into cells Each cell contains its atoms Assign a set of adjacent cells to each processor Each processor computes values for its cells, communicating with neighbors when their data is needed Caglar and Griebel, World scientific, 1999 Simulated 108 atoms on up to 512 processors Linear speedup for 160,000 atoms on 64 processors

Load balancing

Nanocomposites Matrix-nanotube interface modeled with springs An extra force term computed for atoms attached to springs Springs can break, requiring substantial increase in computations in that region Spring Polymer matrix

Parallelization Distributed shared memory Balance the load Ensure locality of data Simple lexical approach will result in load imbalance Balanced lexical Adjust the domain size

Breadth first search We want to ensure locality too Balanced Breadth First Search Perform a breadth first search on the atom interaction graph

Use general purpose software Other approaches Use general purpose software Jostle Metis ParMetis

Experimental parameters Nanotube with 1000 atoms Spring probability: 0.05 Probability of a spring breaking in an iteration: 0.01 Load increase factor due to spring break: 200 Disturbance region depth: 3 Number of time steps: 100

Load imbalance

Non-local interactions

Load balancing time

Variation of load with time

Conclusions and future work Neighbor search Parallelization Current approaches appear inadequate General purpose software is too slow Special purpose techniques appear promising Stochastic versions of certain current techniques possible Multi-scale simulation of nano-composites MD at nano-scale and FEM at larger scale