Ab initio simulations for (the WITCH) Penning traps using a graphics cards for faster coulomb interaction calculations S. Van Gorp, M. Breitenfeldt, M.

Slides:



Advertisements
Similar presentations
DEVELOPMENT OF ONLINE EVENT SELECTION IN CBM DEVELOPMENT OF ONLINE EVENT SELECTION IN CBM I. Kisel (for CBM Collaboration) I. Kisel (for CBM Collaboration)
Advertisements

Instructor Notes Lecture discusses parallel implementation of a simple embarrassingly parallel nbody algorithm We aim to provide some correspondence between.
INSTITUUT VOOR KERN- EN STRALINGSFYSICA 16-May-2002Weak Interaction Trap for CHarged particles1 WITCH: Physics goals and Present Status Valentin Kozlov.
Chapter 21 Electric Charge and Electric Field. Charles Allison © 2000 Question An  particle with a charge +2e and a mass of 4m p is on a collision course.
Scalable Multi-Cache Simulation Using GPUs Michael Moeng Sangyeun Cho Rami Melhem University of Pittsburgh.
Christopher McCabe, Derek Causon and Clive Mingham Centre for Mathematical Modelling & Flow Analysis Manchester Metropolitan University MANCHESTER M1 5GD.
IIAA GPMAD A beam dynamics code using Graphics Processing Units GPMAD (GPU Processed Methodical Accelerator Design) utilises Graphics Processing Units.
Topology-Caching for Dynamic Particle Volume Raycasting Jens Orthmann, Maik Keller and Andreas Kolb, University of Siegen.
TUPEC057 Advances With Merlin – A Beam Tracking Code J. Molson, R.J. Barlow, H.L. Owen, A. Toader MERLIN is a.
ECE 562 Computer Architecture and Design Project: Improving Feature Extraction Using SIFT on GPU Rodrigo Savage, Wo-Tak Wu.
 asymmetry parameter measurements in nuclear  -decay as a probe for non-Standard Model physics K.U.Leuven, Univ. Bonn, NPI Rez (Prague), ISOLDE CERN.
Multi Agent Simulation and its optimization over parallel architecture using CUDA™ Abdur Rahman and Bilal Khan NEDUET(Department Of Computer and Information.
Daniel Blackburn Load Balancing in Distributed N-Body Simulations.
Simulating Self Gravitating Planetesimal Disks on the Graphics Processing Unit (GPU) Alice C. Quillen Alex Moore University of Rochester Poster DPS 2008.
Newtonian N-Body Simulator Michelle Teh, Michael Witten April 27, 2007.
Weekly Report Start learning GPU Ph.D. Student: Leo Lee date: Sep. 18, 2009.
ALICE HLT High Speed Tracking and Vertexing Real-Time 2010 Conference Lisboa, May 25, 2010 Sergey Gorbunov 1,2 1 Frankfurt Institute for Advanced Studies,
Adnan Ozsoy & Martin Swany DAMSL - Distributed and MetaSystems Lab Department of Computer Information and Science University of Delaware September 2011.
GPU Graphics Processing Unit. Graphics Pipeline Scene Transformations Lighting & Shading ViewingTransformations Rasterization GPUs evolved as hardware.
GPGPU overview. Graphics Processing Unit (GPU) GPU is the chip in computer video cards, PS3, Xbox, etc – Designed to realize the 3D graphics pipeline.
1 Scalable Distributed Fast Multipole Methods Qi Hu, Nail A. Gumerov, Ramani Duraiswami Institute for Advanced Computer Studies Department of Computer.
CuMAPz: A Tool to Analyze Memory Access Patterns in CUDA
Shared memory systems. What is a shared memory system Single memory space accessible to the programmer Processor communicate through the network to the.
1 TCP06 Parksville 8/5/06 Electron capture branching ratios for the nuclear matrix elements in double-beta decay using TITAN ◆ Nuclear matrix elements.
Many-Core Scalability of the Online Event Reconstruction in the CBM Experiment Ivan Kisel GSI, Germany (for the CBM Collaboration) CHEP-2010 Taipei, October.
WITCH Control System Michaël Tandecki & The WITCH collaboration S. Van Gorp, E. Traykov, V. De Leebeeck, G. Soti, F. Wauters, N. Severijns (K.U.Leuven),
Interactive Time-Dependent Tone Mapping Using Programmable Graphics Hardware Nolan GoodnightGreg HumphreysCliff WoolleyRui Wang University of Virginia.
WITCH status + Simbuca, a Penning trap simulation program S. Van Gorp, M. Breitenfeldt, V. De Leebeeck,T. Porobic, G. Soti, M. Tandecki, F. Wauters, N.
Accelerating a Software Radio Astronomy Correlator By Andrew Woods Supervisor: Prof. Inggs & Dr Langman.
Robert Liao Tracy Wang CS252 Spring Overview Traditional GPU Architecture The NVIDIA G80 Processor CUDA (Compute Unified Device Architecture) LAPACK.
GPU Architecture and Programming
INSTITUUT VOOR KERN- EN STRALINGSFYSICA 23-May-2002Weak Interaction Trap for CHarged particles1 Progress of the WITCH project Valentin Kozlov NIPNET meeting,
GPU DAS CSIRO ASTRONOMY AND SPACE SCIENCE Chris Phillips 23 th October 2012.
Tone Mapping on GPUs Cliff Woolley University of Virginia Slides courtesy Nolan Goodnight.
Development of the distributed monitoring system for the NICA cluster Ivan Slepov (LHEP, JINR) Mathematical Modeling and Computational Physics Dubna, Russia,
An FPGA Implementation of the Ewald Direct Space and Lennard-Jones Compute Engines By: David Chui Supervisor: Professor P. Chow.
Nathal Severijns, JRA-11 TRAPSPEC TRAPSPEC JRA-11 EURONS PCC Meeting, Groningen, December 2006 JRA-11 (TRAPSPEC) Improvements and developments of.
Simulating the Nonlinear Schrodinger Equation using MATLAB with CUDA
Simon Van Gorp 1/x Thesis defense 28 th of February, 2011 Search for physics beyond the standard electroweak model with the WITCH experiment Simon Van.
By Dirk Hekhuis Advisors Dr. Greg Wolffe Dr. Christian Trefftz.
1)Leverage raw computational power of GPU  Magnitude performance gains possible.
May 8, 2007Farid Harhad and Alaa Shams CS7080 Overview of the GPU Architecture CS7080 Final Class Project Supervised by: Dr. Elias Khalaf By: Farid Harhad.
Simon Van Gorp PhD defense 28th of February, Leuven
IceCube simulation with PPC Photonics: 2000 – up to now Photon propagation code PPC: now.
An Efficient CUDA Implementation of the Tree-Based Barnes Hut n-body Algorithm By Martin Burtscher and Keshav Pingali Jason Wengert.
Mayur Jain, John Verboncoeur, Andrew Christlieb [jainmayu, johnv, Supported by AFOSR Abstract Electrostatic particle based models Developing.
Simbuca [1], using a graphics card to simulate Coulomb interactions in a Penning trap Simon Van Gorp [1]: S. Van Gorp et al. (2011), Nuclear Instruments.
1 November 11, 2015 A Massively Parallel, Hybrid Dataflow/von Neumann Architecture Yoav Etsion November 11, 2015.
David Angulo Rubio FAMU CIS GradStudent. Introduction  GPU(Graphics Processing Unit) on video cards has evolved during the last years. They have become.
Barnes Hut N-body Simulation Martin Burtscher Fall 2009.
WITCH - a first determination of the beta-neutrino angular correlation coefficient in 35 Ar decay S. Van Gorp, M. Breitenfeldt, V. De Leebeeck,T. Porobic,
Space Charge with PyHEADTAIL and PyPIC on the GPU Stefan Hegglin and Adrian Oeftiger Space Charge Working Group meeting –
Computer Architecture Lecture 24 Parallel Processing Ralph Grishman November 2015 NYU.
IceCube simulation with PPC Photonics: 2000 – up to now Photon propagation code PPC: now.
3/12/2013Computer Engg, IIT(BHU)1 CUDA-3. GPGPU ● General Purpose computation using GPU in applications other than 3D graphics – GPU accelerates critical.
INSTITUUT VOOR KERN- EN STRALINGSFYSICA 18-September-2002Weak Interaction Trap for CHarged particles1 Present status of WITCH experiment Valentin Kozlov.
Barnes Hut N-body Simulation Martin Burtscher Fall 2009.
Experimental Tests of the Standard Model of Weak Interactions D.Zakoucky, Nuclear Physics Institute ASCR, Rez near Prague Experimental facility WITCH at.
The WITCH Experiment 5th International Symposium on Symmetries in Subatomic Physics June 18-22, 2012 KVI, Groningen G. Ban 1, M. Breitenfeldt 2, V. De.
WITCH - a first determination of the beta-neutrino angular correlation S. Van Gorp, M. Breitenfeldt, V. De Leebeeck,T. Porobic, G. Soti, M. Tandecki, N.
General Purpose computing on Graphics Processing Units
MCP Analysis on ALpha Howard Chiao.
ALICE HLT tracking running on GPU
Graphics Processing Unit
Faster File matching using GPGPU’s Deephan Mohan Professor: Dr
Electric Fields and Potential
NVIDIA Fermi Architecture
Kenneth Moreland Edward Angel Sandia National Labs U. of New Mexico
6- General Purpose GPU Programming
N-Body Gravitational Simulations
Presentation transcript:

Ab initio simulations for (the WITCH) Penning traps using a graphics cards for faster coulomb interaction calculations S. Van Gorp, M. Breitenfeldt, M. Tandecki, F. Wauters, E. Traykov,, N. Severijns (K.U.Leuven, Belgium), M. Beck, P. Friedag, C. Weinheimer (Univ. Munster, Germany), A. Herlert (ISOLDE-CERN, Geneva, Switserland), V. Kozlov, F. Gluck (Univ. Karlsruhe, Germany), D. Zakoucky (NPI-Rez, Prague, Czech)

Simon Van Gorp – TCP Saariselkä WITCH introduction Simulation motivation Coulomb interactions Tree codes Chamomile scheme: Coulomb on a GPU. Performance GPU vs CPU Simonion, a Penning Trap simulation program Integrator Buffer gas collisions Conclusion and Outlook 2/14 Overview

Simon Van Gorp – TCP Saariselkä WITCH Introduction 3/14 The WITCH experiment looks for scalar currents in nuclear beta decay. A scattering free source is needed because of the low recoil energy. A retardation barrier is applied and the # ions reaching an MCP detector are counted.

Simulation Motivation Simon Van Gorp – TCP Saariselkä The retardation spectrometer is a combination of E and B fields. A particle tracking routine was developed to understand the behavior of the ions by F. Gluck. A good source of ions is needed to describe the process correctly. Understand space charge effects in the Penning traps. 4/14 WITCH: 10 6 ions per cycle -> Computer simulations which are dominated by the Coulomb interaction calculation

Coulomb interactions Simon Van Gorp – TCP Saariselkä Coulomb force scales with O(N 2 ) Tree methods (Barnes Hut, PM, P 3 M, FMM) reduces this to O(N log N) 5/14 Space is divided in nodes. Which are subdivided A node has the total charge and mass, and is located on the centre of mass. Approx. long range force by aggregating particles into one particle and use the force of this one particle Scaled Coulomb Force puts more weight to the charge of one ion to simulate more ions. Works well [1] [1]: D. Beck et al, Hyp. Int. 132, 2001

Why a GPU? Simon Van Gorp – TCP Saariselkä /14 GPU CPU -high parallelismSTILL TO DO -very fast floating point calculations -SIMD structure (pipelining!) Geforce 8800 GTX

Chamomile scheme Simon Van Gorp – TCP Saariselkä Calculating gravitational interactions on a Graphics Card via the Chamomile scheme from Hamada and Iitaka (in 2007). 7/14 Why a GPU? -parallelism! -only 20 float operations -CUDA programming language for GPU’s J-bag I-pot force [2]: T. Hamada and T. Iitaka, arXiv.org:astro-ph/ , 2007

Chamomile scheme: practical usage Simon Van Gorp – TCP Saariselkä Function provided by Hamada and Iitaka: Gravitational force ≈ Coulomb Force Conversion coefficient: Needed: -64 bit linux -NVIDIA Graphics Card that supports CUDA - CUDA environment v2.3 Not needed: -CUDA knowledge -… 8/14

Simon Van Gorp – TCP Saariselkä Ar + ions No Buffergas 9/14 Movie

Simon Van Gorp – TCP Saariselkä GPU vs CPU GPU blows the CPU away. The effect becomes more visible with even more particles simulated. Simulated is a quadrupole excitation for 100ms with buffer gas. This takes 3 days with a GPU compared to 3-4 years with a CPU! GPU improvement factorCPU and GPU simulation time

Simon Van Gorp – TCP Saariselkä /14 Simonion overview Simonion is a modular Penning Trap simulation package. Reading external fieldmaps Trap excitations 3 different integrators 2 buffergas routines Possibility to include external fieldmaps Can run on CPU and GPU Compile with g++ or icpc A root analysis file is provided A Makefile is provided

Integrators and buffer gas models Simon Van Gorp – TCP Saariselkä Integrators: 4 th and 5 th order Runga Kutta with adaptive stepsize and error control. 1 st order (predictor corrector) Gear method. Buffer gas models: Langevin or polarizability model (= for all mases) Ion Mobility based model ( ≈ for all mases) 12/14

Simon Van Gorp – TCP Saariselkä Other GPU bases libraries: Kirin: + corrector and predictor integrator [3] 2 x faster Sapporo: double precision, corrector on CPU [4] 2 x faster ? N-Body from Nvidia [5] 2 x faster ? Tree codes Scaled Coulomb force 13/14 Possible improvements [3]: Belleman et al, New Astronomy, 13(2), ,2007 [4]: Evghenii et al, New Astronomy, 14(7), ,2009 [5]:Nyland et al, GPU Gems 3, Chapter 31, Addison-Wesley, 2007

Simon Van Gorp – TCP Saariselkä A Penning Trap simulation package Simonion is presented that is stable and intuitive to use. The first program that uses a GPU to calculate Coulomb interactions much faster! GPU computing is a new field of which we barely scratched the surface. Possible improvements like tree codes and scaled Coulomb force will push the limit to 10 6 particles 14/14 Conclusion

Thank you for your attention

Simon Van Gorp – TCP Saariselkä /24