IceCube simulation with PPC Photonics: 2000 – up to now Photon propagation code PPC: 2009 - now.

Slides:



Advertisements
Similar presentations
Construction process lasts until coding and testing is completed consists of design and implementation reasons for this phase –analysis model is not sufficiently.
Advertisements

Lecture 38: Chapter 7: Multiprocessors Today’s topic –Vector processors –GPUs –An example 1.
Inpainting Assigment – Tips and Hints Outline how to design a good test plan selection of dimensions to test along selection of values for each dimension.
SPICE Mie [mi:] Dmitry Chirkin, UW Madison. Updates to ppc and spice PPC: Randomized the simulation based on system time (with us resolution) Added the.
1 100M CUDA GPUs Oil & GasFinanceMedicalBiophysicsNumericsAudioVideoImaging Heterogeneous Computing CPUCPU GPUGPU Joy Lee Senior SW Engineer, Development.
Development of a track trigger based on parallel architectures Felice Pantaleo PH-CMG-CO (University of Hamburg) Felice Pantaleo PH-CMG-CO (University.
Linked to ORCA/PINGU J. Brunner. Calibrations Main stream External input Simulation Reconstruction.
Chapter Nine NetWare-Based Networking. Objectives Identify the advantages of using the NetWare network operating system Describe NetWare’s server hardware.
M. Kowalski Search for Neutrino-Induced Cascades in AMANDA II Marek Kowalski DESY-Zeuthen Workshop on Ultra High Energy Neutrino Telescopes Chiba,
CS-3013 & CS-502, Summer 2006 Memory Management1 CS-3013 & CS-502 Summer 2006.
CUDA and the Memory Model (Part II). Code executed on GPU.
Phun with Photonics Phun with Photonics Berkeley IceCube Collaboration Meeting Michelangelo D’Agostino UC Berkeley March 20, 2005.
Efficient Pseudo-Random Number Generation for Monte-Carlo Simulations Using GPU Siddhant Mohanty, Subho Shankar Banerjee, Dushyant Goyal, Ajit Mohanty.
Shekoofeh Azizi Spring  CUDA is a parallel computing platform and programming model invented by NVIDIA  With CUDA, you can send C, C++ and Fortran.
2012/06/22 Contents  GPU (Graphic Processing Unit)  CUDA Programming  Target: Clustering with Kmeans  How to use.
CuMAPz: A Tool to Analyze Memory Access Patterns in CUDA
IceCube simulation with PPC on GPUs Dmitry Chirkin, UW Madison photon propagation code graphics processing unit.
Sean Grullon with Gary Hill Maximum likelihood reconstruction of events using waveforms.
© 2008, Renesas Technology America, Inc., All Rights Reserved 1 Purpose  This training module provides an overview of optimization techniques used in.
IceCube: String 21 reconstruction Dmitry Chirkin, LBNL Presented by Spencer Klein LLH reconstruction algorithm Reconstruction of digital waveforms Muon.
I3PropagatorMMC module Dmitry Chirkin, LBNL IceCube meeting, Uppsala, 2004.
Chapter Nine NetWare-Based Networking. Introduction to NetWare In 1983, Novell introduced its NetWare network operating system Versions 3.1 and 3.1—collectively.
Greg Sullivan University of Maryland Data Filtering and Software IceCube Collaboration Meeting Monday, March 21, 2005.
Implementing a dual readout calorimeter in SLIC and testing Geant4 Physics Hans Wenzel Fermilab Friday, 2 nd October 2009 ALCPG 2009.
GPU Architecture and Programming
Photon propagation and ice properties Bootcamp UW Madison Dmitry Chirkin, UW Madison r air bubble photon.
© Janice Regan, CMPT 300, May CMPT 300 Introduction to Operating Systems Memory: Relocation.
MMC and UCR icetray modules Dima Chirkin, LBNL Presented by Spencer Klein Tau 2-bang Coincident showers.
IceCube simulation with PPC Dmitry Chirkin, UW Madison photon propagation code.
QCAdesigner – CUDA HPPS project
Photon propagation and ice properties Bootcamp UW Madison Dmitry Chirkin, UW Madison r air bubble photon.
Standard Candle, Flasher, and Cascade Simulations in IceCube Michelangelo D’Agostino UC Berkeley PSU Analysis Meeting June 21-24, 2006.
Ice Investigation with PPC Dmitry Chirkin, UW (photon propagation code)
1)Leverage raw computational power of GPU  Magnitude performance gains possible.
IceCube simulation with PPC Dmitry Chirkin, UW Madison, 2010.
Ice model update Dmitry Chirkin, UW Madison IceCube Collaboration meeting, Calibration session, March 2014.
Photonics Tables Bin Optimization Kyle Mandli Paolo Desiati University of Wisconsin – Madison Wuppertal AMANDA Collaboration Meeting.
PDAC-10 Middleware Solutions for Data- Intensive (Scientific) Computing on Clouds Gagan Agrawal Ohio State University (Joint Work with Tekin Bicer, David.
Review of Ice Models What is an “ice model”? PTD vs. photonics What models are out there? Which one(s) should/n’t we use? Kurt Woschnagg, UCB AMANDA Collaboration.
IceCube simulation with PPC Photonics: 2000 – up to now Photon propagation code PPC: now.
GPU Photon Transport Simulation Studies Mary Murphy Undergraduate, UW-Madison Dmitry Chirkin IceCube at UW-Madison Tareq AbuZayyad IceCube at UW-River.
All lepton generation and propagation with MMC Dmitry Chirkin, UCB/LBNL AMANDA meeting, Uppsala, 2004.
IceCube simulation with PPC Dmitry Chirkin, UW Madison, 2010 effective scattering coefficient (from Ryan Bay)
DirectFit reconstruction of the Aya’s two HE cascade events Dmitry Chirkin, UW Madison Method of the fit: exhaustive search simulate cascade events with.
Update on neutrino simulation with GENIE Carla Distefano.
LLVM IR, File - Praakrit Pradhan. Overview The LLVM bitcode has essentially two things A bitstream container format Encoding of LLVM IR.
Geant4 Simulation for KM3 Georgios Stavropoulos NESTOR Institute WP2 meeting, Paris December 2008.
Muon Energy reconstruction in IceCube and neutrino flux measurement Dmitry Chirkin, University of Wisconsin at Madison, U.S.A., MANTS meeting, fall 2009.
Introduction To Software Development Environment.
Heterogeneous Processing KYLE ADAMSKI. Overview What is heterogeneous processing? Why it is necessary Issues with heterogeneity CPU’s vs. GPU’s Heterogeneous.
Photon propagation and ice properties Bootcamp UW Madison Dmitry Chirkin, UW Madison r air bubble photon.
Water Properties Phone Conference1 Light Generation in KM3 and Photonics Corey Reed (Nikhef) Photon Studies KM3 MC MMC+Photonics Summary.
Light Propagation in the South Pole Ice
Overview Modern chip designs have multiple IP components with different process, voltage, temperature sensitivities Optimizing mix to different customer.
CS427 Multicore Architecture and Parallel Computing
South Pole Ice model Dmitry Chirkin, UW, Madison.
INTRODUCING Adams/CHASSIS
South Pole Ice (SPICE) model
SPICECUBE.
Freeze-In and Hole Ice Studies with Flashers
Presented by: Isaac Martin
Ice Investigation with PPC
Experimental setup (SPICE)
Fast Track Fitting in the SiD01 Detector
Search for coincidences and study of cosmic rays spectrum
Summary of yet another Photonics Workshop AMANDA/IceCube Collaboration Meeting Berkeley, March 19, 2005.
Photonics Workshop AMANDA/IceCube Collaboration Meeting Berkeley, March 19, 2005 Going the last mile…
6- General Purpose GPU Programming
Wendy Taylor STT Meeting Fermilab September 28, 2001
Presentation transcript:

IceCube simulation with PPC Photonics: 2000 – up to now Photon propagation code PPC: now

Photonics: conventional on CPU First, run photonics to fill space with photons, tabulate the result Create such tables for nominal light sources: cascade and uniform half-muon Simulate photon propagation by looking up photon density in tabulated distributions  Table generation is slow  Simulation suffers from a wide range of binning artifacts  Simulation is also slow! (most time is spent loading the tables)

PPC simulation on GPU graphics processing unit execution threads propagation steps (between scatterings) photon absorbed new photon created (taken from the pool) threads complete their execution (no more photons) Running on an NVidia GTX 295 CUDA-capable card, ppc is configured with: 384 threads in 33 blocks (total of threads) average of ~ 1024 photons per thread (total of photons per call)

Direct photon tracking with PPC simulating flasher/standard candle photons same code for muon/cascade simulation using Henyey-Greenstein scattering function with =0.8 using tabulated (in 10 m depth slices) layered ice structure employing 6-parameter ice model to extrapolate in wavelength transparent folding of acceptance and efficiencies precise tracking through layers of ice, no interpolation needed much faster than photonics for E -2 nugen and unweighted CORSIKA:  corsika files (4 sec each) in 24 hours  E -2 nugen files in 24 hours IC-40 i.e., E -2 nugen files in ~3-4 days on 6-GPU cudatest in ~1 day on 3 cuda00X computers photon propagation code

Photon Propagation Code: PPC There are 5 versions of the ppc: original c++ "fast" c++ in Assembly for CUDA GPU icetray module All versions verified to produce identical results comparison with i3mcml

ppc-gpu

ppc icetray module moved from sandbox to outdated code removed (old slower code that has not changed since 09/09) new code has a wrapper: private/ppc/i3ppc.cxx, which compiles by cmake system into the libppc.so it is necessary to compile an additional library libxppc.so by running make in private/ppc/gpu:  “make glib” compiles gpu-accelerated version (needs cuda tools)  “make clib” compiles cpu version (from the same sources!) link to libxppc.so and libcudart.so (if gpu version) from build/lib directory this library file must be loaded before the libppc.so wrapper library  Should perhaps compile the CPU version by default by cmake system?

Configuration files ppc needs the following tables: wv.dat wavelength-tabulated DOM acceptance calculated from qe_dom2007a table of efficiency.h file as.dat overall DOM efficiency and parameters of the angular sensitivity polynomial expansion before running the program create a link to one of the provided files: as.nominal nominal (measured in the lab) as.holeice hole ice (corrected by hole ice model) rnd.txt table of random number multipliers for the multiply-with-carry random number generator tilt.par files describing the ice tilt tilt.dat delete unless using icemodel.sp2 (SPICE^2) icemodel.par file with 6 parameters of the icemodel icemodel.dat main ice properties table: depth/b_e(400)/a_dust(400)/\delta\tau (as in report icecube/ ) before running the program create a link to one of the provided files: *All models* to be used with as.holeice unless otherwise specified icemodel.aha AHA model icemodel.sp1 SPICE model icemodel.sp2 SPICE^2 model icemodel.sp2+ SPICE^2+ model icemodel.sp2+n SPICE^2+ model with nominal ice icemodel.sp2x SPICE^2x model **ATTENTION** You must delete files tilt.par and tilt.dat if using AHA or SPICE, but not SPICE^2 Only SPICE^2 and higher have been fitted to use the tilted ice description. Can specify alternative directory containing these tables with the PPCTABLESDIR env. variable:  this could be a part of simprod or a separate project (for consistency), or part of ppc

Other configuration parameters A few configuration parameters may be set as “#define”s in private/ppc/gpu/ini.cxx file: #define OFLA // omit the flasher DOM #define ROMB // use rhomb cells aligned with the array #define ASENS // enable angular sensitivity #define TILT // enable tilted ice layers #define MKOW // use Marek Kowalski's photon yield parametrization #define MLTP 30 // number of multiprocessors #define WNUM 33 // optimized for 30 multiprocessors #define OVER 10 // size of photon bunches along the muon track #define NTHR 384 // NTHR*NBLK should not exceed the count of different random number multipliers #define NPHO 1024 // maximum number of photons propagated by one thread #define HNUM // size of the output hit buffer, must hold hits from up to NPHO*NTHR*NBLK photons #define MAXLYS 180 // maximum number of ice layers #define MAXGEO 5200 // maximum number of OMs #define OVR 5 // over-R: DOM radius "oversize" scaling factor

Outlook ppc has been used to simulate lots of IC40 data with V of simulation  simulates file nugen E -2 set (sufficient for an entire analysis) in 1 day on 3 cuda001/2/3 computers (18 GPUs) need to:  verify that it works for V of simulation  add code to treat high-efficient DOMs correctly  verify that it works for IC59  improve flasher simulation (interface with photoflash)  figure out the best way to compile