S. Guatelli, A. Mantero, J. Moscicki, M. G. Pia Geant4 medical simulations in a distributed computing environment 4th Workshop on Geant4 Bio-medical Developments.

Slides:



Advertisements
Similar presentations
NGS computation services: APIs and.
Advertisements

Maria Grazia Pia Simulation in a Distributed Computing Environment Simulation in a Distributed Computing Environment S. Guatelli 1, A. Mantero 1, P. Mendez.
LIBRA: Lightweight Data Skew Mitigation in MapReduce
28 April, 2005ISGC 2005, Taiwan The Efficient Handling of BLAST Applications on the GRID Hurng-Chun Lee 1 and Jakub Moscicki 2 1 Academia Sinica Computing.
Maria Grazia Pia Geant4 LowE Workshop 30-31/5/2002 ow Energy e.m. Workshop CERN, May 2002.
Beowulf Supercomputer System Lee, Jung won CS843.
Maria Grazia Pia, INFN Genova 1 Part V The lesson learned Summary and conclusions.
Monte Carlo simulation for radiotherapy in a distributed computing environment S. Chauvie 2,3, S. Guatelli 2, A. Mantero 2, J. Moscicki 1, M.G. Pia 2 CERN.
Other GEANT4 capabilities Event biasing Parameterisation (fast simulation) Persistency Parallelisation and integration in a distributed computing environment.
A general purpose dosimetric system for brachytherapy
DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Distributed Computing in Physics Parallel Geant4 Simulation in Medical.
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
DIANE Overview Germán Carrera, Alfredo Solano (CNB/CSIC) EMBRACE COURSE Monday 19th of February to Friday 23th. CNB-CSIC Madrid.
Task Farming on HPCx David Henty HPCx Applications Support
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
Introduction to Parallel Programming MapReduce Except where otherwise noted all portions of this work are Copyright (c) 2007 Google and are licensed under.
IEEE Nuclear Science Symposium and Medical Imaging Conference Short Course The Geant4 Simulation Toolkit Sunanda Banerjee (Saha Inst. Nucl. Phys., Kolkata,
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
PTC ½ day – Experience in PS2 and SPS H. Bartosik, Y. Papaphilippou.
KISTI Grid CA Status Report KISTI Supercomputing Center Sangwan Kim APGridPMA Meeting Mar 8, 2010 Academia Sinica, Taipei, Taiwan.
March 3rd, 2006 Chen Peng, Lilly System Biology1 Cluster and SGE.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Data Analysis using Java Mobile Agents Mark Dönszelmann, Information, Process and Technology Group, IT, CERN ATLAS Software Workshop Analysis Tools Meeting,
Geant4 Acceptance Suite for Key Observables CHEP06, T.I.F.R. Mumbai, February 2006 J. Apostolakis, I. MacLaren, J. Apostolakis, I. MacLaren, P. Mendez.
DIANE Project CHEP 03 DIANE Distributed Analysis Environment for semi- interactive simulation and analysis in Physics Jakub T. Moscicki,
Loosely Coupled Parallelism: Clusters. Context We have studied older archictures for loosely coupled parallelism, such as mesh’s, hypercubes etc, which.
Maria Grazia Pia, INFN Genova Test & Analysis Project aka “statistical testing” Maria Grazia Pia, INFN Genova on behalf of the T&A team
A General Purpose Brachytherapy Software Simulation + Analysis (isodose calculation) 2/10/2002 Geant4 Workshop CERN Susanna Guatelli Univ. and INFN Genova.
Enabling Grids for E-sciencE EGEE-III INFSO-RI Using DIANE for astrophysics applications Ladislav Hluchy, Viet Tran Institute of Informatics Slovak.
Grid Computing at Yahoo! Sameer Paranjpye Mahadev Konar Yahoo!
Technological Transfer from HEP to Medical Physics How precise Brachytherapy MonteCarlo simulations can be applied in Clinics Reality Problem: How to achieve.
Distributed System Concepts and Architectures 2.3 Services Fall 2011 Student: Fan Bai
UAB Dynamic Tuning of Master/Worker Applications Anna Morajko, Paola Caymes Scutari, Tomàs Margalef, Eduardo Cesar, Joan Sorribes and Emilio Luque Universitat.
Maria Grazia Pia, INFN Genova Introduction to medical physics applications Maria Grazia Pia, INFN Genova Geant4 Workshop,
ROOT and Federated Data Stores What Features We Would Like Fons Rademakers CERN CC-IN2P3, Nov, 2011, Lyon, France.
Parallel & Distributed Systems and Algorithms for Inference of Large Phylogenetic Trees with Maximum Likelihood Alexandros Stamatakis LRR TU München Contact:
Benchmarks of medical dosimetry simulation on the grid S. Chauvie 1, A. Lechner 4, P. Mendez Lorenzo 5, J. Moscicki 5, M.G. Pia 6 G.A.P. Cirrone 2, G.
LOGO Development of the distributed computing system for the MPD at the NICA collider, analytical estimations Mathematical Modeling and Computational Physics.
IRCC & Mauriziano Hospital & INFN & S Croce e Carle Hospital
By Jeff Dean & Sanjay Ghemawat Google Inc. OSDI 2004 Presented by : Mohit Deopujari.
PROOF and ALICE Analysis Facilities Arsen Hayrapetyan Yerevan Physics Institute, CERN.
Super Scaling PROOF to very large clusters Maarten Ballintijn, Kris Gulbrandsen, Gunther Roland / MIT Rene Brun, Fons Rademakers / CERN Philippe Canal.
PROOF tests at BNL Sergey Panitkin, Robert Petkus, Ofer Rind BNL May 28, 2008 Ann Arbor, MI.
Cloud Computing project NSYSU Sec. 1 Demo. NSYSU EE IT_LAB2 Outline  Our system’s architecture  Flow chart of the hadoop’s job(web crawler) working.
Susanna Guatelli Geant4 in a Distributed Computing Environment S. Guatelli 1, P. Mendez Lorenzo 2, J. Moscicki 2, M.G. Pia 1 1. INFN Genova, Italy, 2.
Geant4 Training 2004 Short Course Katsuya Amako (KEK) Gabriele Cosmo (CERN) Giuseppe Daquino (CERN) Susanna Guatelli (INFN Genova) Aatos Heikkinen (Helsinki.
CIP HPC CIP - HPC HPC = High Performance Computer It’s not a regular computer, it’s bigger, faster, more powerful, and more.
NGS computation services: APIs and.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
Introduction to Performance Tuning Chia-heng Tu PAS Lab Summer Workshop 2009 June 30,
PROOF on multi-core machines G. GANIS CERN / PH-SFT for the ROOT team Workshop on Parallelization and MultiCore technologies for LHC, CERN, April 2008.
GRID & Parallel Processing Koichi Murakami11 th Geant4 Collaboration Workshop / LIP - Lisboa (10-14/Oct./2006) 1 GRID-related activity in Japan Go Iwai,
Brief introduction about “Grid at LNS”
F. Foppiano3, S. Guatelli2, J. Moscicki1, M.G. Pia2 CERN1 INFN Genova2
GWE Core Grid Wizard Enterprise (
Simulation in a Distributed Computing Environment
Grid Computing Colton Lewis.
NGS computation services: APIs and Parallel Jobs
F. Foppiano, S. Guatelli, B. Mascialino, M. G. Pia, M. Piergentili
Introduction to medical physics applications
Support for ”interactive batch”
Other GEANT4 capabilities
Simulation in a Distributed Computing Environment
Short Course Siena, 5-6 October 2006
The Hadrontherapy Geant4 advanced example
Simulation in a Distributed Computing Environment
Distributed Simulation with Geant4
The Geant4 Hadrontherapy Advanced Example
Presentation transcript:

S. Guatelli, A. Mantero, J. Moscicki, M. G. Pia Geant4 medical simulations in a distributed computing environment 4th Workshop on Geant4 Bio-medical Developments Geant4 Physics Validation INFN Genova, July 2005

Outline Problem: how to obtain quick response Brief introduction of DIANE How to parallelize a Geant4 application Project –Parallelization of two medical physics Geant4 applications Brachytherapy IMRT application –Study of the performance Using a dedicated cluster Using the GRID First results Work in progress

Execution time - Brachy Number of events for a sufficient statistic for a dosimetric study of a single brachytherapic source: 20 M events Execution time of 20 M events on a Pentium IV, 3 GHz –16650 s ~ 5 h Clinical use: quick response means order of minutes

Execution times – IMRT Number of events for a sufficient statistic: 10 9 events Execution time of 10 9 events: – s ~ 228 h ~ 9 days and half Quick response required for clinical use

Speed adequate for clinic use Transparent configuration in sequential or parallel mode Transparent access to the GRID through an intermediate software layer Parallelisation Access to distributed computing resources

Project Parallelization of the Geant4 IMRT and Brachytherapy application Parallelization through DIANE Performance test –Run on a single machine –Run on a dedicated cluster –Run on the GRID

DIANE DIstributed ANalysis Environment speed OK but expensive hardware investment + maintenance IRCC LAN SWITCHSWITCH Node01 Node02 Node03 Node04 IMRT Geant4 Simulation and Anaphe Analysis on a dedicated Beowulf Cluster S. Chauvie et al., IRCC Torino, Siena 2002 Previous studies for parallelization of a Geant4 based medical application Alternative strategy DIANE ParallelisationAccess to the GRID Transparent access to a distributed computing environment

DIANE DIstributed ANalysis Environment prototype for an intermediate layer between applications and the GRID Parallel cluster processing – make fine tuning and customisation easy – transparently using GRID technology – application independent Hide complex details of underlying technology Developed by J. Moscicki, CERN

Practical example How to “dianize” the Geant4 application: –Look the Geant4 extended example: ExDIANE in the parallel directory –Completely transparent to the user: same G4 code –Documentation at specific for Geant4 applications availablewww.cern.ch/diane/

Run through DIANE #-*-python-*- Application = "G4Analysis" WorkerInitData = { 'G4ApplicationComponentName' : "G4MedLinac", ‘initMacroFile' : ""“ /control/verbose 1/run/verbose 1 /control/saveHistory /run/initialize/tracking/storeTrajectory 1 /Jaws/X1/DistanceFromAxis -5.0 cm /Jaws/X2/DistanceFromAxis 5.0 cm /Jaws/Y1/DistanceFromAxis -5.0 cm/ Jaws/Y2/DistanceFromAxis 5.0 cm /Jaws/update/energy 6.0 MeV /sourceType MeV """} JobInitData = { 'runParams' : { 'seed' : 0, 'eventNumber' : , 'macroFileTemplate': "/run/beamOn " }, 'eventsPerWorker' :10000, 'workerInit' : WorkerInitData } Example of a macro file

Run on parallel mode on a dedicated cluster Type the command: –Diane.startjob –j macrofileName.mac –wms=IPLIST IPLIST if a file name containing the list of the names of the machines the user intends to use

Practical Example example: Geant4 simulation with analysis the total number of events of the simulation is divided in tasks each task produces a file with histograms job result = sum of histograms produced by tasks master-worker model client starts a job workers perform tasks and produce histograms master integrates the results

Resources of this project Resources of this project Dedicated cluster of 4 pc’s in Genova (Pentium IV, 3 GHz) –For preliminary tests Dedicated 30 pc’s (biprocessors) cluster (Xeon, 2.8 GHz) –Thanks to H.C. Lee, Academia Sinica Computing Center, Taiwan LSF cluster at CERN –To study a real case of running on a cluster, used by more users GRID –Run on a distributed computing environment

Running on a single CPU To study the efficiency of DIANE Plot of: with respect to the number of events Execution time using DIANE means running sequentially the dianized simulation, dividing the job in tasks Useful study to optimize the number of events for task Brachy-Iridium source simulation Overhead of DIANE Preliminary

Preliminary results Running on a dedicated cluster –Divide the total number of events in tasks –Dispatch the tasks on more workers Execution times of the Brachy with respect to the number of CPU’s used No merging of the output files of the simulations

Preliminary results Running on a dedicated cluster Efficiency: The efficiency is higher with higher number of tasks N is the number of CPUs Preliminary

Performance Optimization Why the efficiency is higher with higher number of tasks The execution time is bigger because there is still one task to end Splitting the job in more tasks increases the balance in execution times of the workers

Work in progress Optimization of the method to merge the output files of the tasks –In the present situation the merging introduces a significant overhead on the results –We found problems with adding histograms with PI –A. Pfeiffer and L. Moneta are helping in this last task Last step: running on the GRID Refine the results