DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Distributed Computing in Physics Parallel Geant4 Simulation in Medical.

Slides:



Advertisements
Similar presentations
Maria Grazia Pia Simulation in a Distributed Computing Environment Simulation in a Distributed Computing Environment S. Guatelli 1, A. Mantero 1, P. Mendez.
Advertisements

MINJAE HWANG THAWAN KOOBURAT CS758 CLASS PROJECT FALL 2009 Extending Task-based Programming Model beyond Shared-memory Systems.
Maria Grazia Pia, INFN Genova 1 Part V The lesson learned Summary and conclusions.
Monte Carlo simulation for radiotherapy in a distributed computing environment S. Chauvie 2,3, S. Guatelli 2, A. Mantero 2, J. Moscicki 1, M.G. Pia 2 CERN.
Other GEANT4 capabilities Event biasing Parameterisation (fast simulation) Persistency Parallelisation and integration in a distributed computing environment.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
Piccolo – Paper Discussion Big Data Reading Group 9/20/2010.
Copyright © 1998 Wanda Kunkle Computer Organization 1 Chapter 2.1 Introduction.
The Origin of the VM/370 Time-sharing system Presented by Niranjan Soundararajan.
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
TPB Models Development Status Report Presentation to the Travel Forecasting Subcommittee Ron Milone National Capital Region Transportation Planning Board.
Bob Thome, Senior Director of Product Management, Oracle SIMPLIFYING YOUR HIGH AVAILABILITY DATABASE.
CERN - IT Department CH-1211 Genève 23 Switzerland t The High Performance Archiver for the LHC Experiments Manuel Gonzalez Berges CERN, Geneva.
Building a Real Workflow Thursday morning, 9:00 am Lauren Michael Research Computing Facilitator University of Wisconsin - Madison.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Performance Concepts Mark A. Magumba. Introduction Research done on 1058 correspondents in 2006 found that 75% OF them would not return to a website that.
Goodbye rows and tables, hello documents and collections.
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
DIANE Project CHEP 03 DIANE Distributed Analysis Environment for semi- interactive simulation and analysis in Physics Jakub T. Moscicki,
Offline Coordinators  CMSSW_7_1_0 release: 17 June 2014  Usage:  Generation and Simulation samples for run 2 startup  Limited digitization and reconstruction.
Wenjing Wu Andrej Filipčič David Cameron Eric Lancon Claire Adam Bourdarios & others.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
CMAQ Runtime Performance as Affected by Number of Processors and NFS Writes Patricia A. Bresnahan, a * Ahmed Ibrahim b, Jesse Bash a and David Miller a.
1 PRAGUE site report. 2 Overview Supported HEP experiments and staff Hardware on Prague farms Statistics about running LHC experiment’s DC Experience.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
Technological Transfer from HEP to Medical Physics How precise Brachytherapy MonteCarlo simulations can be applied in Clinics Reality Problem: How to achieve.
S. Guatelli, A. Mantero, J. Moscicki, M. G. Pia Geant4 medical simulations in a distributed computing environment 4th Workshop on Geant4 Bio-medical Developments.
Building a Real Workflow Thursday morning, 9:00 am Lauren Michael Research Computing Facilitator University of Wisconsin - Madison.
ROOT and Federated Data Stores What Features We Would Like Fons Rademakers CERN CC-IN2P3, Nov, 2011, Lyon, France.
LOGO Development of the distributed computing system for the MPD at the NICA collider, analytical estimations Mathematical Modeling and Computational Physics.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
11 CLUSTERING AND AVAILABILITY Chapter 11. Chapter 11: CLUSTERING AND AVAILABILITY2 OVERVIEW  Describe the clustering capabilities of Microsoft Windows.
MROrder: Flexible Job Ordering Optimization for Online MapReduce Workloads School of Computer Engineering Nanyang Technological University 30 th Aug 2013.
GRID activities in Wuppertal D0RACE Workshop Fermilab 02/14/2002 Christian Schmitt Wuppertal University Taking advantage of GRID software now.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
Infrastructure for Data Warehouses. Basics Of Data Access Data Store Machine Memory Buffer Memory Cache Data Store Buffer Bus Structure.
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
PROOF tests at BNL Sergey Panitkin, Robert Petkus, Ofer Rind BNL May 28, 2008 Ann Arbor, MI.
PROOF Benchmark on Different Hardware Configurations 1 11/29/2007 Neng Xu, University of Wisconsin-Madison Mengmeng Chen, Annabelle Leung, Bruce Mellado,
Parallelization Geant4 simulation is an embarrassingly parallel computational problem – each event can possibly be treated independently 1.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
15-Feb-02Steve Traylen, RAL WP6 Test Bed Report1 RAL/UK WP6 Test Bed Report Steve Traylen, WP6 PPGRID/RAL, UK
Susanna Guatelli Geant4 in a Distributed Computing Environment S. Guatelli 1, P. Mendez Lorenzo 2, J. Moscicki 2, M.G. Pia 1 1. INFN Genova, Italy, 2.
Maria Grazia Pia, INFN Genova and CERN1 Geant4 highlights of relevance for medical physics applications Maria Grazia Pia INFN Genova and CERN.
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
Some GPU activities at the CMS experiment Felice Pantaleo EP-CMG-CO EP-CMG-CO 1.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Background Computer System Architectures Computer System Software.
ETICS An Environment for Distributed Software Development in Aerospace Applications SpaceTransfer09 Hannover Messe, April 2009.
Next Generation of Apache Hadoop MapReduce Owen
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Thomas Baus Senior Sales Consultant Oracle/SAP Global Technology Center Mail: Phone:
Architecture of a platform for innovation and research Erik Deumens – University of Florida SC15 – Austin – Nov 17, 2015.
Enabling Grids for E-sciencE LRMN ThIS on the Grid Sorina CAMARASU.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
GRID & Parallel Processing Koichi Murakami11 th Geant4 Collaboration Workshop / LIP - Lisboa (10-14/Oct./2006) 1 GRID-related activity in Japan Go Iwai,
F. Foppiano3, S. Guatelli2, J. Moscicki1, M.G. Pia2 CERN1 INFN Genova2
Solid State Disks Testing with PROOF
Simulation in a Distributed Computing Environment
Kilohertz Decision Making on Petabytes
Grid Canada Testbed using HEP applications
F. Foppiano, S. Guatelli, B. Mascialino, M. G. Pia, M. Piergentili
Support for ”interactive batch”
Other GEANT4 capabilities
Simulation in a Distributed Computing Environment
Simulation in a Distributed Computing Environment
Distributed Simulation with Geant4
Presentation transcript:

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Distributed Computing in Physics Parallel Geant4 Simulation in Medical and Space Science Applications Jakub T. Moscicki, CERN/IT Maria G. Pia, INFN Genova Alfonso Mantero, INFN Genova Susanna Guatelli, INFN Genova

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Applications of Distributed Technology and GRID Examples of interdisciplinary applications Geant4 simulation and analysis speed-up factor ~ 30 times DIANE R&D Project: application-oriented gateway to GRID developed for LHC CERN IT/API – INFN Geant4/LowEnergy collaboration cern.ch/diane LHC: ntuple analysis and simulation radiotherapy: brachytherapy, IMRT space missions: ESA Bepi Colombo, LISA

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Why Distributed Computing? share limited hardware resources lend when not needed, borrow when needed optimize load of CPUs avoid redundancy: save common disk space distributed collaborations e.g. LHC community share and manage access to distributed data replication, security, consistency move processing close to available resources e.g. data process in parallel

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 What is GRID ? global, unified resource access system a la WWW: easy and universal access virtual organisations over administrative boundaries black-box: sumbit here, run anywhere world of virtual happiness but... in pratice to work efficiently and correctly every generic system must be customized to match specific experiment's needs and their configuration technology in constant evolution mature and universally accessible GRID still to come

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 DIstributed ANalysis Environment parallel cluster processing make fine tuning and customization easy transparently using GRID technology accessible via a Wide Area Network application independent

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 DIstributed ANalysis Environment hide complex details of underlying technology easy to use dedicated to master-worker model most of typical jobs: ntuple analysis, event level distributed simulation

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Preliminary Benchmark Results

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Standard Geant4 Simulation the goal of simulation: study the experimental configuration and the physics reach for Bepi Colombo ESA mission to Mercury requires high statistics  many events 20 Mio events ~ 3 hours up to 100 Mio events might be useful estimated time ~16 hours analysis implemented with AIDA/Anaphe

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Distributed Geant4 Simulation increase performance shift from batch to semi-interactive simulation user can study the results of the simulation faster and more often generate more events – debug simulation faster correctness and ease of use preserve reproducability of the results parallel should look as local to users main goals:

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Benchmarking Environment parallel cluster configuration 70 redhat 61 nodes 7 Intel STL2 (2 x PIII 1GHz, 512MB) 31 ASUS P2B-D ( 2 x PIII 600MHz, 512MB) 15 Celsius 620 (2 x PIII, 550MHz, 512MB) the rest – Kayak 450 Mhz (2 x PIII, 450Mhz, 128MB) reference sequential machine pcgeant2 (2x Xeon 1700Mhz, 1GB) notice different CPU speeds and memory size

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Scalability Test – Job Time not normalized execution time: average gain 15 times

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Normalized Efficiency normalized efficiency: average real gain ~30 times

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Benchmarking Commentary non-exclusive access to interactive machines 'load-noise' background, unpredictible load peaks different CPU and RAM on nodes AFS used to fetch physics config data try to remove the noise: repeat simulations many times to get the correct mean work at night and off-peak hours (what about US people using CERN computing facilities ?) etc... interpretation of results scaling factors for different CPU speeds results agree with expectations

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Summary

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Scalability Tests prototype deployment of Geant4-DIANE proved significant performance improvement scalability tests: 140 Mio Events 70 nodes in the cluster 1 hour total parallel execution putting together DIANE and Geant4 is fairly easy done in few days...

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Easy to use user-friendliness application developer (e.g. Geant4 simulation) is shielded from complexity of underlying technology not affecting the original code of application standalone and distributed cases is the same code good separation of the subsystems application does not need to know that it runs in distributed environment... the distributed framework (DIANE) does not need to care about what actions application performs internally

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Universally Applicable DIANE is application independent easy to customize and use in applications other than Geant4 e.g. it has been originally developed for ntuple analysis DIANE may bridge applications to the GRID world without necessarily waiting for fully-fledged GRID infrastructure to become available with smooth transition to GRID technologies as they become available DIANE and distributed computing technology may be applied in a variety of other scientific/research domains

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 In progress: Optimization time of job execution = slowest machine......or most loaded one at the moment often had to wait a long time for last worker to finish example of customization exploit dual-processor mode use larger number of smaller workers fast machines run workers sequentially many times benchmark in dedicated cluster

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 In progress: Medical Applications plan to run Geant4 simulation for radiotherapy in couple of days new possibilities: precise MC-based treatment planning FAST small hospitals may access distributed resources worldwide

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 References more informarion: cern.ch/diane aida.freehep.org cern.ch/anaphe

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 The end

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 From sequential to parallel simulation

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Structure of the simulation initialization phase (constant) load ~10-15 Mb of physics tables, config data etc. reference sequential machine: ~ 4 minutes (user time) cluster nodes: ~ 5-6 minutes beamOn ~ f( event number ) small job: 1-5 Mio events medium job: Mio events big job: > 50 Mio events

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Reproducability initial seed of the random engine make sure that every parallel simulation starts with a seed uniquely determined by the job's initial seed number of times engine is used depends on the initial seed make sure that correlations between the workers' seeds are avoided our solution: use two uncorrelated random engines one to generate a table of initial seeds (one seed for each worker) another for the simulation inside the worker

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 DIANE – G4 prototype Parallelization of Geant4 simulation is a joint project between Geant4 – DIANE – Anaphe DIANE is an R&D project in IT/API to study distributed analysis and simulation and create a prototype initiated early 2001 with very limited resources Anaphe is an analysis project supported by IT provides the analysis framework for HEP The pilot programme includes G4 simulation which produces AIDA/Anaphe histograms Collaboration started late spring 2002

DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Reproducability parameters which need to be fixed to reproduce the simulation: total number of events initial seed... but also: number of workers number of events per worker