Download presentation
Presentation is loading. Please wait.
Published byAmy Francis Modified over 9 years ago
1
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 1 Experiences of Submitting UKDMC and LISA GEANT4 Jobs FIRST I should say thanks for the opportunity of using the growing resource of the GRID particularly for Particle Astrophysics (LISA and UKDMC) which has no direct funding to (or from) the GRIDPP It is increasingly clear we NEED use of the GRID in order to carry out accurate simulations for both LISA and Dark Matter – either to investigate signals or possible backgrounds on micro and macroscopic levels. Specifically this is processor intensive work I am a novice – only 5 weeks usage of the GRID
2
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 2 Experiences of Submitting UKDMC and LISA GEANT4 Jobs 1.Experimental configuration a)Dark Matter b)LISA 2.Output of >100 Jobs 3.Benefits 4.Comments on Functionality and UI
3
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 3 UKDMC Experiment
4
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 4 Two-Phase Liquid Xenon, ZEPLIN III ZEPLIN III, our near future detector will offer extreme levels of signal-to-background discrimination. Interpretation will only be possible with full Monte Carlo studies (a novelty for DM)
5
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 5 Prototype Simulation: Full Lab Geometry
6
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 6 GEANT 4 Due to the need to develop Monte Carlo simulations for Dark Matter experiments, I have become involved in the development of Geant4 – particularly exploiting the low energy, radioactivity decay and neutron extensions of the toolkit (see advanced example DMX within the release package) From basic simulations of our prototype system it is clear that greater computing power is required in order to produce high statistics and accurately model our detectors. In addition it is envisaged to develop a simulation of the underground environment, UNEX, to give accurate spectra and particle types within the experimental area. Furthermore, Imperial is also involved in LISA a gravitational wave experiment, where charging of proof masses becomes critical. To simulate the charging rate requires large number of cosmic events with rare hadronic showers resulting in residual charge
7
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 7 One High Energy event: LXe GXe PMT mirror source
8
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 8 roomelastic inelasticoutside Neutron s
9
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 9 LISA/STEP Gravitational wave experiment STEP = test of the equivalence principle Both rely on floating proof masses with no electrical connections prone to charging effects from cosmic rays However, charging rate is relatively low (~1 in 5000 particles)
10
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 10 LISA Geometry and Geant4 Images
11
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 11 LISA Geometry and Geant4 Images
12
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 12 OUTPUT from Grid Running Over 100 jobs have been run on the grid to try and estimate the charging rate in the proof mass and test different cuts and processes within Geant4. 1.9 million events have been run in ~300 hours of CPU time The preliminary outcome is as follows:
13
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 13 6secs and Convergence of Charging Rate
14
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 14 6secs and Convergence of Charging Rate Initial indications of charging rate
15
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 15 GRID JOB SUBMISSION – My Experience With the auspicious title of DataGrid User….. After running >100 jobs I have had some experience of running of jobs on the DataGrid – mostly good However, there are a few things that if implemented would be useful, although their unavailability may just be the youthful nature of the GRID and therefore already present, at least in the design… (some of which are due to my ignorance)
16
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 16 Things Missing, apparently (1) 1.Status of job events run, nearcompletion? 2.Run-time Partial grab of output files check on job: RB release 1.4 (Dave’s talk) 3.Length of identifier – cumbersome 4.Saving identifiers to file ease of management of many jobs 5.Request output to be saved to file automatically when job completed 6.Proxy expiration and file loss – can protect against it, but can occur 7.File back up – prevent losses when things crash, and therefore reduce number of repeat jobs 8.Job clearing and file clearing – particularly if job crashes/disappears
17
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 17 Things Missing, apparently (2) 10.Diagnostics memory usage (single event leaks) – max/average CPU Time – can access at runtime with Globus Disc access = efficiency of local staging, etc… 11.Forced killing of jobs? Clearing of files or keeping of partial files – cancelling jobs loses everything 12.Run time limit/Disk Usage/Memory Usage in case of problems/diagnostic? 14.Node limit Batch script to run jobs sequentially without clogging up the farm – from proxy-request? 15.Shared Disc for data? – Input files are ~500 Mbytes and copied 32 times…
18
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 18 Things Missing, apparently (3) 16.What decides resource management? Queued at IC or at RAL – speed of processor? Disc transfer time? 17.Jobs cleared before get output (RB dies) 18.Housekeeping/cleaning of tmp files 19.Script to save output to your account without 3 rd party access? 20.Prone to abuse? 21.Tidy up – clear up dangling jobs and tmp files for a given user
19
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 19 Things Missing, apparently (4) 22.Reliability? 23.Compilers and inter-site homogeneity IC = egcs 1.1.2 whereas RAL = gcc 2.95.2 24.Level of resource available and average usage/users/CPU power – stage requests, think about optimising problem, look elsewhere 25.More nodes with my VO would naturally be helpful IC = 16 nodes (user limit 14) RAL = 8 nodes elsewhere? Apologies if some of these comments are due to me…
20
Friday May 10th 2002 Alex Howard Imperial College LondonSlide 20 Conclusions The GRID is clearly a very powerful resource that has enabled me to run a lot of jobs in a very short space of time It is clear that Dark Matter and Spacecraft charging studies at this time NEED the GRID - particularly for accurate Monte Carlo simulations of future detectors (ZEPLIN III) and Spacecraft charging rates (LISA/STEP) In running jobs some things could perhaps be more elegant/convenient to use, but on the whole it is not too difficult
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.