Download presentation
Presentation is loading. Please wait.
Published bySophia Bates Modified over 9 years ago
1
PERFORMANCE AND ANALYSIS WORKFLOW ISSUES US ATLAS Distributed Facility Workshop 13-14 November 2012, Santa Cruz
2
IMPORTANCE OF ANALYSIS JOBS Number of analysis jobs are increasing Production jobs are mostly CPU limited, well controlled, hopefully optimized and can be monitored through other already existing system Analysis jobs we know very little about and potentially could: be inefficient, wreck havoc at storage elements, networks. Twice failure rate of production jobs 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 2
3
ANALYSIS QUEUES PERFORMANCE Idea Find what is performance of ATLAS analysis jobs on the grid There is no framework that everybody uses, that could be instrumented Understand numbers: each site has it’s hard limits in terms of storage, cpus, network, software. Improve ATLAS software ATLAS files, way we use them Site’s configurations Requirements: Monitoring framework Tests simple, realistic, accessible, versatile as possible Running on most of the resources we have Fast turn around Test codes that are “recommended way to do it” Web interface for most important indicators 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 3
4
TEST FRAMEWORK 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 4 HammerCloud ORACLE DB CERN Results ORACLE DB CERN Results WEB site SVN configuration, test scripts Continuous Job performance Generic ROOT IO scripts Realistic analysis jobs Site performance Site optimization One-off new releases (Athena, ROOT) new features, fixes All T2D sites (currently 42 sites) Large number of monitored parameters Central database Wide range of visualization tools Continuous Job performance Generic ROOT IO scripts Realistic analysis jobs Site performance Site optimization One-off new releases (Athena, ROOT) new features, fixes All T2D sites (currently 42 sites) Large number of monitored parameters Central database Wide range of visualization tools
5
TEST FRAMEWORK 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 5 Pilot numbers obtained from panda db 5-50 jobs per day per site Each job runs at least 24 tests 5 read modes + 1 full analysis job Over 4 different files Takes data on machine status Cross reference to Panda DB Currently 2 million results in DB WEB site http://ivukotic.web.cern.ch/ivukotic/HC/index.asp
6
SUMMARY RESULTS Setup times 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 6
7
SUMMARY RESULTS Stage-in 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 7 Space for improvement 60 s = 41 MB/s The Fix
8
SUMMARY RESULTS Execution time 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 8 GPFS not mounted – can’t run in direct mode
9
SUMMARY RESULTS Stage out 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 9
10
SUMMARY RESULTS Total time = setup + stage in + exec + stage out [s] – as measured by pilot 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 10
11
SUMMARY – GOING DEEPER CPU efficiency Measures only event loop Defined as CPU time / WALL time Keep in mind – very slow machine can have very high CPU eff. All you want to do is make it as high as possible 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 11 FACTS: 1.Unless doing bootstrapping or some weird calculation, users code is negligible compared to unzipping. 2.ROOT can unzip at 40MB/s
12
SUMMARY – GOING DEEPER CPU efficiency 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 12 Direct access site
13
13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 13 GOING DEEPER - CASE OF SWITCH STACKING Test files are local to both UC and IU sites. Lower band is IU. Only part of the machines are affected. (the best ones)
14
We check CPU eff. VS. Load Network in/out Memory Swap 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 14 GOING DEEPER - CASE OF SWITCH STACKING
15
CASE OF SWITCH STACKING 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 15 Machines can do much better as seen in copy2scratch mode. Drained node as bad as busy one. Manual checks show connections to servers much bellow 1Gbps. Stack performance depend on: its configuration (software) what is connected where Optimal switch stacking not exactly trivial. I suspect a lot of sites have the same issues. NET2 and BNL show very similar pattern. Will be investigated till the bottom.
16
Finally Two big issues discovered. Just that was worth the effort Bunch of smaller problems with queues, misconfigurations found and solved FUTURE Fixing remaining issues Investigate Virtual Queues Per site web interface Automatic procedure to follow performance Automatic mailing Investigating non-US sites 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 16
17
13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 17 WORKFLOW ISSUES For most users this is the workflow: Skimming/slimming data usually prun and no complex code often filter_and_merge.py Merging data only part of people do it unclear how to do it on the grid moving small files around very inefficient Getting data locally DaTRI requests to USA processed slowly Most people dq2-get Storing it locally Not much space in tier-3’s Accessing data from localgroupdisk Analyzing data Mostly local queues Rarely proof People willing to wait for few hours and manually merge results
18
SLIM SKIM SERVICE Idea Establish service to which users submit parameters of their skim&slim job, uses opportunistically CPU’s and FAX as data source and provides optimized dataset. Practically WebUI to submit request and follow job progress Oracle DB for a backend Currently UC3 will be used for processing. Output data will be dq2-put into MWT2 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 18 Work started. Performance and turn around time are what will make or brake this service. Work started. Performance and turn around time are what will make or brake this service.
19
APPENDIX The Fix timer_command.py is part of the pilot3 code. Used very often in all of the transforms. Serves to start any command as a subprocess and kills it if not finished before a given timeout. Not exactly trivial. For some commands was waiting 60 seconds even when command finished. Was also trying to close all the possible file descriptors before executing child process. That could take from 0.5s – few tens of seconds depending on site’s settings. Fixed in the last pilot version. Total effect estimate: Quarter of computing time is spent on analysis jobs Average analysis job is less than 30 min. Fix speeds up job in average 3 minutes - 10% Applied to 40 Tier2’s the fix equivalent of adding one full tier2 of capacity 13/11/2012ILIJA VUKOTIC IVUKOTIC@UCHICAGO.EDU 19 BACK
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.