PanDA Update Kaushik De Univ. of Texas at Arlington XRootD Workshop, UCSD January 27, 2015.

Slides:



Advertisements
Similar presentations
Resource Management §A resource can be a logical, such as a shared file, or physical, such as a CPU (a node of the distributed system). One of the functions.
Advertisements

Integrating Network Awareness in ATLAS Distributed Computing Using the ANSE Project J.Batista, K.De, A.Klimentov, S.McKee, A.Petroysan for the ATLAS Collaboration.
Integrating Network and Transfer Metrics to Optimize Transfer Efficiency and Experiment Workflows Shawn McKee, Marian Babik for the WLCG Network and Transfer.
MultiJob PanDA Pilot Oleynik Danila 28/05/2015. Overview Initial PanDA pilot concept & HPC Motivation PanDA Pilot workflow at nutshell MultiJob Pilot.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES News on monitoring for CMS distributed computing operations Andrea.
I/O Systems ◦ Operating Systems ◦ CS550. Note:  Based on Operating Systems Concepts by Silberschatz, Galvin, and Gagne  Strongly recommended to read.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
ATLAS Off-Grid sites (Tier-3) monitoring A. Petrosyan on behalf of the ATLAS collaboration GRID’2012, , JINR, Dubna.
ATLAS federated xrootd monitoring requirements Rob Gardner July 26, 2012.
CERN IT Department CH-1211 Geneva 23 Switzerland t The Experiment Dashboard ISGC th April 2008 Pablo Saiz, Julia Andreeva, Benjamin.
Integration Program Update Rob Gardner US ATLAS Tier 3 Workshop OSG All LIGO.
Enabling Grids for E-sciencE Overview of System Analysis Working Group Julia Andreeva CERN, WLCG Collaboration Workshop, Monitoring BOF session 23 January.
Tier 3 Data Management, Tier 3 Rucio Caches Doug Benjamin Duke University.
PhysX CoE: LHC Data-intensive workflows and data- management Wahid Bhimji, Pete Clarke, Andrew Washbrook – Edinburgh And other CoE WP4 people…
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
PanDA A New Paradigm for Computing in HEP Kaushik De Univ. of Texas at Arlington NRC KI, Moscow January 29, 2015.
FAX UPDATE 1 ST JULY Discussion points: FAX failover summary and issues Mailing issues Panda re-brokering to sites using FAX cost and access Issue.
FAX UPDATE 26 TH AUGUST Running issues FAX failover Moving to new AMQ server Informing on endpoint status Monitoring developments Monitoring validation.
DDM-Panda Issues Kaushik De University of Texas At Arlington DDM Workshop, BNL September 29, 2006.
PanDA Summary Kaushik De Univ. of Texas at Arlington ADC Retreat, Naples Feb 4, 2011.
Efi.uchicago.edu ci.uchicago.edu Towards FAX usability Rob Gardner, Ilija Vukotic Computation and Enrico Fermi Institutes University of Chicago US ATLAS.
PanDA: Exascale Federation of Resources for the ATLAS Experiment
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
ROOT and Federated Data Stores What Features We Would Like Fons Rademakers CERN CC-IN2P3, Nov, 2011, Lyon, France.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Overview of STEP09 monitoring issues Julia Andreeva, IT/GS STEP09 Postmortem.
DDM Monitoring David Cameron Pedro Salgado Ricardo Rocha.
Efi.uchicago.edu ci.uchicago.edu FAX status developments performance future Rob Gardner Yang Wei Andrew Hanushevsky Ilija Vukotic.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
MultiJob pilot on Titan. ATLAS workloads on Titan Danila Oleynik (UTA), Sergey Panitkin (BNL) US ATLAS HPC. Technical meeting 18 September 2015.
Post-DC2/Rome Production Kaushik De, Mark Sosebee University of Texas at Arlington U.S. Grid Phone Meeting July 13, 2005.
Network awareness and network as a resource (and its integration with WMS) Artem Petrosyan (University of Texas at Arlington) BigPanDA Workshop, CERN,
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
PD2P The DA Perspective Kaushik De Univ. of Texas at Arlington S&C Week, CERN Nov 30, 2010.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
PanDA Status Report Kaushik De Univ. of Texas at Arlington ANSE Meeting, Nashville May 13, 2014.
XROOTD AND FEDERATED STORAGE MONITORING CURRENT STATUS AND ISSUES A.Petrosyan, D.Oleynik, J.Andreeva Creating federated data stores for the LHC CC-IN2P3,
PANDA: Networking Update Kaushik De Univ. of Texas at Arlington SC15 Demo November 18, 2015.
PanDA & BigPanDA Kaushik De Univ. of Texas at Arlington BigPanDA Workshop, CERN October 21, 2013.
PERFORMANCE AND ANALYSIS WORKFLOW ISSUES US ATLAS Distributed Facility Workshop November 2012, Santa Cruz.
Julia Andreeva on behalf of the MND section MND review.
SDN Provisioning, next steps after ANSE Kaushik De Univ. of Texas at Arlington US ATLAS Planning, CERN June 29, 2015.
Conclusions on Monitoring CERN A. Read ADC Monitoring1.
FAX UPDATE 12 TH AUGUST Discussion points: Developments FAX failover monitoring and issues SSB Mailing issues Panda re-brokering to FAX Monitoring.
MND review. Main directions of work  Development and support of the Experiment Dashboard Applications - Data management monitoring - Job processing monitoring.
Global ADC Job Monitoring Laura Sargsyan (YerPhI).
Efi.uchicago.edu ci.uchicago.edu Data Federation Strategies for ATLAS using XRootD Ilija Vukotic On behalf of the ATLAS Collaboration Computation and Enrico.
1. 2 Overview Extremely short summary of the physical part of the conference (I am not a physicist, will try my best) Overview of the Grid session focused.
Efi.uchicago.edu ci.uchicago.edu Ramping up FAX and WAN direct access Rob Gardner on behalf of the atlas-adc-federated-xrootd working group Computation.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Efi.uchicago.edu ci.uchicago.edu Storage federations, caches & WMS Rob Gardner Computation and Enrico Fermi Institutes University of Chicago BigPanDA Workshop.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Network integration with PanDA Artem Petrosyan PanDA UTA,
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Future of Distributed Production in US Facilities Kaushik De Univ. of Texas at Arlington US ATLAS Distributed Facility Workshop, Santa Cruz November 13,
WLCG Operations Coordination report Maria Alandes, Andrea Sciabà IT-SDC On behalf of the WLCG Operations Coordination team GDB 9 th April 2014.
ATLAS Distributed Computing ATLAS session WLCG pre-CHEP Workshop New York May 19-20, 2012 Alexei Klimentov Stephane Jezequel Ikuo Ueda For ATLAS Distributed.
PanDA & Networking Kaushik De Univ. of Texas at Arlington ANSE Workshop, CalTech May 6, 2013.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Monitoring the Readiness and Utilization of the Distributed CMS Computing Facilities XVIII International Conference on Computing in High Energy and Nuclear.
PanDA Configurator and Network Aware Brokerage Fernando Barreiro Megino, Kaushik De, Tadashi Maeno 14 March 2015, US ATLAS Distributed Facilities Meeting,
PD2P, Caching etc. Kaushik De Univ. of Texas at Arlington ADC Retreat, Naples Feb 4, 2011.
Efi.uchicago.edu ci.uchicago.edu Sharing Network Resources Ilija Vukotic Computation and Enrico Fermi Institutes University of Chicago Federated Storage.
PanDA & Networking Kaushik De Univ. of Texas at Arlington UM July 31, 2013.
U.S. ATLAS Grid Production Experience
Future of WAN Access in ATLAS
Outline Benchmarking in ATLAS Performance scaling
PanDA in a Federated Environment
The ADC Operations Story
Univ. of Texas at Arlington BigPanDA Workshop, ORNL
Presentation transcript:

PanDA Update Kaushik De Univ. of Texas at Arlington XRootD Workshop, UCSD January 27, 2015

Overview  At last XRootD workshop (OSG-AHM, April 2014) we showed first results from overflow jobs – looked promising but high failure rates caused us to switch off further testing  Failover jobs were working very well  In this talk I will focus on new results and future plans  This work benefits from synergy with networking projects  While XRootD is an important component of US ATLAS core program, additional funding from DOE and NSF was crucial to success of this work  DOE-ASCR/HEP funded BigPanda project (BNL-UTA-ANL)  NSF funded CC-NIE:ANSE project (CalTech-UM-Vanderbilt-UTA)  Future work will be funded through Russian MegaPanDA grant January 27, 2015Kaushik De2

FAX and PanDA  All WLCG ATLAS sites are becoming FAX enabled  But we are not replacing FTS with FAX – they are complementary  Gradually shift some PanDA workload to FAX  FAX allows more interesting use cases – focus on new capabilities in PanDA that are only possible through FAX  FAX deployment, robust operations and new software development must evolve together  Many details already in other ATLAS talks at this meeting  ATLAS Status and Plans: Ilija Vukotic (UC)  ATLAS data formats and FAX: Doug Benjamin (Duke)  ATLAS experience: Rob Gardner (UC)  Event Service: Vakho Tsulaia (LBNL)  And others – I will try not to repeat their excellent content January 27, 2015Kaushik De3

PanDA Scale January 27, M jobs every month at >150 sites Processed more than one exabytes in 2013 and 2014 Kaushik De4

Case 1: Rescuing Failed Transfers  PanDA pilot retries FTS transfer of input files a few times for production  For analysis jobs, if local file cannot be read, jobs fail  For sites with FAX enabled, why not try FAX to rescue job?  This was implemented and working well – snapshot of past 6 hours January 27, 2015Kaushik De5

Case 2: Faster User analysis  Also first use case for network integration with PanDA  Goal - reduce waiting time for user jobs  User analysis jobs normally go to sites with local input data  This can occasionally lead to long wait times (jobs are re-brokered if possible, or PD2P data caching will make more copies eventually to reduce congestion)  While nearby sites with good network access may be idle  Brokerage uses concept of ‘nearby’ sites  Use cost metric generated with Hammercloud tests  Calculate weight based on usual brokerage criteria (availability of CPU resources, data location, release…) plus new network transfer cost  Jobs will be sent to the site with best overall weight  Throttling is used to manage load on network January 27, 2015Kaushik De6

Network Data Repositories  Native data repositories  Historical data stored from collectors  SSB – site status board for sonar and perfSonar data  FAX data is kept independently and uploaded  AGIS (ATLAS Grid Information System)  Most recent / processed data only – updated periodically  Mixture of push/pull – depending on source of data  schedConfigDB  Internal Oracle DB used by PanDA for fast access  Uses standard ATLAS collector January 27, 2015Kaushik De7

January 27, 2015Kaushik De8

Jobs using FAX for remote access January 27, 2015Kaushik De9 About 4-10% of jobs Access Data Remotely Through Federated XRootD Remote site Is selected Based on Network Performance Data since October 2014

Early Example from Oct, 2014 January 27, 2015Kaushik De10 How do we measure success? Let’s look at example task.

Jobs from task on oct 3 January 27, 2015Kaushik De11

Job wait times for example task January 27, 2015Kaushik De12

Other Tasks and other Metrics?  There are three metrics to quantify performance  Average wait times with/without FAX  Execution time of jobs – averages and tails  Success rate of jobs with/without FAX  We do not yet have comprehensive analytics  Ilija is exporting and studying PanDA job data in Hadoop  Future optimizations  Algorithm has many tunable parameters  Work is underway (for example, site black-listing)  Complex optimization problem over three dimensions  We have moved from development to operations  Which is always harder  ATLAS scale is challenging - >150 sites, millions jobs/day January 27, 2015Kaushik De13

Measure of Success January 27, 2015Kaushik De14 Many sites used for overflow Failure rate is manageable

Maybe not so Rosy? January 27, 2015Kaushik De15

Conclusions for Case 2  Network data collection working well  Additional algorithms to combine network data will be tried  HC tests working well – PS data not robust yet  PanDA brokerage worked well  Achieved goal of reducing wait time – though anecdotally  Well balanced local vs remote access  Need fine tuning – we have a lot of data now (4M jobs)  We have overflow working for FAX  We need deeper study to optimize January 27, 2015Kaushik De16

Case 3: FAX for Data Transfer  Production jobs move data  PanDA transfers input data for MC production, reprocessing, group production (DPD making), and other privileged processing  Transfer is done asynchronously, if needed (preference given to sites which already have data by brokerage)  Once transfer is completed, PanDA receives callback, which activates processing  Work started to send callback to PanDA frrom FAX  Once callback is available, we will instrument a few sites January 27, 2015Kaushik De17

Other Possibilities  FAX for data caching  PanDA will send jobs to sites which may have data (not guaranteed)  Workload aware data access  PanDA can choose FAX direct access for IO intensive workloads, and copy data for CPU intensive workloads  Optimized ROOT access  pRoot?  FAX for writing output files  So far, all use cases involve read access  PanDA can use the flexibility of writing via FAX  FAX for event service  Event service needs object store – FAX can provide this January 27, 2015Kaushik De18

Conclusion  Finally, good results from 2 out of our 3 use cases discussed at previous FAX meetings  Lot of work to move from development to operations  Ongoing work and future work  We need to focus on operations and analytics  Many new possibilities  FAX is now integral part of PanDA WMS January 27, 2015Kaushik De19