Xrootd Monitoring Atlas Software Week CERN November 27 – December 3, 2010 Andrew Hanushevsky, SLAC.

Slides:



Advertisements
Similar presentations
Potential Data Access Architectures using xrootd OSG All Hands Meeting Harvard University March 7-11, 2011 Andrew Hanushevsky, SLAC
Advertisements

Xrootd Roadmap Atlas Tier 3 Meeting University of Chicago September 12-13, 2011 Andrew Hanushevsky, SLAC
Xrootd Update OSG All Hands Meeting University of Nebraska March 19-23, 2012 Andrew Hanushevsky, SLAC
1 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Mass Storage For BaBar at SLAC Andrew Hanushevsky Stanford.
M ONITORING SERVER PERFORMANCE Unit objectives Use Task Manager to monitor server performance and resource usage Use Event Viewer to identify and troubleshoot.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
Scalla Back Through The Future Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 8-April-10
XRootD Roadmap To Start The Second Decade Root Workshop Saas-Fee March 11-14, 2013 Andrew Hanushevsky, SLAC
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Experience of xrootd monitoring for ALICE at RDIG sites G.S. Shabratova JINR A.K. Zarochentsev SPbSU.
The Next Generation Root File Server Andrew Hanushevsky Stanford Linear Accelerator Center 27-September-2004
Scalla/xrootd Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 19-August-2009 Atlas Tier 2/3 Meeting
ATLAS Off-Grid sites (Tier-3) monitoring A. Petrosyan on behalf of the ATLAS collaboration GRID’2012, , JINR, Dubna.
M1G Introduction to Database Development 6. Building Applications.
Xrootd Demonstrator Infrastructure OSG All Hands Meeting Harvard University March 7-11, 2011 Andrew Hanushevsky, SLAC
Scalla/xrootd Andrew Hanushevsky, SLAC SLAC National Accelerator Laboratory Stanford University 19-May-09 ANL Tier3(g,w) Meeting.
Scalla/xrootd Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 29-October-09 ATLAS Tier 3 Meeting at ANL
HPS Online Software Discussion Jeremy McCormick, SLAC Status and Plans.
Ramiro Voicu December Design Considerations  Act as a true dynamic service and provide the necessary functionally to be used by any other services.
1 Wenguang WangRichard B. Bunt Department of Computer Science University of Saskatchewan November 14, 2000 Simulating DB2 Buffer Pool Management.
March 6, 2009Tofigh Azemoon1 Real-time Data Access Monitoring in Distributed, Multi Petabyte Systems Tofigh Azemoon Jacek Becla Andrew Hanushevsky Massimiliano.
Scalla/xrootd Introduction Andrew Hanushevsky, SLAC SLAC National Accelerator Laboratory Stanford University 6-April-09 ATLAS Western Tier 2 User’s Forum.
Xrootd Monitoring for the CMS Experiment Abstract: During spring and summer 2011 CMS deployed Xrootd front- end servers on all US T1 and T2 sites. This.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Connect. Communicate. Collaborate Place organisation and project logos in this area AAIEye – A Monitoring Tool For AAI’s Mika Suvanto, CSC TNC 2008, Bruges.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
ROOT-CORE Team 1 PROOF xrootd Fons Rademakers Maarten Ballantjin Marek Biskup Derek Feichtinger (ARDA) Gerri Ganis Guenter Kickinger Andreas Peters (ARDA)
STATUS OF DCACHE N2N AND MONITORING REPORT I. CURRENT SITUATION xrootd4j is a part of dCache implemented in such a way that each change requires new dCache.
David Adams ATLAS DIAL/ADA JDL and catalogs David Adams BNL December 4, 2003 ATLAS software workshop Production session CERN.
Accelerating Debugging In A Highly Distributed Environment CHEP 2015 OIST Okinawa, Japan April 28, 2015 Andrew Hanushevsky, SLAC
Performance and Scalability of xrootd Andrew Hanushevsky (SLAC), Wilko Kroeger (SLAC), Bill Weeks (SLAC), Fabrizio Furano (INFN/Padova), Gerardo Ganis.
Xrootd Present & Future The Drama Continues Andrew Hanushevsky Stanford Linear Accelerator Center Stanford University HEPiX 13-October-05
Xrootd Update Alice Tier 1/2 Workshop Karlsruhe Institute of Technology (KIT) January 24-26, 2012 Andrew Hanushevsky, SLAC
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
ATLAS XRootd Demonstrator Doug Benjamin Duke University On behalf of ATLAS.
Scalla/xrootd Andrew Hanushevsky, SLAC SLAC National Accelerator Laboratory Stanford University 08-June-10 ANL Tier3 Meeting.
Scalla Advancements xrootd /cmsd (f.k.a. olbd) Fabrizio Furano CERN – IT/PSS Andrew Hanushevsky Stanford Linear Accelerator Center US Atlas Tier 2/3 Workshop.
Xrootd Monitoring and Control Harsh Arora CERN. Setting Up Service  Monalisa Service  Monalisa Repository  Test Xrootd Server  ApMon Module.
XRootD & ROOT Considered Root Workshop Saas-Fee September 15-18, 2015 Andrew Hanushevsky, SLAC
XROOTD AND FEDERATED STORAGE MONITORING CURRENT STATUS AND ISSUES A.Petrosyan, D.Oleynik, J.Andreeva Creating federated data stores for the LHC CC-IN2P3,
PERFORMANCE AND ANALYSIS WORKFLOW ISSUES US ATLAS Distributed Facility Workshop November 2012, Santa Cruz.
Performance Testing Test Complete. Performance testing and its sub categories Performance testing is performed, to determine how fast some aspect of a.
Scalla Authorization xrootd /cmsd Andrew Hanushevsky SLAC National Accelerator Laboratory CERN Seminar 10-November-08
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
Scalla In’s & Out’s xrootdcmsd xrootd /cmsd Andrew Hanushevsky SLAC National Accelerator Laboratory OSG Administrator’s Work Shop Stanford University/SLAC.
SRM Space Tokens Scalla/xrootd Andrew Hanushevsky Stanford Linear Accelerator Center Stanford University 27-May-08
Scalla As a Full-Fledged LHC Grid SE Wei Yang, SLAC Andrew Hanushevsky, SLAC Alex Sims, LBNL Fabrizio Furano, CERN SLAC National Accelerator Laboratory.
Scalla + Castor2 Andrew Hanushevsky Stanford Linear Accelerator Center Stanford University 27-March-07 Root Workshop Castor2/xrootd.
MND review. Main directions of work  Development and support of the Experiment Dashboard Applications - Data management monitoring - Job processing monitoring.
Federated Data Stores Volume, Velocity & Variety Future of Big Data Management Workshop Imperial College London June 27-28, 2013 Andrew Hanushevsky, SLAC.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
Enabling Grids for E-sciencE CMS/ARDA activity within the CMS distributed system Julia Andreeva, CERN On behalf of ARDA group CHEP06.
Status of tests in the LCG 3D database testbed Eva Dafonte Pérez LCG Database Deployment and Persistency Workshop.
CERN - IT Department CH-1211 Genève 23 Switzerland CASTOR F2F Monitoring at CERN Miguel Coelho dos Santos.
Testing Infrastructure Wahid Bhimji Sam Skipsey Intro: what to test Existing testing frameworks A proposal.
Data Analysis w ith PROOF, PQ2, Condor Data Analysis w ith PROOF, PQ2, Condor Neng Xu, Wen Guan, Sau Lan Wu University of Wisconsin-Madison 30-October-09.
WLCG Transfers Dashboard A unified monitoring tool for heterogeneous data transfers. Alexandre Beche.
Database Issues Peter Chochula 7 th DCS Workshop, June 16, 2003.
ATLAS Off-Grid sites (Tier-3) monitoring A. Petrosyan on behalf of the ATLAS collaboration GRID’2012, , JINR, Dubna.
New Features of Xrootd SE Wei Yang US ATLAS Tier 2/Tier 3 meeting, University of Texas, Arlington,
SLACFederated Storage Workshop Summary Andrew Hanushevsky SLAC National Accelerator Laboratory April 10-11, 2014 SLAC.
WLCG Transfers monitoring EGI Technical Forum Madrid, 17 September 2013 Pablo Saiz on behalf of the Dashboard Team CERN IT/SDC.
Federating Data in the ALICE Experiment
Glast Collaboration Data Server and Data Catalog
SLAC National Accelerator Laboratory
Data Federation with Xrootd Wei Yang US ATLAS Computing Facility meeting Southern Methodist University, Oct 11-12, 2011.
Simulation use cases for T2 in ALICE
Scalla/XRootd Advancements
Presentation transcript:

Xrootd Monitoring Atlas Software Week CERN November 27 – December 3, 2010 Andrew Hanushevsky, SLAC

December 1, 2010 Outline Introduction to xrootd monitoring What’s available How it works What you probably really want 2 Atlas Software Week

December 1, 2010 What is xrootd monitoring? Server-side services that report information Two services configured via the xrootd config file Real-time detailed monitoring xrootd.monitor directive Periodic summary monitoring xrd.report directive Details in “Xrd/Xrootd Configuration Reference” Atlas Software Week

December 1, 2010 Why Two Services? Real-time and periodic data vastly different Each service designed for unique data requirements Real-time data is fast paced and continuous Setting this up is not for the faint-of-heart Periodic data is rather leisurely but bursty Has few requirements and is relatively easy to setup It’s likely you will only use periodic summaries As we shall see as we go on 4 Atlas Software Week

December 1, 2010 Real-Time Monitoring Flow 5 Atlas Software Week xrootd Requests Client Events Clientxrootd Data Collector DB RT Data Display Data Collector DB Detailed Stream Meta Stream Start Session sessionId, user, pid, client, server, start T, authinfo FRM Staging stageid, user, pid, client, file path, stage T, duration, server Open File fileid, user, pid, client, server, file path, open T File I/O fileid, I/O length, offset, window T Close File fileid, bytes read, bytes written, close T App data user, pid, client, server, application specific data End Session sessionId, duration, end T Decoder Loader Decoder Loader DB Query Rendering File

December 1, 2010 Real-Time Data Handling Is Hard Potentially > 50 MB/Sec monitoring stream Needs fast data collector (i.e. monitoring server) Part of the base package Complex inter-related data Needs sophisticated tools to probe relationships Base package interfaces with mySQL Provides basic web-interface rendering of data This is a lot of work to put up! 6 Atlas Software Week

December 1, 2010 But Can Get Very Good Insights TopPerformersTable 7 Atlas Software Week

December 1, 2010 Summary Monitoring Is Easier Summary data periodically reported Very large amount of data available You pick which is to be reported by category Use the xrd.report directive mpxstats Centrally collect it via provided mpxstats tool Merges and converts xml streams to keyword/value pairs Feed data into your favorite monitoring system Ganglia, GRIS, Nagios, MonALISA, etc 8 Atlas Software Week

December 1, 2010 Summary Monitoring Data Flow 9 Atlas Software Week DataServers Monitoring Host mpxstats xrd.report monhost:1999 all every 15s monhost:1999 ganglia Xrootd version Xrootd version CPU usage CPU usage Bytes In/Out Bytes In/Out Number of Connections Number of Connections Number of delays Number of delays Number of files open Number of files open Space by space token Space by space token Space by volume Space by volume Number of transactions Number of transactions Number of threads Number of threads Number of requests/type Number of requests/type Typical Data Collected This is a centralized data flow; you can also do a distributed flow! Here mpxstats runs on each data server and you route data to localhost.

December 1, 2010 Easy To Render Basic Metrics 10 Atlas Software Week

December 1, 2010 Summary xrootd provides a wealth of monitoring data From super detailed to basic summaries Your needs will determine what you collect We suggest sticking with periodic summary data However, you must have a monitoring system Ganglia, Gris, Nagios, MonaLisa, or other 11 Atlas Software Week

December 1, 2010 Acknowledgements 12 Atlas Software Week Software Contributors Alice: Derek Feichtinger ATLAS: Charles Waldman, Wei Yang CERN: Fabrizio Furano, Lukasz Janyst, Andreas Peters CMS: Brian Bockelman Fermi/GLAST: Tony Johnson (Java) FZK: Artem Trunov BeStMan LBNL: Alex Sim, Junmin Gu, Vijaya Natarajan (BeStMan team) LSST: Daniel Wang Root: Gerri Ganis, Beterand Bellenet, Fons Rademakers SLAC: Tofigh Azemoon, Andrew Hanushevsky,Wilko Kroeger Operational Collaborators BNL, CERN, FZK, IN2P3, OSG, SLAC, UC, UTA, UVIC, UWisc Partial Funding US Department of Energy Contract DE-AC02-76SF00515 with Stanford University