MONALISA MONITORING AND CONTROL Costin Grigoras. O UTLINE MonALISA services and clients Usage in ALICE Online SE discovery mechanism Data management 3.

Slides:



Advertisements
Similar presentations
SSRS 2008 Architecture Improvements Scale-out SSRS 2008 Report Engine Scalability Improvements.
Advertisements

ALICE G RID SERVICES IP V 6 READINESS
The Premier Software Usage Analysis and Reporting Toolset CELUG Presentation – May 12, 2010 LT-Live : License Tracker’s License Server Monitor.
CERN LCG Overview & Scaling challenges David Smith For LCG Deployment Group CERN HEPiX 2003, Vancouver.
A Java Architecture for the Internet of Things Noel Poore, Architect Pete St. Pierre, Product Manager Java Platform Group, Internet of Things September.
MONITORING WITH MONALISA Costin Grigoras. M ONITORING WITH M ON ALISA What is MonALISA ? MonALISA communication architecture Monitoring modules ApMon.
June 2003 Iosif Legrand MONitoring Agents using a Large Integrated Services Architecture Iosif Legrand California Institute of Technology.
Monitoring and controlling VRVS Reflectors Catalin Cirstoiu 3/7/2003.
NGOP J.Fromm K.Genser T.Levshina M.Mengel V.Podstavkov.
October 2003 Iosif Legrand Iosif Legrand California Institute of Technology.
Operating Systems Concepts 1. A Computer Model An operating system has to deal with the fact that a computer is made up of a CPU, random access memory.
F Fermilab Database Experience in Run II Fermilab Run II Database Requirements Online databases are maintained at each experiment and are critical for.
Platform as a Service (PaaS)
ALICE DATA ACCESS MODEL Outline ALICE data access model - PtP Network Workshop 2  ALICE data model  Some figures.
Chapter 3.1:Operating Systems Concepts 1. A Computer Model An operating system has to deal with the fact that a computer is made up of a CPU, random access.
G RID SERVICES IP V 6 READINESS
ALICE data access WLCG data WG revival 4 October 2013.
5 Chapter Five Web Servers. 5 Chapter Objectives Learn about the Microsoft Personal Web Server Software Learn how to improve Web site performance Learn.
Online Monitoring with MonALISA Dan Protopopescu Glasgow, UK Dan Protopopescu Glasgow, UK.
© 2006 Cisco Systems, Inc. All rights reserved.Cisco Public 1 Version 4.0 Identifying Application Impacts on Network Design Designing and Supporting Computer.
Open Science Grid The OSG Accounting System: GRATIA by Philippe Canal (FNAL) & Matteo Melani (SLAC) Mumbai, India CHEP2006.
ACAT 2003 Iosif Legrand Iosif Legrand California Institute of Technology.
Ramiro Voicu December Design Considerations  Act as a true dynamic service and provide the necessary functionally to be used by any other services.
1 Ramiro Voicu, Iosif Legrand, Harvey Newman, Artur Barczyk, Costin Grigoras, Ciprian Dobre, Alexandru Costan, Azher Mughal, Sandor Rozsa Monitoring and.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES P. Saiz (IT-ES) AliEn job agents.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES PhEDEx Monitoring Nicolò Magini CERN IT-ES-VOS For the PhEDEx.
Monitoring, Accounting and Automated Decision Support for the ALICE Experiment Based on the MonALISA Framework.
February 2006 Iosif Legrand 1 Iosif Legrand California Institute of Technology February 2006 February 2006 An Agent Based, Dynamic Service System to Monitor,
The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within.
N EWS OF M ON ALISA SITE MONITORING
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Site operations Outline Central services VoBox services Monitoring Storage and networking 4/8/20142ALICE-USA Review - Site Operations.
Update on replica management
DDM Monitoring David Cameron Pedro Salgado Ricardo Rocha.
Overview of ALICE monitoring Catalin Cirstoiu, Pablo Saiz, Latchezar Betev 23/03/2007 System Analysis Working Group.
Overview of DAQ at CERN experiments E.Radicioni, INFN MICE Daq and Controls Workshop.
Monitoring with MonALISA Costin Grigoras. What is MonALISA ?  Caltech project started in 2002
Xrootd Monitoring and Control Harsh Arora CERN. Setting Up Service  Monalisa Service  Monalisa Repository  Test Xrootd Server  ApMon Module.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
April 2003 Iosif Legrand MONitoring Agents using a Large Integrated Services Architecture Iosif Legrand California Institute of Technology.
PPDG February 2002 Iosif Legrand Monitoring systems requirements, Prototype tools and integration with other services Iosif Legrand California Institute.
JAliEn Java AliEn middleware A. Grigoras, C. Grigoras, M. Pedreira P Saiz, S. Schreiner ALICE Offline Week – June 2013.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
AliEn central services Costin Grigoras. Hardware overview  27 machines  Mix of SLC4, SLC5, Ubuntu 8.04, 8.10, 9.04  100 cores  20 KVA UPSs  2 * 1Gbps.
ALICE DATA ACCESS MODEL Outline 05/13/2014 ALICE Data Access Model 2  ALICE data access model  Infrastructure and SE monitoring.
+ AliEn site services and monitoring Miguel Martinez Pedreira.
ECHO A System Monitoring and Management Tool Yitao Duan and Dawey Huang.
Company LOGO Network Management Architecture By Dr. Shadi Masadeh 1.
October 2006 Iosif Legrand 1 Iosif Legrand California Institute of Technology An Agent Based, Dynamic Service System to Monitor, Control and Optimize Distributed.
03/09/2007http://pcalimonitor.cern.ch/1 Monitoring in ALICE Costin Grigoras 03/09/2007 WLCG Meeting, CHEP.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
DAQ & ConfDB Configuration DB workshop CERN September 21 st, 2005 Artur Barczyk & Niko Neufeld.
+ AliEn status report Miguel Martinez Pedreira. + Touching the APIs Bug found, not sending site info from ROOT to central side was causing the sites to.
MONITORING WITH MONALISA Costin Grigoras. M ON ALISA COMMUNICATION ARCHITECTURE MonALISA software components and the connections between them Data consumers.
1 R. Voicu 1, I. Legrand 1, H. Newman 1 2 C.Grigoras 1 California Institute of Technology 2 CERN CHEP 2010 Taipei, October 21 st, 2010 End to End Storage.
A System for Monitoring and Management of Computational Grids Warren Smith Computer Sciences Corporation NASA Ames Research Center.
TIFR, Mumbai, India, Feb 13-17, GridView - A Grid Monitoring and Visualization Tool Rajesh Kalmady, Digamber Sonvane, Kislay Bhatt, Phool Chand,
Amazon Web Services. Amazon Web Services (AWS) - robust, scalable and affordable infrastructure for cloud computing. This session is about:
Storage discovery in AliEn
Federating Data in the ALICE Experiment
Platform as a Service (PaaS)
Platform as a Service (PaaS)
California Institute of Technology
ALICE Monitoring
A Messaging Infrastructure for WLCG
CHAPTER 3 Architectures for Distributed Systems
Storage elements discovery
AliEn central services (structure and operation)
Publishing ALICE data & CVMFS infrastructure monitoring
Presentation transcript:

MONALISA MONITORING AND CONTROL Costin Grigoras

O UTLINE MonALISA services and clients Usage in ALICE Online SE discovery mechanism Data management /02/24 - ALICE T1/T2 Torino

M ON ALISA COMMUNICATION ARCHITECTURE MonALISA software components and the connections between them Data consumers Multiplexing layer Helps firewalled endpoints connect Registration and discovery JINI-Lookup Services Secure & Public MonALISA services Proxies Clients HL services Agents Network of Data gathering services Fully Distributed System with no Single Point of Failure /02/24 - ALICE T1/T2 Torino

G RID MONITORING ARCHITECTURE Monitoring follows the general AliEn deployment layout: one service per site collects and aggregates site-local monitoring information 4 Long History DB LCG Tools MonALISA AliEn Site ApMon AliEn Job Agent ApMon AliEn Job Agent ApMon AliEn Job Agent MonALISA LCG Site ApMon AliEn CE ApMon AliEn SE ApMon Cluster Monitor ApMon AliEn TQ ApMon AliEn Job Agent ApMon AliEn Job Agent ApMon AliEn Job Agent ApMon AliEn CE ApMon AliEn SE ApMon Cluster Monitor ApMon AliEn IS ApMon AliEn Optimizers ApMon AliEn Brokers ApMon MySQL Servers ApMon CastorGrid Scripts ApMon API Services MonALISARepository Aggregated Data rss vsz cpu time run time job slots free space nr. of files open files Queued JobAgents cpu ksi2k job status disk used processes load net In/out jobs status sockets migrated mbytes active sessions MyProxy status Alerts Actions /02/24 - ALICE T1/T2 Torino

S ERVICE COMPONENT The Service is the data collecting entity that runs on each VoBox to gather and aggregate local data Many available modules that listen for / poll data Local host monitoring (CPU, memory, network traffic, processes and sockets in each state, LM sensors, APC UPSs), log files tailing SNMP generic & specific modules; Condor, PBS, LSF and SGE (accounting & host monitoring), Ganglia Ping, tracepath, traceroute, pathload, xrootd Ciena, Optical switches (TL1); Netflow/Sflow (Force10) Calling external applications/scripts that output the values as text XDR-formatted UDP messages (ApMon) In-memory buffer for recent data Can also store persistently in a local database (not used in ALICE) Data aggregation filters Creating high-level views like cluster-wide total traffic IN/OUT, number of processes in each state … Derived data available to clients like the original stream Subscriber mechanism Clients can ask for past data and/or subscribe to arbitrary cuts in the monitoring data stream and they are notified in real time of new data /02/24 - ALICE T1/T2 Torino

A P M ON ApMon: embeddable APlication MONitoring library Lightweight library of APIsAPIs C, C++, Java, Perl, Python Send any app-specific information to ML Service(s) UDP/8884 (open XDR binary format)XDR Flexible configuration hardcoded in the app configuration file or URL Dynamic options reload while the app is running Very high throughput (50 KHz of parameters to a single service) ROOT wrapper as TMonaLisaWriterTMonaLisaWriter /02/24 - ALICE T1/T2 Torino

A P M ON ApMon optional features Background application monitoring 10 parameters / PID Used CPU & wall time, % of the machine CPU Partition stats, size of workdir, open files Memory usage (resident, virtual and %), page faults Background system monitoring parameters / host Load, CPU, memory & swap usage Network interfaces (in/out/IPs/errs) Sockets in each state, processes in each state Disk IO, swap IO /02/24 - ALICE T1/T2 Torino

A LI E N - SPECIFIC PARAMETERS Site services collect data from all local components AliEn services, proxies’ status, critical local directories on the VoBox Xrootd storage nodes (ALICE:: :: _xrootd_* ; XrdStatus) Machine parameters (CPU, load, memory, sockets, processes, network traffic, disk and swap IO) Xrootd internal parameters and per transfer details Storage space as seen by xrootd Job agents ( _JobAgent) CPU and memory usage, payload job ID, status Jobs (ALICE:: :: _Jobs) Full machine parameters (same as above) CPU and memory usage of the process tree itself (+ Si2K normalized values) Owner and masterjob / subjob IDs Payload status code (STARTING, RUNNING, SAVING) Available bandwidth, traceroute/tracepath between services /02/24 - ALICE T1/T2 Torino

A GGREGATED VALUES PER SITE Values are aggregated in various categories (*_Summary) Job summaries Number of jobs in each state (absolute and rates) min/max/avg/total resource usage (absolute and rates) Per cluster and per user Top memory consumers Job agent overview Number of JAs in each state Average TTL Queued JAs (that don’t report yet active monitoring data) Histograms of number of executed jobs per JA Aggregated Xrootd network traffic per IP class, target site name, LAN/WAN, absolute total Cluster worker nodes’ summary status Xrootd and PROOF cluster summary /02/24 - ALICE T1/T2 Torino

O PEN DATA ACCESS Adapting ML to your needs You can recycle the VoBox ML instance for your own purposes Sending extra data to it from local services or adding modules to it For example cluster monitoring with Or an ApMon-based host or application sensor Start a separate, independent service in the “alice” group In case you want to deploy custom filters / alarms Or start your own monitoring group For several sites / redundant monitoring Services and clients can belong/connect to several groups Any number of consumers can subscribe to any cut in the data stream But try to find the minimal expression that gives the data you want (>100KHz of monitoring data for 8M+ parameters) /02/24 - ALICE T1/T2 Torino

M ON ALISA CLIENTS Two important clients GUI client Interactive exploring of all the parameters Can plot history or real-time values Customizable history query interval Subscribes to those particular series and updates the plots in real time Storage client (aka Repository) Subscribes to a set of parameters and stores them in database structures suitable for long-term archival Is usually complemented by a web interface presenting these values Can also be embedded in another controlling application WebServices & REST clients Limited functionality: they lack the subscription mechanism /02/24 - ALICE T1/T2 Torino

I NTERACTIVE CLIENT Select “alice” only /02/24 - ALICE T1/T2 Torino

I NTERACTIVE CLIENT Browsing parameters 4 hierarchical levels of parameters (Farm, Cluster, Node, Function) /02/24 - ALICE T1/T2 Torino

I NTERACTIVE CLIENT Dynamic views /02/24 - ALICE T1/T2 Torino

W EB INTERFACE Base for the servicehttp://alimonitor.cern.ch/ Single package including Headless client Apache Tomcat PostgreSQL database Web interface includes examples of dynamic views History charts Real-time bar plots Pie, spider, histograms Status tables Most views generated with one.properties file With this foundation you can build a custom repository Dynamic pages (JSP or servlets) Other plots with the included JFreeChart libraryJFreeChart Data aggregation filters Alarms, actions /02/24 - ALICE T1/T2 Torino

M ON ALISA IN ALICE O NLINE Vasco Barroso, ALICE DAQ /02/24 - ALICE T1/T2 Torino

Continuously updating web interface, real time data /02/24 - ALICE T1/T2 Torino

Using the GUI for browsing the data /02/24 - ALICE T1/T2 Torino

SE DISCOVERY MECHANISM Base parameters for the discovery algorithm Network topology traceroute / tracepath between pairs of VoBox services 1 stream available bandwidth measurements SE functional tests Performed centrally every 2h, targeting the declared redirector add/get/rm suite using the entire AliEn stack Or just get if the storage is full The dynamically discovered xrootd data servers are tested individually, with a simplified suite Monitor discrepancies between declared volume and total space currently seen by the redirector Above issues can be seen herehere Plus many other related tests, like insufficiently large TCP buffer sizes /02/24 - ALICE T1/T2 Torino

TCP BUFFER SIZE EFFECT ON WAN TRANSFERS Online version of this plot: Recommended Recommended TCP buffer sizes: at least 8 if not 16MB /02/24 - ALICE T1/T2 Torino ICMP throttling (?) Discreet effect of the congestion control algorithm on congested links (x 8.39Mbps) 4MB buffers 8MB16MB 1000 km2000 km4000 km

AS- LEVEL NETWORK TOPOLOGY VIEW traceroute / tracepath aggregated at AS level /02/24 - ALICE T1/T2 Torino

R EPLICA DISCOVERY MECHANISM Base logic of SE selection Closest working replicas are used for both reading and writing Sorting the SEs by the network distance to the client making the request Combining network topology data with the geographical location Leaving as last resort the SEs that fail the respective functional test Weighted with their free space and recent reliability Writing is slightly randomized for more ‘democratic’ data distribution /02/24 - ALICE T1/T2 Torino

D ISTANCE METRIC FUNCTION Distance between any to IP addresses distance(IP, IP) Same C-class network Common domain name Same AS Same country (+ f (RTT between the respective AS-es if known) ) If distance between the AS-es is known, use it Same continent Far, far away distance(IP, Set ): Client's public IP to all known IPs for the storage (storage nodes, redirectors, VoBoxes near it…) /02/24 - ALICE T1/T2 Torino 0 1

W EIGHT FACTORS Distance modifiers Free space modifies the distance with f (ln(free space / 50TB)) Storage-reported space usage has priority over the catalogue view on the space Recent history of add, resp. get contribute with 75% * last day success ratio + 25% * last week success ratio To all these a per-SE knob allows tuning to particular situations Isolated SEs that need to attract more data Avoiding SEs to be upgraded / decommissioned /02/24 - ALICE T1/T2 Torino

D ATA MANAGEMENT Other data-related operations Data transfers Still relying on Andreas’ xrd3cp Falling back to the plain xrdcp in/out Data deletion AliEn should queue all physical deletes In practice “dark” data creeps in `xrd ls` is veeery slow, resync with catalogue in O(months) Removed files (63.05 TB), kept files (509.3 TB), directories from ALICE::LBL::SE, took 89d 15:13 … ALICE::CERN::EOS, took 38d 16:08 Still cannot `ls` dCache SEs (tokens are not passed by the `xrd` 3.x cmd) SE incidents Full or partial decommissioning New hardware Lost files Handled on a case-by-case basis /02/24 - ALICE T1/T2 Torino

U PCOMING OPERATION Authen and SE keys expiring Current central services certificate expires Apr 25 Is it enough to generate a new public key from the existing private one ? How to deploy it without affecting the running system ? In sync on all SEs ? A cron job watching an URL and acting on change ? When would be a good time to do this operation ? We have 2 more months to plan, deploy and execute it A good opportunity to also upgrade Xrootd /02/24 - ALICE T1/T2 Torino

2Q||!2Q ? Questions ? /02/24 - ALICE T1/T2 Torino