General Grid Monitoring Infrastructure (GGMI) Peter kacsuk and Norbert Podhorszki MTA SZTAKI.

Slides:



Advertisements
Similar presentations
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
Advertisements

WP2: Data Management Gavin McCance University of Glasgow November 5, 2001.
IEEE NSS 2003 Performance of the Relational Grid Monitoring Architecture (R-GMA) CMS data challenges. The nature of the problem. What is GMA ? And what.
Legacy code support for commercial production Grids G.Terstyanszky, T. Kiss, T. Delaitre, S. Winter School of Informatics, University.
Grid Application Builders Teach In31/01/02Antony Wilson Information & Monitoring Services WP3.
WP 1 Grid Workload Management Massimo Sgaravatto INFN Padova.
USING THE GLOBUS TOOLKIT This summary by: Asad Samar / CALTECH/CMS Ben Segal / CERN-IT FULL INFO AT:
Hungarian GRID Projects and Cluster Grid Initiative P. Kacsuk MTA SZTAKI
Visual Solution to High Performance Computing Computer and Automation Research Institute Laboratory of Parallel and Distributed Systems
Sergey Belov, LIT JINR 15 September, NEC’2011, Varna, Bulgaria.
CrossGrid Task 3.3 Grid Monitoring Trinity College Dublin (TCD) Brian Coghlan Paris MAR-2002.
MCTS Guide to Microsoft Windows Server 2008 Network Infrastructure Configuration Chapter 8 Introduction to Printers in a Windows Server 2008 Network.
GRID technology by SZTAKI MTA SZTAKI Laboratory of Parallel and Distributed Systems
MultiJob PanDA Pilot Oleynik Danila 28/05/2015. Overview Initial PanDA pilot concept & HPC Motivation PanDA Pilot workflow at nutshell MultiJob Pilot.
Sergey Belov, Tatiana Goloskokova, Vladimir Korenkov, Nikolay Kutovskiy, Danila Oleynik, Artem Petrosyan, Roman Semenov, Alexander Uzhinskiy LIT JINR The.
CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 CMS Applications Progress towards GridPP milestones Data management.
Automated Tracing and Visualization of Software Security Structure and Properties Symposium on Visualization for Cyber Security 2012 (VizSec’12) Seattle,
Track 1: Cluster and Grid Computing NBCR Summer Institute Session 2.2: Cluster and Grid Computing: Case studies Condor introduction August 9, 2006 Nadya.
Hungarian Supercomputing GRID
Enabling Workflow in UPnP Networks Andreas BobekUniversity of Rostock Faculty of Computer Science and Electrical Engineering Andreas Bobek, Hendrik Bohn,
Cluster Programming Technology and its Application in Meteorology Computer and Automation Research Institute Hungarian Academy of Sciences Hungarian Meteorological.
ATLAS Off-Grid sites (Tier-3) monitoring A. Petrosyan on behalf of the ATLAS collaboration GRID’2012, , JINR, Dubna.
DynamicBLAST on SURAgrid: Overview, Update, and Demo John-Paul Robinson Enis Afgan and Purushotham Bangalore University of Alabama at Birmingham SURAgrid.
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
Introduction on R-GMA Shi Jingyan Computing Center IHEP.
Computer and Automation Research Institute Hungarian Academy of Sciences Presentation and Analysis of Grid Performance Data Norbert Podhorszki and Peter.
GRM + Mercury in P-GRADE Monitoring of P-GRADE applications in the Grid using GRM and Mercury.
Support for Debugging Automatically Parallelized Programs Robert Hood Gabriele Jost CSC/MRJ Technology Solutions NASA.
Computer and Automation Research Institute Hungarian Academy of Sciences Automatic checkpoint of CONDOR-PVM applications by P-GRADE Jozsef Kovacs, Peter.
SZTAKI in DataGrid 2003 What to do this year. Topics ● Application monitoring (GRM) ● Analysis and Presentation (Pulse) ● Performance of R-GMA.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Grid Workload Management Massimo Sgaravatto INFN Padova.
INFSO-RI Enabling Grids for E-sciencE Supporting legacy code applications on EGEE VOs by GEMLCA and the P-GRADE portal P. Kacsuk*,
Stuart Wakefield Imperial College London Evolution of BOSS, a tool for job submission and tracking W. Bacchi, G. Codispoti, C. Grandi, INFN Bologna D.
OMIS Approach to Grid Application Monitoring Bartosz Baliś Marian Bubak Włodzimierz Funika Roland Wismueller.
1 Advanced features of the P-GRADE portal Peter Kacsuk, Gergely Sipos Peter Kacsuk, Gergely Sipos MTA.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Code Applications Tamas Kiss Centre for Parallel.
Grid Architecture William E. Johnston Lawrence Berkeley National Lab and NASA Ames Research Center (These slides are available at grid.lbl.gov/~wej/Grids)
WP3 Information and Monitoring Steve Fisher / RAL 23/9/2003.
Performance evaluation on grid Zsolt Németh MTA SZTAKI Computer and Automation Research Institute.
Computer and Automation Research Institute Hungarian Academy of Sciences SZTAKI’s work in DataGrid WP September Norbert Podhorszki Laboratory of.
1 P-GRADE Portal: a workflow-oriented generic application development portal Peter Kacsuk MTA SZTAKI, Hungary Univ. of Westminster, UK.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Applications.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Status report on Application porting at SZTAKI.
Website: Answering Continuous Queries Using Views Over Data Streams Alasdair J G Gray Werner.
LCG workshop on Operational Issues CERN November, EGEE CIC activities (SA1) Accounting: current status
FP6−2004−Infrastructures−6-SSA E-infrastructure shared between Europe and Latin America Alexandre Duarte CERN IT-GD-OPS UFCG LSD 1st EELA Grid School.
 CMS data challenges. The nature of the problem.  What is GMA ?  And what is R-GMA ?  Performance test description  Performance test results  Conclusions.
The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research Infrastructures contract no Workflow repository, user.
CMS Usage of the Open Science Grid and the US Tier-2 Centers Ajit Mohapatra, University of Wisconsin, Madison (On Behalf of CMS Offline and Computing Projects)
The impact of R-GMA (upon WP1 and WP4). EDG (Paris) 6 Mar James MagowanImpact of R-GMA Grid Monitoring Architecture (GMA) We use it not only for.
JSS Job Submission Service Massimo Sgaravatto INFN Padova.
Grid Workload Management (WP 1) Massimo Sgaravatto INFN Padova.
1 Porting applications to the NGS, using the P-GRADE portal and GEMLCA Peter Kacsuk MTA SZTAKI Hungarian Academy of Sciences Centre for.
INFSO-RI Enabling Grids for E-sciencE R-GMA Gergely Sipos and Péter Kacsuk MTA SZTAKI Credit to Valeria Ardizzone.
WP1 Status and plans Francesco Prelz, Massimo Sgaravatto 4 th EDG Project Conference Paris, March 6 th, 2002.
© Geodise Project, University of Southampton, Workflow Support for Advanced Grid-Enabled Computing Fenglian Xu *, M.
Grid Execution Management for Legacy Code Architecture Exposing legacy applications as Grid services: the GEMLCA approach Centre.
Origami: Scientific Distributed Workflow in McIDAS-V Maciek Smuga-Otto, Bruce Flynn (also Bob Knuteson, Ray Garcia) SSEC.
SAM Status Update Piotr Nyczyk LCG Management Board CERN, 5 June 2007.
Chapter 11 Analysis Methodology Spring Incident Response & Computer Forensics.
TIFR, Mumbai, India, Feb 13-17, GridView - A Grid Monitoring and Visualization Tool Rajesh Kalmady, Digamber Sonvane, Kislay Bhatt, Phool Chand,
InSilicoLab – Grid Environment for Supporting Numerical Experiments in Chemistry Joanna Kocot, Daniel Harężlak, Klemens Noga, Mariusz Sterzel, Tomasz Szepieniec.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Applications.
Tamas Kiss University Of Westminster
Peter Kacsuk – Sipos Gergely MTA SZTAKI
EO Applications Parallel Session
Grid Application Support Group Case study Schrodinger equations on the Grid Status report 16. January, Created by Akos Balasko
Peter Kacsuk MTA SZTAKI
MMG: from proof-of-concept to production services at scale
Presentation transcript:

General Grid Monitoring Infrastructure (GGMI) Peter kacsuk and Norbert Podhorszki MTA SZTAKI

General Grid Monitoring Infrastructure (GGMI) Grid Status Monitoring Infrastructure GSMI (R-GMA) Grid Appl. Monitoring Infrastructure GAMI PROVE GRM PULSE R-GMA Browser

Performance comparison of GSMI and GAMI ● Performance measurement – Loop, instrumented with GRM – For loop N, generates 2N+2 events (loop begin + loop end+ start + exit) – 1 machine – 3 machines ● M1: Producer and ProducerServlet ● M2: All the other servlets including ConsumerServlet ● M3: Consumer

● AppTotal = time between inserting first and last event ● Total = time between inserting first event and receiving the last event in Consumer ● On 3 machines, N=1000 R-GMA starts loosing events. Only 1848, 1780 from 2002 events received. ● For N=10K, test never finishes (one night at least) Performance of GSMI

● AppTotal = time between inserting first and last event ● Total = time between inserting first event and receiving the last event in Consumer ● No loosing events ● Linear scaling Performance of GAMI

R-GMA GAMI !!! !!! R-GMA vs GAMI on 3 machines GAMI 1 machine vs 3 machines

GAMI structure PROVE Application Process Site 1 Local Host Host 1Host 2Host 1 Local Monitor LM Site 2 Local Monitor LM Main Monitor MM Appl. Process

GAMI ● To deliver trace data from the application to the user efficiently. – Uses TCP Socket communication – Data in XDR format and could be optimised for TCP transmission – Two sw. hops between application and GRM: local and main monitors – One hw. hop: host of main monitor

Steps of application monitoring ● Step 1: user submits a job (gets GID from the broker) ● Step 2: user starts PROVE with parameter GID ● Step 3: PROVE looks for the execution site (search in R- GMA) ● Step 4: PROVE looks for the address of GAMI Main Monitor of the execution site (search in R-GMA) ● Step 5: PROVE subscribes for application trace at the GAMI Main Monitor ● Step 6: GAMI Main Monitor associates the application job id (GID) with the Unix process ids.

Problems of Application Monitoring ● Problem 1: To find the execution site of the application by PROVE – Where is it running? -> machineX.siteY ● Problem 2: To find the monitor to be connected – What is the address of GAMI Main Monitor running at siteY? ● Problem 3: To find the application by the GAMI Main Monitor – What processes (PIDs) belong to application GID? ● Solution: The info needed for solving these problems should be published in R-GMA => integration of R- GMA and GAMI needed

Problems ● Problem 1: To find the execution site of the application by PROVE – Where is it running? -> machineX.siteY – Broker  R-GMA (discussion with WP1) ● Problem 2: To find the monitor to be connected – What is the address of GAMI Main Monitor running at siteY? – GAMI Main Monitor  R-GMA

Problems ● Problem 3: To find the application by the GAMI Main Monitor – What processes (PIDs) belong to application GID? ● Problem to be solved: 5 levels of job/process ids – GID (generated by the resource broker) – Condor G – ID – GRAM ID – Local job manager ID – Process ID ● Discussion with WP1

Temporary solution for the 3 rd problem ● User defines unique id for the application ● Application process publishes this id to the GAMI Local Monitor ● PROVE will use this id for collecting trace data

General Grid Monitoring Infrastructure (GGMI) Grid Status Monitoring Infrastructure GSMI (R-GMA) Grid Appl. Monitoring Infrastructure GAMI PROVE GRM PULSE R-GMA Browser User support tools

Tools ● Pulse: – Analysis and presentation of Grid performance data ● R-GMA browser: – web-based browser for available shemas and producers within the R-GMA ● GRM: – Instrumentation library for trace collection – On-line and off-line monitoring of sequential and MPI applications ● PROVE: – On-line and off-line visualization of trace for sequential and MPI applications

Documents and reports ● User's Manual for the stand-alone GRM/PROVE ● GRM/PROVE User's Guide ● Versions of GRM ● Peformance Monitoring, Analysis and Presentation for Grid Applications – Technical report about GRM within the EU-DataGrid project ●

Publications ● From Cluster Monitoring to Grid Monitoring Based on GRM – EuroPar’2001, Manchester ● Application Monitoring in the Grid with GRM and PROVE – Proc. of the International Conference on Computational Science - ICCS 2001, San Francisco ● Presentation and Analysis of Grid Performance Data – EuroPar'2003, Klagenfurt ● Pulse: A Tool for Presentation and Analysis of Grid Performance Data – MIPRO'2003, Opatija ●

Summary ● Advantages of the concept: – Gives a full Grid monitoring infrastructure including both ● Status monitoring ● Application monitoring – Supports on-line and off-line mpi application monitoring and visualization – Increases the chance that it will be used by LCG-2 – No special or extra requirement for R-GMA – Integration will be done by SZTAKI – Gives the potential of competing the US solutions ● already two prestigious papers at EuroPar’01 and EuroPar’03 ● Further potential publication in JOGC