LOGO Scheduling system for distributed MPD data processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna.

Slides:



Advertisements
Similar presentations
Network II.5 simulator ..
Advertisements

MAP REDUCE PROGRAMMING Dr G Sudha Sadasivam. Map - reduce sort/merge based distributed processing Best for batch- oriented processing Sort/merge is primitive.
MapReduce Online Created by: Rajesh Gadipuuri Modified by: Ying Lu.
Distributed System Structures Network Operating Systems –provide an environment where users can access remote resources through remote login or file transfer.
Utilizing the GDB debugger to analyze programs Background and application.
1 OBJECTIVES To generate a web-based system enables to assemble model configurations. to submit these configurations on different.
P-GRADE and WS-PGRADE portals supporting desktop grids and clouds Peter Kacsuk MTA SZTAKI
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL March 25, 2003 CHEP 2003 Data Analysis Environment and Visualization.
Distributed Application Management Using PLuSH Jeannie Albrecht, Christopher Tuttle, Alex C. Snoeren, and Amin Vahdat UC San Diego CSE {jalbrecht, ctuttle,
Sun Grid Engine Grid Computing Assignment – Fall 2005 James Ruff Senior Department of Mathematics and Computer Science Western Carolina University.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
CPS216: Advanced Database Systems (Data-intensive Computing Systems) How MapReduce Works (in Hadoop) Shivnath Babu.
DIRAC API DIRAC Project. Overview  DIRAC API  Why APIs are important?  Why advanced users prefer APIs?  How it is done?  What is local mode what.
Module 15: Monitoring. Overview Formulate requirements and identify resources to monitor in a database environment Types of monitoring that can be carried.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Apache Airavata GSOC Knowledge and Expertise Computational Resources Scientific Instruments Algorithms and Models Archived Data and Metadata Advanced.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
National Alliance for Medical Image Computing Grid Computing with BatchMake Julien Jomier Kitware Inc.
Track 1: Cluster and Grid Computing NBCR Summer Institute Session 2.2: Cluster and Grid Computing: Case studies Condor introduction August 9, 2006 Nadya.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
A Distributed Computing System Based on BOINC September - CHEP 2004 Pedro Andrade António Amorim Jaime Villate.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
DISTRIBUTED COMPUTING
MapReduce: Simplified Data Processing on Large Clusters Jeffrey Dean and Sanjay Ghemawat.
OracleAS Reports Services. Problem Statement To simplify the process of managing, creating and execution of Oracle Reports.
J OINT I NSTITUTE FOR N UCLEAR R ESEARCH OFF-LINE DATA PROCESSING GRID-SYSTEM MODELLING FOR NICA 1 Nechaevskiy A. Dubna, 2012.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
1 Dryad Distributed Data-Parallel Programs from Sequential Building Blocks Michael Isard, Mihai Budiu, Yuan Yu, Andrew Birrell, Dennis Fetterly of Microsoft.
CPS216: Advanced Database Systems (Data-intensive Computing Systems) Introduction to MapReduce and Hadoop Shivnath Babu.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
The PROGRESS Grid Service Provider Maciej Bogdański Portals & Portlets 2003 Edinburgh, July 14th-17th.
A Guide to Oracle9i1 Database Instance startup and shutdown.
Resource Brokering in the PROGRESS Project Juliusz Pukacki Grid Resource Management Workshop, October 2003.
LOGO PROOF system for parallel MPD event processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna.
Grid Computing at Yahoo! Sameer Paranjpye Mahadev Konar Yahoo!
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
09/02 ID099-1 September 9, 2002Grid Technology Panel Patrick Dreher Technical Panel Discussion: Progress in Developing a Web Services Data Analysis Grid.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Development of the distributed monitoring system for the NICA cluster Ivan Slepov (LHEP, JINR) Mathematical Modeling and Computational Physics Dubna, Russia,
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
SEE-GRID-SCI The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research Infrastructures contract no.
LOGO Development of the distributed computing system for the MPD at the NICA collider, analytical estimations Mathematical Modeling and Computational Physics.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
Copyright © 2010, SAS Institute Inc. All rights reserved. SAS ® Using the SAS Grid.
Chapter 1 Database Access from Client Applications.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
PROGRESS: GEW'2003 Using Resources of Multiple Grids with the Grid Service Provider Michał Kosiedowski.
EGEE 3 rd conference - Athens – 20/04/2005 CREAM JDL vs JSDL Massimo Sgaravatto INFN - Padova.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Alien and GSI Marian Ivanov. Outlook GSI experience Alien experience Proposals for further improvement.
Large-scale accelerator simulations: Synergia on the Grid turn 1 turn 27 turn 19 turn 16 C++ Synergia Field solver (FFT, multigrid) Field solver (FFT,
Introduction Goal: connecting multiple computers to get higher performance – Multiprocessors – Scalability, availability, power efficiency Job-level (process-level)
Next Generation of Apache Hadoop MapReduce Owen
Joint Institute for Nuclear Research Synthesis of the simulation and monitoring processes for the data storage and big data processing development in physical.
Active-HDL Server Farm Course 11. All materials updated on: September 30, 2004 Outline 1.Introduction 2.Advantages 3.Requirements 4.Installation 5.Architecture.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
1 An unattended, fault-tolerant approach for the execution of distributed applications Manuel Rodríguez-Pascual, Rafael Mayo-García CIEMAT Madrid, Spain.
LOGO Mock Data Challenge for the MPD experiment on the NICA cluster Potrebenikov Yu. K., Schinov B. G., Rogachevsky O.V., Gertsenberger K. V. Laboratory.
SQL Database Management
Compute and Storage For the Farm at Jlab
Experience of PROOF cluster Installation and operation
PROOF system for parallel NICA event processing
Mock Data Challenge for the MPD experiment on the HybriLIT cluster
CHAPTER 7.
Controlling a large CPU farm using industrial tools
Haiyan Meng and Douglas Thain
Process Description and Control
MonteCarlo production for the BaBar experiment on the Italian grid
Presentation transcript:

LOGO Scheduling system for distributed MPD data processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna

NICA scheme Gertsenberger K.V.2

Multipurpose Detector (MPD) The software MPDRoot is developed for the MPD event simulation, reconstruction of experimental or simulated data and following physical analysis of heavy ion collisions registered by the MultiPurpose Detector at the NICA collider. 3Gertsenberger K.V.

Development of the NICA cluster 2 main directions of the development:  data storage development for the experiment  organization of parallel processing of the MPD events 4 development and expansion distributed cluster for the MPD experiment based on LHEP farm development and expansion distributed cluster for the MPD experiment based on LHEP farm Gertsenberger K.V.

Current NICA cluster in LHEP 5Gertsenberger K.V.

Data storage on the NICA cluster 6Gertsenberger K.V. Distributed file system GlusterFS  it aggregates existing file systems in a common distributed file system  automatic replication works as background process  background self- checking service restores corrupted files in case of hardware or software failure

Parallel MPD data processing PROOF server parallel data processing in ROOT macros on the parallel architectures concurrent data processing MPD-scheduler scheduling system for the task distribution to parallelize data processing on the cluster nodes 7Gertsenberger K.V.

MPD-scheduler  Developed on C++ language with ROOT classes’ support. SVN: mpdroot/macro/mpd_scheduler  Uses scheduling system the Sun Grid Engine system (qsub command) for execution in cluster mode.  SGE combines cluster machines at the LHEP farm (nc10, nc11 and nc13) into the pool of worker nodes with 34 logical processors.  Jobs for distributed execution on the NICA cluster are described and passed to MPD-scheduler as XML file: $ mpd-scheduler my_job.xml 8Gertsenberger K.V.

9 The description starts and ends with tag. Tag sets information about macro being executed by MPDRoot:  name – file path of a ROOT macro to execute, necessary parameter  start_event – number of the first event to process for all input files, optional  count_event – count of the events to process for all input files, optional  add_args – additional arguments of the ROOT macro, if required Job description. Tag. Gertsenberger K.V.

10 Tag defines files to process by macro above:  input – input file path  output – result file path  start_event – number of the first event in the input file, optional  count_event – count of the events to process in the input file, optional  paralell_mode – processor count to parallel event processing of input file, optional  merge – whether merge result part files in parallel_mode, default: “true” Gertsenberger K.V. Job description. Tag.

11 … … db_input – string for defining a list of files from MPD simulation database mpd.jinr.ru – net address of the server with simulation database and some selection parameters: range of the collision energy, type of the particle generator, particles of the collision, description and other. The list of special variables of argument “output”: ${counter} = file counter with start value and step being equal 1 ${input} = input file path ${file_name} = name of the input file without extension ${file_name_with_ext} = name of the input file with extension Gertsenberger K.V. Processing event files from MPD simulation database.

12 Tag describes run parameters and the allocated resources for the job:  mode – execution mode: ‘global’ – distributed processing on the NICA cluster, ‘local’ – multithreaded execution on a multicore computer  count – maximum count of the processors allocated for this job  config – path of a bash file with environment variables (including ROOT environment variables) being executed before macro  logs – log file path for multithreaded mode Gertsenberger K.V. Job description. Tag.

13 Tag with argument line is used to run a non-ROOT command. Running non-ROOT command on the NICA cluster Gertsenberger K.V. Job description. Non-ROOT command.

Local use MPD-scheduler can be used to parallel event processing on user multicore machine in local mode 14Gertsenberger K.V. <file input=“~/mpdroot/macro/mpd/evetest1.root" output="~/mpdroot/macro/mpd/mpddst1.root“ start_event=”0” count_event=”0”/> <file input="~/mpdroot/macro/mpd/evetest2.root" output="~/mpdroot/macro/mpd/mpddst2.root“ start_event=”0” count_event=”1000” parallel_mode=“5” merge=“true”/>

MPD-scheduler on the NICA cluster 15Gertsenberger K.V.15Gertsenberger K.V. SGE SGE = Sun Grid Engine server SGE = Sun Grid Engine worker *.root GlusterFS SGE batch system (10) (14) qsub evetest1.root SGE MPD-scheduler evetest2.root evetest3.root free free mpddst2.root job_reco.xml job_command.xml mpddst1.root mpddst3.root job_command.xml

The speedup of the one reconstruction on the NICA cluster 16Gertsenberger K.V.

The description of the scheduling system on mpd.jinr.ru 17Gertsenberger K.V.

Conclusions  The distributed NICA cluster was deployed based on LHEP farm for the NICA/MPD experiment (Fairsoft, ROOT/PROOF, MPDRoot, Gluster, Sun Grid Engine). 128 cores  The data storage was organized with the GlusterFS distributed file system: /nica/mpd[1-8]. 10 TB  The system for the distributed job execution – MPD-scheduler was developed to run MPDRoot macros concurrently on the cluster. It’s based on the Sun Grid Engine scheduling system.  The web site mpd.jinr.ru in section Computing – NICA cluster – Batch processing presents the manual for the developed MPD scheduling system. 18Gertsenberger K.V.