LOGO PROOF system for parallel MPD event processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna.

Slides:



Advertisements
Similar presentations
1 ALICE Grid Status David Evans The University of Birmingham GridPP 14 th Collaboration Meeting Birmingham 6-7 Sept 2005.
Advertisements

Objektorienteret Middleware Presentation 2: Distributed Systems – A brush up, and relations to Middleware, Heterogeneity & Transparency.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Statistics of CAF usage, Interaction with the GRID Marco MEONI CERN - Offline Week –
1 Status of the ALICE CERN Analysis Facility Marco MEONI – CERN/ALICE Jan Fiete GROSSE-OETRINGHAUS - CERN /ALICE CHEP Prague.
31 January 2007Craig E. Ward1 Large-Scale Simulation Experimentation and Analysis Database Programming Using Java.
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
Computer System Architectures Computer System Software
The ALICE Analysis Framework A.Gheata for ALICE Offline Collaboration 11/3/2008 ACAT'081A.Gheata – ALICE Analysis Framework.
LOGO Scheduling system for distributed MPD data processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna.
J OINT I NSTITUTE FOR N UCLEAR R ESEARCH OFF-LINE DATA PROCESSING GRID-SYSTEM MODELLING FOR NICA 1 Nechaevskiy A. Dubna, 2012.
Hadoop/MapReduce Computing Paradigm 1 Shirish Agale.
Transparent Grid Enablement Using Transparent Shaping and GRID superscalar I. Description and Motivation II. Background Information: Transparent Shaping.
1 Dryad Distributed Data-Parallel Programs from Sequential Building Blocks Michael Isard, Mihai Budiu, Yuan Yu, Andrew Birrell, Dennis Fetterly of Microsoft.
Parallel Processing - introduction  Traditionally, the computer has been viewed as a sequential machine. This view of the computer has never been entirely.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
BESIII Production with Distributed Computing Xiaomei Zhang, Tian Yan, Xianghu Zhao Institute of High Energy Physics, Chinese Academy of Sciences, Beijing.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
By Garrett Kelly. 3 types or reasons for distributed applications Data Data used by the application is distributed Computation Computation is distributed.
Grid Computing at Yahoo! Sameer Paranjpye Mahadev Konar Yahoo!
Development of the distributed monitoring system for the NICA cluster Ivan Slepov (LHEP, JINR) Mathematical Modeling and Computational Physics Dubna, Russia,
LOGO Development of the distributed computing system for the MPD at the NICA collider, analytical estimations Mathematical Modeling and Computational Physics.
Topic Distributed DBMS Database Management Systems Fall 2012 Presented by: Osama Ben Omran.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
Copyright © 2012, SAS Institute Inc. All rights reserved. SAS ® GRID AT PHAC SAS OTTAWA PLATFORM USERS SOCIETY, NOVEMBER 2012.
Parallelization Geant4 simulation is an embarrassingly parallel computational problem – each event can possibly be treated independently 1.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Background Computer System Architectures Computer System Software.
Alien and GSI Marian Ivanov. Outlook GSI experience Alien experience Proposals for further improvement.
Silberschatz, Galvin and Gagne ©2009Operating System Concepts – 8 th Edition Chapter 4: Threads.
Introduction Goal: connecting multiple computers to get higher performance – Multiprocessors – Scalability, availability, power efficiency Job-level (process-level)
Distributed File System. Outline Basic Concepts Current project Hadoop Distributed File System Future work Reference.
Page 1 2P13 Week 1. Page 2 Page 3 Page 4 Page 5.
Next Generation of Apache Hadoop MapReduce Owen
Joint Institute for Nuclear Research Synthesis of the simulation and monitoring processes for the data storage and big data processing development in physical.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
LOGO The unified database for the fixed target experiment Gertsenberger K. V. Laboratory of High Energy Physics, JINR for collaboration 30 September.
Cofax Scalability Document Version Scaling Cofax in General The scalability of Cofax is directly related to the system software, hardware and network.
Lecture 13 Parallel Processing. 2 What is Parallel Computing? Traditionally software has been written for serial computation. Parallel computing is the.
Grid technologies for large-scale projects N. S. Astakhov, A. S. Baginyan, S. D. Belov, A. G. Dolbilov, A. O. Golunov, I. N. Gorbunov, N. I. Gromova, I.
Threads by Dr. Amin Danial Asham. References Operating System Concepts ABRAHAM SILBERSCHATZ, PETER BAER GALVIN, and GREG GAGNE.
LIT participation LIT participation Ivanov V.V. Laboratory of Information Technologies Meeting on proposal of the setup preparation for external beams.
INTRODUCTION TO HIGH PERFORMANCE COMPUTING AND TERMINOLOGY.
LOGO Mock Data Challenge for the MPD experiment on the NICA cluster Potrebenikov Yu. K., Schinov B. G., Rogachevsky O.V., Gertsenberger K. V. Laboratory.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
MPD Data Acquisition System: Architecture and Solutions
Experience of PROOF cluster Installation and operation
PROOF system for parallel NICA event processing
Introduction to Distributed Platforms
Mock Data Challenge for the MPD experiment on the HybriLIT cluster
evoluzione modello per Run3 LHC
Definition of Distributed System
Parallel Processing - introduction
Bernd Panzer-Steindel, CERN/IT
15-826: Multimedia Databases and Data Mining
Distributed System 電機四 陳伯翰 b
Operating System Concepts
Distributed System Structures 16: Distributed Structures
Nuclear Physics Data Management Needs Bruce G. Gibbard
Support for ”interactive batch”
Laura Bright David Maier Portland State University
MonteCarlo production for the BaBar experiment on the Italian grid
Introduction To Distributed Systems
Containers on Azure Peter Lasne Sr. Software Development Engineer
Presentation transcript:

LOGO PROOF system for parallel MPD event processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna

NICA scheme Gertsenberger K.V.2

Multipurpose Detector (MPD) The software MPDRoot is developed for the MPD event simulation, reconstruction of experimental or simulated data and following physical analysis of heavy ion collisions registered by the MultiPurpose Detector at the NICA collider. 3Gertsenberger K.V.

 high interaction rate (up to 6 KHz)  high particle multiplicity, about 1000 charged particles for the central collision at the NICA energy  one event reconstruction takes tens of seconds in MPDRoot now, 1M events – months  large data stream from the MPD: is estimated at 5 to 10 PB of raw data per year 1m simulated events ~ 50 TB  MPD event data can be processed concurrently  the ability to use multicore / multiprocessor machines, computing clusters and, subsequently, GRID system 4Gertsenberger K.V. Prerequisites of the parallel processing

Current NICA cluster in LHEP 5Gertsenberger K.V.

Data storage on the NICA cluster 6Gertsenberger K.V. Distributed file system GlusterFS  it aggregates existing file systems in a common distributed file system  automatic replication works as background process  background self- checking service restores corrupted files in case of hardware or software failure

Parallel MPD event processing PROOF server parallel data processing in ROOT macros on the parallel architectures concurrent event processing MPD-scheduler scheduling system for the task distribution to parallelize data processing on the cluster nodes 7Gertsenberger K.V.

Parallel data processing with PROOF  PROOF (Parallel ROOT Facility) is a part of the ROOT software, no additional installations  PROOF uses data independent parallelism based on the lack of correlation for MPD events  good scalability  Parallelization for three parallel architectures: 1.PROOF-Lite parallelizes the data processing on one multiprocessor/multicores machine 2.PROOF parallelizes processing on heterogeneous computing cluster 3.Parallel data processing in GRID system  Transparency: the same program code can execute both sequentially and concurrently 8Gertsenberger K.V.

Using PROOF in MPDRoot  The last parameter of the reconstruction: run_type (default, “local”). Speedup on the user multicore machine: $ root reco.C(“evetest.root”, “mpddst.root”, 0, 1000, “proof”) parallel processing of 1000 events with thread count being equal logical processor count $ root reco.C(“evetest.root”, “mpddst.root”, 0, 500, “proof:workers=3”) parallel processing of 500 events with three concurrent threads Speedup on the NICA cluster: $ root reco.C(“evetest.root”, “mpddst.root”, 0, 1000, parallel processing of 1000 events on all cluster’s cores of the PoD farm $ root reco.C(“evetest.root”, …, 0, 500, parallel processing of 500 events on the PoD cluster with 15 workers  XRootD files support 9Gertsenberger K.V.

The speedup of the reconstruction on 4-cores machine 10Gertsenberger K.V.

PROOF on the NICA cluster 11Gertsenberger K.V. proof proof = master server proof = slave node *.root GlusterFS Proof On Demand Cluster (10) (14) $ root reco.C(“evetest.root”,”mpddst.root”, 0, 3, event count evetest.root event №1 event №2 mpddst.root event №0

The speedup of the reconstruction on the NICA cluster 12Gertsenberger K.V.

The description of PROOF system on mpd.jinr.ru 13Gertsenberger K.V.

Conclusions  The distributed NICA cluster was deployed on LHEP farm for the NICA/MPD experiment (Fairsoft, ROOT/PROOF, MPDRoot, Gluster). 128 cores  The data storage was organized with the GlusterFS distributed file system: /nica/mpd[1-8]. 10 TB  PROOF On Demand cluster containing nc10 (with POD server), nc11 and nc13 machines with 34 processor cores was implemented to parallelize event data processing for the MPD experiment. PROOF support was added to the reconstruction macro.  The web site mpd.jinr.ru in section Computing – NICA cluster – PROOF parallelize presents the manual for the PROOF system. 14Gertsenberger K.V.