PROOF system for parallel NICA event processing

Slides:



Advertisements
Similar presentations
1 ALICE Grid Status David Evans The University of Birmingham GridPP 14 th Collaboration Meeting Birmingham 6-7 Sept 2005.
Advertisements

CMU SCS : Multimedia Databases and Data Mining Extra: intro to hadoop C. Faloutsos.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
 Need for a new processing platform (BigData)  Origin of Hadoop  What is Hadoop & what it is not ?  Hadoop architecture  Hadoop components (Common/HDFS/MapReduce)
CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/27 A Control Software for the ALICE High Level Trigger Timm.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Homework 2 In the docs folder of your Berkeley DB, have a careful look at documentation on how to configure BDB in main memory. In the docs folder of your.
CPS216: Advanced Database Systems (Data-intensive Computing Systems) How MapReduce Works (in Hadoop) Shivnath Babu.
Statistics of CAF usage, Interaction with the GRID Marco MEONI CERN - Offline Week –
1 Status of the ALICE CERN Analysis Facility Marco MEONI – CERN/ALICE Jan Fiete GROSSE-OETRINGHAUS - CERN /ALICE CHEP Prague.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
31 January 2007Craig E. Ward1 Large-Scale Simulation Experimentation and Analysis Database Programming Using Java.
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
Computer System Architectures Computer System Software
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
Computing on the Cloud Jason Detchevery March 4 th 2009.
LOGO Scheduling system for distributed MPD data processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna.
CS525: Special Topics in DBs Large-Scale Data Management Hadoop/MapReduce Computing Paradigm Spring 2013 WPI, Mohamed Eltabakh 1.
J OINT I NSTITUTE FOR N UCLEAR R ESEARCH OFF-LINE DATA PROCESSING GRID-SYSTEM MODELLING FOR NICA 1 Nechaevskiy A. Dubna, 2012.
Hadoop/MapReduce Computing Paradigm 1 Shirish Agale.
LHC Computing Plans Scale of the challenge Computing model Resource estimates Financial implications Plans in Canada.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
BESIII Production with Distributed Computing Xiaomei Zhang, Tian Yan, Xianghu Zhao Institute of High Energy Physics, Chinese Academy of Sciences, Beijing.
4/20/02APS April Meeting1 Database Replication at Remote sites in PHENIX Indrani D. Ojha Vanderbilt University (for PHENIX Collaboration)
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
LOGO PROOF system for parallel MPD event processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna.
Grid Computing at Yahoo! Sameer Paranjpye Mahadev Konar Yahoo!
ROOT for Data Analysis1 Intel discussion meeting CERN 5 Oct 2003 Ren é Brun CERN Distributed Data Analysis.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Development of the distributed monitoring system for the NICA cluster Ivan Slepov (LHEP, JINR) Mathematical Modeling and Computational Physics Dubna, Russia,
DISTRIBUTED COMPUTING. Computing? Computing is usually defined as the activity of using and improving computer technology, computer hardware and software.
Kjell Orsborn UU - DIS - UDBL DATABASE SYSTEMS - 10p Course No. 2AD235 Spring 2002 A second course on development of database systems Kjell.
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015.
LOGO Development of the distributed computing system for the MPD at the NICA collider, analytical estimations Mathematical Modeling and Computational Physics.
CS525: Big Data Analytics MapReduce Computing Paradigm & Apache Hadoop Open Source Fall 2013 Elke A. Rundensteiner 1.
Topic Distributed DBMS Database Management Systems Fall 2012 Presented by: Osama Ben Omran.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Hadoop/MapReduce Computing Paradigm 1 CS525: Special Topics in DBs Large-Scale Data Management Presented By Kelly Technologies
Parallelization Geant4 simulation is an embarrassingly parallel computational problem – each event can possibly be treated independently 1.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Background Computer System Architectures Computer System Software.
Alien and GSI Marian Ivanov. Outlook GSI experience Alien experience Proposals for further improvement.
Introduction Goal: connecting multiple computers to get higher performance – Multiprocessors – Scalability, availability, power efficiency Job-level (process-level)
Next Generation of Apache Hadoop MapReduce Owen
Joint Institute for Nuclear Research Synthesis of the simulation and monitoring processes for the data storage and big data processing development in physical.
Active-HDL Server Farm Course 11. All materials updated on: September 30, 2004 Outline 1.Introduction 2.Advantages 3.Requirements 4.Installation 5.Architecture.
ATLAS TIER3 in Valencia Santiago González de la Hoz IFIC – Instituto de Física Corpuscular (Valencia)
LOGO The unified database for the fixed target experiment Gertsenberger K. V. Laboratory of High Energy Physics, JINR for collaboration 30 September.
Lecture 13 Parallel Processing. 2 What is Parallel Computing? Traditionally software has been written for serial computation. Parallel computing is the.
Grid technologies for large-scale projects N. S. Astakhov, A. S. Baginyan, S. D. Belov, A. G. Dolbilov, A. O. Golunov, I. N. Gorbunov, N. I. Gromova, I.
INTRODUCTION TO HIGH PERFORMANCE COMPUTING AND TERMINOLOGY.
LOGO Mock Data Challenge for the MPD experiment on the NICA cluster Potrebenikov Yu. K., Schinov B. G., Rogachevsky O.V., Gertsenberger K. V. Laboratory.
Emanuele Leonardi PADME General Meeting - LNF January 2017
MPD Data Acquisition System: Architecture and Solutions
Experience of PROOF cluster Installation and operation
WP18, High-speed data recording Krzysztof Wrona, European XFEL
Introduction to Distributed Platforms
Overview of the Belle II computing
Mock Data Challenge for the MPD experiment on the HybriLIT cluster
evoluzione modello per Run3 LHC
15-826: Multimedia Databases and Data Mining
Distributed System 電機四 陳伯翰 b
US CMS Testbed.
Grid Canada Testbed using HEP applications
Ch 4. The Evolution of Analytic Scalability
Nuclear Physics Data Management Needs Bruce G. Gibbard
Support for ”interactive batch”
Presentation transcript:

PROOF system for parallel NICA event processing MPD/BM@N Collaboration Joint Institute for Nuclear Research, Dubna

NICA accelerator complex Gertsenberger K.V.

MPD and MpdRoot software The software MpdRoot is developed for the MPD event simulation, reconstruction of experimental or simulated data and following physical analysis of heavy ion collisions registered by the MultiPurpose Detector at the NICA collider. (based on ROOT and FairRoot) The MpdRoot software is available in the GitLab https://git.jinr.ru/nica/mpdroot Gertsenberger K.V.

BM@N and BmnRoot software The software BmnRoot is developed for the BM@N event simulation, reconstruction of experimental or simulated data and following physical analysis of collisions of elementary particles and ions with a fixed target at the NICA collider. (based on ROOT and FairRoot) The BmnRoot software is available in the GitLab https://git.jinr.ru/nica/bmnroot Gertsenberger K.V.

Prerequisites of the parallel processing/storing high interaction rate (up to 7 KHz) high particle multiplicity, up to 1000 charged particles for the central collision at the NICA energy one event reconstruction can take a lot of time in MpdRoot now large data stream from the MPD: is estimated at 5-10 PB of raw data per year 100m simulated events ~ 5 PB MPD event data can be processed concurrently! Gertsenberger K.V.

Current NICA cluster (prototype) Gertsenberger K.V.

Current data storage on the NICA cluster GlusterFS distributed file system free and open source aggregates existing file systems in a common distributed file system has no metadata server automatic replication works as background process background self-checking service restores corrupted files in case of hardware or software failure Gertsenberger K.V.

Parallel event processing NICA cluster concurrent data processing on cluster nodes PROOF server parallel event data processing in ROOT macros on the parallel architectures MPD-Scheduler scheduling system for task distribution to parallelize MPD data processing on the cluster nodes Gertsenberger K.V.

Parallel data processing with PROOF PROOF (Parallel ROOT Facility) is a part of the ROOT software, no additional installations PROOF uses data independent parallelism based on the lack of correlation for MPD events  good scalability Parallelization for three parallel architectures: PROOF-Lite parallelizes the data processing on one multiprocessor/multicores machine PROOF parallelizes processing on heterogeneous computing cluster Parallel data processing in GRID system Gertsenberger K.V.

Using PROOF in reconstruction The last parameter of the reconstruction macro: run_type (default, “local”). Speedup on the user multicore machine, e.g. for MpdRoot: $ root reco.C(“evetest.root”, “mpddst.root”, 0, 1000, “proof”) parallel processing of 1000 events with thread count being equal logical processor count $ root reco.C(“evetest.root”, “mpddst.root”, 0, 500, “proof:workers=3”) parallel processing of 500 events with three concurrent threads Speedup on the NICA cluster: $ root reco.C(“evetest.root”, “mpddst.root”, 0, 1000, “proof:mpd@nc10.jinr.ru:21001”) parallel processing of 1000 events on all cluster’s cores of the PoD farm $ root reco.C(“evetest.root”, …, 0, 500, “proof:mpd@nc10.jinr.ru:21001:workers=15”) parallel processing of 500 events on the PoD cluster with 15 workers Gertsenberger K.V.

The speedup of the reconstruction on 4-cores Gertsenberger K.V.

PROOF on the NICA cluster (file splitting mode) event count $ root reco.C(“evetest.root”,”mpddst.root”, 0, 3, “proof:mpd@nc10.jinr.ru:21001”) GlusterFS mpddst.root *.root evetest.root event №0 event №1 event №2 proof proof proof proof proof proof proof = master server proof = slave node Proof On Demand cluster (64 cores) Gertsenberger K.V.

The speedup on the NICA cluster Gertsenberger K.V.

«PROOF Parallelization» section on mpd.jinr.ru Gertsenberger K.V.

Conclusions The distributed NICA cluster contains 364 processor cores for the NICA experiments (Fairsoft, ROOT/PROOF, BmnRoot, MpdRoot, GlusterFS, Sun Grid Engine). The data storage based on GlusterFS distributed file system provides 83+4 TB shared space (mirrored data). PROOF On Demand cluster with 64 processor cores was implemented to parallelize event data processing for the NICA experiments. PROOF support was added to the reconstruction macro. Parallel reconstruction with PROOF can be used on user multicore machine or on the NICA cluster (just one parameter has to be changed). The web site mpd.jinr.ru in section Computing – NICA cluster – PROOF parallelization presents the manual for the PROOF system. Gertsenberger K.V.