Barthélémy von Haller CERN PH/AID For the ALICE Collaboration The ALICE data quality monitoring system.

Slides:



Advertisements
Similar presentations
Data Quality Monitoring workshop. What this presentation is not What it is and how it is organized  Definition of DQM  Overview of systems and.
Advertisements

Kondo GNANVO Florida Institute of Technology, Melbourne FL.
March 24-28, 2003Computing for High-Energy Physics Configuration Database for BaBar On-line Rainer Bartoldus, Gregory Dubois-Felsmann, Yury Kolomensky,
CHEP 2012 – New York City 1.  LHC Delivers bunch crossing at 40MHz  LHCb reduces the rate with a two level trigger system: ◦ First Level (L0) – Hardware.
1 Databases in ALICE L.Betev LCG Database Deployment and Persistency Workshop Geneva, October 17, 2005.
DQM news Technical side. Web tools  Advantages  Can be ran out of P2 with access rights  Centrally maintained -> can’t be altered on machines  Modern.
CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/27 A Control Software for the ALICE High Level Trigger Timm.
A Database Visualization Tool for ATLAS Monitoring Objects A Database Visualization Tool for ATLAS Monitoring Objects J. Batista, A. Amorim, M. Brandão,
1 The ATLAS Online High Level Trigger Framework: Experience reusing Offline Software Components in the ATLAS Trigger Werner Wiedenmann University of Wisconsin,
First year experience with the ATLAS online monitoring framework Alina Corso-Radu University of California Irvine on behalf of ATLAS TDAQ Collaboration.
New Features of APV-SRS-LabVIEW Data Acquisition Program Eraldo Oliveri on behalf of Riccardo de Asmundis INFN Napoli [Certified LabVIEW Developer] NYC,
Test results Test definition (1) Istituto Nazionale di Fisica Nucleare, Sezione di Roma; (2) Istituto Nazionale di Fisica Nucleare, Sezione di Bologna.
New CERN CAF facility: parameters, usage statistics, user support Marco MEONI Jan Fiete GROSSE-OETRINGHAUS CERN - Offline Week –
Applying Distributed Systems concepts to SCADA By Padmanabha Kamath.
Quality Control B. von Haller 8th June 2015 CERN.
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
Architecture Of ASP.NET. What is ASP?  Server-side scripting technology.  Files containing HTML and scripting code.  Access via HTTP requests.  Scripting.
Central Online Grading System COGS Dec15-21 dec1521.sd.ece.iastate.edu.
Tools and software process for the FLP prototype B. von Haller 9. June 2015 CERN.
Framework for Automated Builds Natalia Ratnikova CHEP’03.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
1 Alice DAQ Configuration DB
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
Experiences, limitations and suggested improvements The ALICE DQM Software and ROOT ROOT Users Workshop Barthelemy von Haller & Adriana Telesca for the.
Introduction and Features of Java. What is java? Developed by Sun Microsystems (James Gosling) A general-purpose object-oriented language Based on C/C++
Requirements for a Next Generation Framework: ATLAS Experience S. Kama, J. Baines, T. Bold, P. Calafiura, W. Lampl, C. Leggett, D. Malon, G. Stewart, B.
AMORE Quality Assurance Integration Filimon Roukoutakis CERN Physics Department ALICE DAQ ALICE Offline Week CERN, 11/10/2007.
David N. Brown Lawrence Berkeley National Lab Representing the BaBar Collaboration The BaBar Mini  BaBar  BaBar’s Data Formats  Design of the Mini 
Online Monitoring and Analysis for Muon Tomography Readout System M. Phipps, M. Staib, C. Zelenka, M. Hohlmann Florida Institute of Technology Department.
1 Performance Optimization In QTP Execution Over Video Automation Testing Speaker : Krishnesh Sasiyuthaman Nair Date : 10/05/2012.
Roberto Divià, CERN/ALICE 1 CHEP 2009, Prague, March 2009 The ALICE Online Data Storage System Roberto Divià (CERN), Ulrich Fuchs (CERN), Irina Makhlyueva.
Gnam Monitoring Overview M. Della Pietra, D. della Volpe (Napoli), A. Di Girolamo (Roma1), R. Ferrari, G. Gaudio, W. Vandelli (Pavia) D. Salvatore, P.
What Is Java? According to Sun in a white paper: Java: A simple, object-oriented, network-savvy, interpreted, robust, secure, architecture-neutral, portable,
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
CMS pixel data quality monitoring Petra Merkel, Purdue University For the CMS Pixel DQM Group Vertex 2008, Sweden.
Overview of DAQ at CERN experiments E.Radicioni, INFN MICE Daq and Controls Workshop.
4 th Workshop on ALICE Installation and Commissioning January 16 th & 17 th, CERN Muon Tracking (MUON_TRK, MCH, MTRK) Conclusion of the first ALICE COSMIC.
News on GEM Readout with the SRS, DATE & AMORE
AliRoot survey P.Hristov 11/06/2013. Offline framework  AliRoot in development since 1998  Directly based on ROOT  Used since the detector TDR’s for.
1 ITS Quality Assurance (& DQM) P. Cerello, P. Christakoglou, W. Ferrarese, M. Nicassio, M. Siciliano ALICE OFFLINE WEEK – April 2008.
September 2007CHEP 07 Conference 1 A software framework for Data Quality Monitoring in ATLAS S.Kolos, A.Corso-Radu University of California, Irvine, M.Hauschild.
Physics & Data Quality Monitoring at CMS Emilio Meschi (original design, run-control, mentoring) CL (core functionality, rules & alarms library, tech support)
Jean-Roch Vlimant, CERN Physics Performance and Dataset Project Physics Data & MC Validation Group McM : The Evolution of PREP. The CMS tool for Monte-Carlo.
1 Checks on SDD Data Piergiorgio Cerello, Francesco Prino, Melinda Siciliano.
MICE CM28 Oct 2010Jean-Sebastien GraulichSlide 1 Detector DAQ o Achievements Since CM27 o DAQ Upgrade o CAM/DAQ integration o Online Software o Trigger.
Online Monitoring System at KLOE Alessandra Doria INFN - Napoli for the KLOE collaboration CHEP 2000 Padova, 7-11 February 2000 NAPOLI.
Quality assurance for TPC. Quality assurance ● Process: ● Detect the problems ● Define, what is the problem ● What do we expect? ● Defined in the TDR.
CWG9 Data Quality Monitoring, Quality Assurance and Visualization B. von Haller CERN.
Maria del Carmen Barandela Pazos CERN CHEP 2-7 Sep 2007 Victoria LHCb Online Interface to the Conditions Database.
The ALICE data quality monitoring Barthélémy von Haller CERN PH/AID For the ALICE Collaboration.
Why do F77 always spoil the landscape ?. Several ideas … slide 2.
P. Vande Vyvre – CERN/PH for the ALICE collaboration CHEP – October 2010.
October 19, 2010 David Lawrence JLab Oct. 19, 20101RootSpy -- CHEP10, Taipei -- David Lawrence, JLab Parallel Session 18: Software Engineering, Data Stores,
CERN IT Department CH-1211 Genève 23 Switzerland t Load testing & benchmarks on Oracle RAC Romain Basset – IT PSS DP.
DAQ thoughts about upgrade 11/07/2012
Monthly video-conference, 18/12/2003 P.Hristov1 Preparation for physics data challenge'04 P.Hristov Alice monthly off-line video-conference December 18,
Slow Control and Run Initialization Byte-wise Environment
Slow Control and Run Initialization Byte-wise Environment
F. Bellini for the DQM core DQM meeting, 04th October 2012
SuperB and its computing requirements
LHC experiments Requirements and Concepts ALICE
Controlling a large CPU farm using industrial tools
Commissioning of the ALICE HLT, TPC and PHOS systems
Jacek Otwinowski (for the DPG QA tools and WP7 groups)
Experience between AMORE/Offline and sub-systems
QA tools – introduction and summary of activities
Support for ”interactive batch”
Module 01 ETICS Overview ETICS Online Tutorials
Offline framework for conditions data
Presentation transcript:

Barthélémy von Haller CERN PH/AID For the ALICE Collaboration The ALICE data quality monitoring system

Data Quality Monitoring 1/17 Online feedback on the quality of data Make sure to take and record high-quality data Identify and solve problem(s) early Data Quality Monitoring (DQM) involves -Online gathering of data -Analysis by user-defined algorithm -Storage of monitoring output -Visualization Oct. 19, 2010 – CHEP 2010Barthélémy von Haller - CERN PH/AID

Data-Acquisition architecture 2/17 Sub-event DA DQM

The AMORE framework AMORE : Automatic MOnitoRing Environment A DQM framework for the ALICE experiment 3/17Oct. 19, 2010 – CHEP 2010Barthélémy von Haller - CERN PH/AID

Publisher – Subscriber paradigm Notification with DIM (Distributed Information Management System) Published objects are encapsulated into « MonitorObject » structure Plugin architecture using ROOT reflection – Modules are dynamic libraries loaded at runtime 4/17 Design & Architecture Oct. 19, 2010 – CHEP 2010Barthélémy von Haller - CERN PH/AID

The Pool 5/17 Current implementation based on a database MySQL : – Light-weight – Freely distributable – Reliable – Performant Oct. 19, 2010 – CHEP 2010Barthélémy von Haller - CERN PH/AID

Subscriber & User Interface Generic GUI – Display any object of any running agent – Possibility of handling automatically the layout – Layout can be saved for future reuse – Fit the basic needs of the users to check what is published by the agents For more complex needs, users can develop their own GUI 6/17Oct. 19, 2010 – CHEP 2010Barthélémy von Haller - CERN PH/AID

7 Agents Monitor Object Sub-directories

8 MCH example : (Muon chambers) Too high occupancy -> electronic problem

Packaging and validation Subversion repositories Software distributed as RPMs Strict release procedure – Build and validate the modules on a test machine in a clean and controlled environment Nightly build – Identify broken code (wrong results, unable to compile) Coverity static build analysis to ensure code quality 9/17Oct. 19, 2010 – CHEP 2010Barthélémy von Haller - CERN PH/AID

First year experience November 2009 : LHC restarts AMORE intensively used in a real world and production environment Up to – 35 agents running – 3400 objects published per second in average – 115 MB published per second in average 10/17Oct. 19, 2010 – CHEP 2010Barthélémy von Haller - CERN PH/AID

First year experience 11

First year experience 12/17 TPC : event display in the detector Oct. 19, 2010 – CHEP 2010Barthélémy von Haller - CERN PH/AID

Performances Improvement implemented – Execution of the agents Software profiling Hardware (blades 16 cores, E5530, 2.4GHz) Parallelization Architecture (32 vs 64 bits), compiler (gcc vs icc) – Database access Server tuning Query optimisation and grouping 13/17Oct. 19, 2010 – CHEP 2010Barthélémy von Haller - CERN PH/AID

Performances: Multithreading 14/17 2 threads:1 thread for the analysis (monitor as many events as possible) 1 thread for the image production (heavy process) Oct. 19, 2010 – CHEP 2010Barthélémy von Haller - CERN PH/AID

Performances: 32 -> 64 bits 15/17 A performance gain is not always the result of a long and detailed profiling. Here, a change of architecture is enough to run 20-30% faster. Oct. 19, 2010 – CHEP 2010Barthélémy von Haller - CERN PH/AID -95%

Plans Fully automatize the process : comparisons to reference data, identification of problems, notification, actions taken Add features to take full advantage of multi- cores architecture – Multiple threads – Multiple processes 16/17Oct. 19, 2010 – CHEP 2010Barthélémy von Haller - CERN PH/AID

Conclusion AMORE has been successfully used since the LHC restart and proved to be very useful Wide range of usages Capable of handling very large number of agents, clients and objects  The architecture is adequate  We are ready for the Heavy Ion runs 17/17Oct. 19, 2010 – CHEP 2010Barthélémy von Haller - CERN PH/AID

Questions 18/17Oct. 19, 2010 – CHEP 2010Barthélémy von Haller - CERN PH/AID

19 Example of a detector (PMD) monitoring the coverage. Plots must be simple to understand for a non-expert shifter. GOOD BAD Oct. 19, 2010 – CHEP 2010

The ALICE experiment CERN : European Organisation for Nuclear Physics LHC : Large Hadron Collider ALICE : A Large Ion Collider Experiment – 18 detectors – Trigger rate : 10 KHz (max) – Bandwidth to mass storage : 1.25 GB/s 20/15

Design & Architecture 21/17

Objects Producers 22/17Oct. 19, 2010 – CHEP 2010 GDC LDC File GDC LDC File Agent Data samples Data Pool Monitor Objects Clients Monitor Objects AliRoot QA AliRoot QA HistogramsEvent Detector Algorithms Monitor Objects High Level Trigger Objects HOMER Prompt Reconstruction Plots, ESD’s, Images Event

Database improvement : grouping of queries 23