ATLAS ONLINE MONITORING. FINISHED! Now what? How to check quality of the data?!! DATA FLOWS!

Slides:



Advertisements
Similar presentations
RPC & LVL1 Mu Barrel Online Monitoring during LS1 M. Della Pietra.
Advertisements

Sander Klous on behalf of the ATLAS Collaboration Real-Time May /5/20101.
GNAM and OHP: Monitoring Tools for the ATLAS Experiment at LHC GNAM and OHP: Monitoring Tools for the ATLAS Experiment at LHC M. Della Pietra, P. Adragna,
June 19, 2002 A Software Skeleton for the Full Front-End Crate Test at BNL Goal: to provide a working data acquisition (DAQ) system for the coming full.
The LAr ROD Project and Online Activities Arno Straessner and Alain, Daniel, Annie, Manuel, Imma, Eric, Jean-Pierre,... Journée de réflexion du DPNC Centre.
Atlas SemiConductor Tracker Andrée Robichaud-Véronneau.
Uli Schäfer Discussions with H.B. after last meeting… All ATLAS standard racks and crates will be monitored. Helfrieds group will equip them with ELMBs.
Data Quality Monitoring for CMS RPC A. Cimmino, D. Lomidze P. Noli, M. Maggi, P. Paolucci.
March 2003 CHEP Online Monitoring Software Framework in the ATLAS Experiment Serguei Kolos CERN/PNPI On behalf of the ATLAS Trigger/DAQ Online Software.
First year experience with the ATLAS online monitoring framework Alina Corso-Radu University of California Irvine on behalf of ATLAS TDAQ Collaboration.
Large Scale and Performance Tests of the ATLAS Online Software CERN ATLAS TDAQ Online Software System D.Burckhart-Chromek, I.Alexandrov, A.Amorim, E.Badescu,
Data Quality Monitoring of the CMS Tracker
DAQ System at the 2002 ATLAS Muon Test Beam G. Avolio – Univ. della Calabria E. Pasqualucci - INFN Roma.
Designing a HEP Experiment Control System, Lessons to be Learned From 10 Years Evolution and Operation of the DELPHI Experiment. André Augustinus 8 February.
Offline Tracker DQM Shift Tutorial. 29/19/20152 Tracker Shifts Overview Online Shifts at P5 (3/day for 24 hours coverage) – One Pixel shifter and one.
JCOP Workshop September 8th 1999 H.J.Burckhart 1 ATLAS DCS Organization of Detector and Controls Architecture Connection to DAQ Front-end System Practical.
Clara Gaspar, October 2011 The LHCb Experiment Control System: On the path to full automation.
IceCube DAQ Mtg. 10,28-30 IceCube DAQ: “DOM MB to Event Builder”
Independent front-end read out subsystems for 17 detectors in three underground sites. All electronics modules of each subsystem are reside in one VME.
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
G. Maron, Agata Week, Orsay, January Agata DAQ Layout Gaetano Maron INFN – Laboratori Nazionali di Legnaro.
André Augustinus 10 September 2001 DCS Architecture Issues Food for thoughts and discussion.
Control in ATLAS TDAQ Dietrich Liko on behalf of the ATLAS TDAQ Group.
Report on the Commissioning Task Force activity Global and sub-detector views on ECS Histogram handling : – histogram handling and PVSS – the ALEPH and.
Gnam Monitoring Overview M. Della Pietra, D. della Volpe (Napoli), A. Di Girolamo (Roma1), R. Ferrari, G. Gaudio, W. Vandelli (Pavia) D. Salvatore, P.
ALICE, ATLAS, CMS & LHCb joint workshop on
OFFLINE TRIGGER MONITORING TDAQ Training 5 th November 2010 Ricardo Gonçalo On behalf of the Trigger Offline Monitoring Experts team.
Clara Gaspar, March 2005 LHCb Online & the Conditions DB.
CMS pixel data quality monitoring Petra Merkel, Purdue University For the CMS Pixel DQM Group Vertex 2008, Sweden.
Introduction CMS database workshop 23 rd to 25 th of February 2004 Frank Glege.
Overview of DAQ at CERN experiments E.Radicioni, INFN MICE Daq and Controls Workshop.
4 th Workshop on ALICE Installation and Commissioning January 16 th & 17 th, CERN Muon Tracking (MUON_TRK, MCH, MTRK) Conclusion of the first ALICE COSMIC.
TRT Offline Software DOE Visit, August 21 st 2008 Outline: oTRT Commissioning oTRT Offline Software Activities oTRT Alignment oTRT Efficiency and Noise.
U.S. ATLAS Executive Committee August 3, 2005 U.S. ATLAS TDAQ FY06 M&O Planning A.J. Lankford UC Irvine.
The Status of the ATLAS Experiment Dr Alan Watson University of Birmingham on behalf of the ATLAS Collaboration.
ALICE Pixel Operational Experience R. Santoro On behalf of the ITS collaboration in the ALICE experiment at LHC.
Part I – Shifter Duties Part II – ACR environment Part III – Run Control & DAQ Part IV – Beam Part V – DCS Part VI – Data Quality Monitoring Part VII.
September 2007CHEP 07 Conference 1 A software framework for Data Quality Monitoring in ATLAS S.Kolos, A.Corso-Radu University of California, Irvine, M.Hauschild.
CHEP March 2003 Sarah Wheeler 1 Supervision of the ATLAS High Level Triggers Sarah Wheeler on behalf of the ATLAS Trigger/DAQ High Level Trigger.
TGC Timing Adjustment Chikara Fukunaga (TMU) ATLAS Timing Workshop 5 July ‘07.
Online Reconstruction 1M.Ellis - CM th October 2008.
Pixel DQM Status R.Casagrande, P.Merkel, J.Zablocki (Purdue University) D.Duggan, D.Hidas, K.Rose (Rutgers University) L.Wehrli (ETH Zuerich) A.York (University.
DQM for the RPC subdetector M. Maggi and P. Paolucci.
Online Monitoring System at KLOE Alessandra Doria INFN - Napoli for the KLOE collaboration CHEP 2000 Padova, 7-11 February 2000 NAPOLI.
TDAQ Experience in the BNL Liquid Argon Calorimeter Test Facility Denis Oliveira Damazio (BNL), George Redlinger (BNL).
The ATLAS DAQ System Online Configurations Database Service Challenge J. Almeida, M. Dobson, A. Kazarov, G. Lehmann-Miotto, J.E. Sloper, I. Soloviev and.
The Detector Performance Study for the Barrel Section of the ATLAS Semiconductor Tracker (SCT) with Cosmic Rays Yoshikazu Nagai (Univ. of Tsukuba) For.
Software for the CMS Cosmic Challenge Giacomo BRUNO UCL, Louvain-la-Neuve, Belgium On behalf of the CMS Collaboration CHEP06, Mumbay, India February 16,
DQM for the RPC subdetector M. Maggi and P. Paolucci.
Online Consumers produce histograms (from a limited sample of events) which provide information about the status of the different sub-detectors. The DQM.
1 DAQ.IHEP Beijing, CAS.CHINA mail to: The Readout In BESIII DAQ Framework The BESIII DAQ system consists of the readout subsystem, the.
14 th IEEE-NPSS Real Time Stockholm - June 9 th 2005 P. F. Zema The GNAM monitoring system and the OHP histogram presenter for ATLAS 14 th IEEE-NPSS Real.
ID Week 13 th of October 2014 Per Johansson Sheffield University.
1 4 July 2006 Alan Barr - SCT DAQ Experience and plans from running the (SCT) DAQ at SR1 HEP Cosmics setup Running modes Problems Future.
TGC On-Line Monitoring Status for 1 beam Before LHC Startup on TomBox: A single machine (pc-tgc-mon-rod-01) receives all TGC events.
ATLAS The ConditionDB is accessed by the offline reconstruction framework (ATHENA). COOLCOnditions Objects for LHC The interface is provided by COOL (COnditions.
M. Caprini IFIN-HH Bucharest DAQ Control and Monitoring - A Software Component Model.
ANDREA NEGRI, INFN PAVIA – NUCLEAR SCIENCE SYMPOSIUM – ROME 20th October
TRTViewer: the ATLAS TRT detector monitoring and diagnostics tool 4 th Workshop on Advanced Transition Radiation Detectors for Accelerator and Space Applications.
EPS HEP 2007 Manchester -- Thilo Pauly July The ATLAS Level-1 Trigger Overview and Status Report including Cosmic-Ray Commissioning Thilo.
TGC L1 Trigger On-Line Monitoring Tel Aviv University:
Risultati del run di integrazione M4
Controlling a large CPU farm using industrial tools
Online Software Status
Data Quality Monitoring of the CMS Silicon Strip Tracker Detector
Level-1 Calo Monitoring
TDAQ commissioning and status Stephen Hillier, on behalf of TDAQ
CMS Pixel Data Quality Monitoring
DQM for the RPC subdetector
CMS Pixel Data Quality Monitoring
Presentation transcript:

ATLAS ONLINE MONITORING

FINISHED! Now what? How to check quality of the data?!! DATA FLOWS!

For slide 2: Detector fully installed with all components and all wires, You want to turn it on... How to know if your detector is doing what you want it to do. Is it registering hits, is it sending data out?

Online Monitoring: Surveillance of the data and its quality Why monitor? To be able to discover in a relatively easy way that the detector is not behaving as it should and take action Examples: detector modules that do not work, synchronization problems btw SCT and TRT, efficiencies sink....

Monitors quality of data Provide checks for the shifter so that data quality problems can be fixed in an early stage. Provide automatic alarms/checks Especially important in the beginning of an experiment for efficiently and quickly solving runtime problems For example: if a certain detector segment has a large noise occupancy or is not producing data the run control can do reconfiguration of the module PURPOSE OF ONLINE MONITORING

HISTOGRAM REPOSITORY (HISTOGRAM DISPLAY)‏ PUBLISH OHP Monitoring display SUBSCRIBE UPDATE HISTO DQM Monitoring display I see tracks in the SCT! OHP Tools for retrieving data from dataflow booking, filling and publishing histograms

For slide 6: Dataflow – Yuriy Partition – sub-detector included in TDAQ Configuration files – tool makes this produces xml file fed into TDAQ Then your detector can run As data flows from the detector specially written tools that work in parallell with the dataflow retrieve information, fill histos & sends this to a repository. Application such as OHP, DQMF subscribe to this information and histograms and display them so that shifters can examine their quality M5 experience M6 first time the Inner Detector as a composite system was included in TDAQ and monitored

Readout System ROB,ROS,SFI,SFO, … LVL2/EF Tier 0 Calibration FARM Gatherer Mon Intelligent Monitoring Display Experts Archiver Data Quelity Assessment Alarms notifications DQMF Slow Ctrl. DBS Var. Ref. DBS Monitoring DBS Data Qual. DBS Monitoring Scheme Mon LVL1 Shifter displays OHP Gatherer

Where and what is monitored Detector monitoring: DCS: detector hardware status and conditions Online Monitoring  ROD Crate, ROS: data quality and integrity  Event Builder: correlation between sub-detectors, consistency of LVL1 information Trigger Monitoring LVL1, LVL2: sample rejected events to check the trigger decision Event Filter: information attached to a sub-set of accepted and rejected events DAQ monitoring: ROS, EB: operational monitoring (buffer occupancies, throughput, s/w and h/w status, errors, etc.)‏

For slide 8: Grab data from all stages of dataflow Gather info from several detector to fill histo Send to databases Display

For slide 11: Several ways to monitor ID uses Athena Monitoring Tools Offline Athena tools are used online via Athena Processing Tools Connects the offline tools to the online detector data Then other services provide the connection between the histogram algorithm and connection to histogram repository – histogram service

Slide 9,10,11,12 Maybe just use the simpler graphical slide 11? Detector divided into sub-detectors, Inner detector Pixel, SCT, TRT Parts of the detector or the whole detector can run Data is read out in a hierarchical manner Each sub-detector has their own Read Out Drivers Read Out Servers receive output from all RODS etc. The services providing online information and histograms get input from many stages of the dataflow from early ROD level to SFI where full events are available This is a separate system parallel to the dataflow and does not interfere. Have own processors and workstations to handle online monitoring

Online Histogram Presenter (OHP)‏ Atlas Data Quality Framework (DQMF)‏ Online Histogram Display (OH)‏ Trigger Monitoring Online Monitoring applications Trigger OHPDQM OH

Monitoring Displays: Online Histogram Presenter (OHP)‏ OHP displays already existing histograms Displayed when histos are updated, i.e. when the detector is running and histos are published OHP subscribes to one or several histograms (or even all) and displays them online If the detector is not running a message will be displayed saying that no histogram is available

Histograms - online monitoring Each detector or detector-subsystem expert has written tools for retrieving data from detector and filling histograms Global Inner Detector uses 7 tools for Athena PT Responsible: Arshak Tonoyan& Heidi Sandaker monitoring of LVL1, BCIDs matching of ID segments at TRT/SCT and SCT/Pixel boundary surface ID noise occupancies monitoring of no. of hits, residuals on combined tracks matching parameters of top and bottom tracks, SCT&TRT extrapolation of SCT segments to TRT: TRT straw efficiencies, residuals, no. of TRT hits on extrapolated track TRT straw efficiencies Produces many histograms available in Online Histogram Display – only 10 allowed for each sub-detector in Online Histogram Presenter

OHP

What will a shifter do Look at histograms Check if they are as they should  Are the histograms actually filling  Do they look like the reference-histograms If not check with the relevant detector desk Inform about what you see is it ok? Maybe a module should be masked off Maybe a detector-module needs to be reconfigured Maybe the run even has to stop (seldom!)?

OHP example: SCT Hits for all tracks in the event THE NUMBER OF HITS PER ALL TRACKS SHOULD BE COMPARED WITH A REFERENCE HISTO

SCT + TRT example: The BCIDs of the SCT and TRT VALUE IS 1 IF EVERYTHING IS OK THEN SCT & TRT ARE READING SAME EVENT! ROD ID DIFFERENT BCID BETWEEN SCT AND TRT - SOMETHING IS WRONG

Online display: Data Quality Monitoring (DQM)‏ Automatic checks on data quality using predefined algorithms on histograms or counters etc. Input from Online Histogram Service Can also be done offline using ROOT files Puts flags or alarms if values are off expected limits or e.g. histograms are empty States are green, yellow, red States per subsystem but also overall state

Milestone 6 (M6) Cosmic Run Global Inner Detector monitoring included for the first time Basic functionalities fully working Looked at the online histograms Did offline work – reported results Work will continue to improve histograms what they display, how and which to choose

CONCLUSIONS Monitoring is crucial for good data-taking Several monitoring applications exist and are working M6 was first time the Global Inner Detector was fully tested Doing shift-work really lets you learn a lot about ATLAS and the jungle of software that is out there

MORE INFO, not included in talk

My contribution to the inner- detector monitoring In addition to doing this work partially from Oslo and partially at CERN, I have taken online monitoring shifts Looked at the online monitoring histograms, and did some offline “analysis” looking into number of tracks registered in the Inner detector as a whole (only SCT and TRT included in M6) and SCT and TRT separately and reported this in a spontaneously organized track-number meeting Reported Inner Detector achievements for Heidi at a weekly SCT Commisioning meeting A great chance to see how everything is working and to learn a lot about the detector and how such a huge experiment works

My contribution to the inner- detector monitoring Helped including inner-detector into TDAQ system – to enable to read data from the inner-detector. Done by several tools; output xml files – these are the configuration files. Config files fed into the TDAQ application - when the run- controllers start taking data our detector will show up in the application and will be sending data Another task: including histograms from the inner-detector into OHP. Yet another xml file had to be written correctly in order to subscribe to the histograms which are already produced and filled when the detector turns on For M6 all the basic functionalities were working for the first time!

Atlas Online Monitoring System Highly distributed, different processes run on different machines at once Low level: hardware states, noisy modules, dead channels, electronic mappings, data quality e.g. Synchronization btw different sub-detectors High-level (Athena): run on full events, check physics quantities momentum, spin etc. Both produce histograms presented in ROOT TDAQ (Trigger and data quality ) group provide useful services e.g.: Online Histogram Service OHS provides input to OHP (Online Histogram Presenter)

Monitoring TDR Input retrieved from:  For hardware, the electronics are monitored from subdetector front-ends by the Detector Control System (DCS)‏  DataFlow elements from RODs, ROSs, SFIs, SFOs done at operational level  Output sent to end-user that means shifter such as me or you, via histograms, flags, values etc First easy monitoring access RODs  Digital Signal Processor installed directly on the ROD board. The CPUs on the ROD crate will send the histograms to the end-user.

Monitoring TDR ROS next level  Larger regions can be monitored (than from RODs)‏ When info from several detectors is needed  Do after Event Builder level  SFI first place where fully built events are available

Monitoring TDR Dataflow ROD crate DAQ all software and hardware that configures, controls, monitors 1 or more ROD crate systems independently on the rest of the data flow. ROS

Online Monitoring with Athena AthenaPT (PT: processing task) tool to integrate TDAQ HLT and Athena software components in one environment Takes care of interface between offline athena and the online needs. Each sub-detectors expert writes algorithms for filling histograms with the needed information, f.ex. For innerdetector fill histograms with noise occupancy for the SCT versus TRT AthenaMonitoring package framework takes care of producing histograms in Athena The interface between the detector and the Histogram repository is already taken care of Running Athena algorithm in PT can mean either online mode with DAQ & OKS partition

Monitoring TDR Online Software Used to configure, control and monitor the TDAQ system Responsible for synchronizing involved sub- systems during start-up and shut-down Online software distiguishes various types of users:  TDAQ Operator: runs TDAQ system in operating room during data-taking period  TDAQ Expert: Can perform changes to configuration has system internal knowledge  Sub-system expert, Detector expert responsible for operation of particular sub-system or detector

Information Service: Deployment DAQ Workstation DAQ Control Workstation IS Server ROS EB DAQ Application DAQ Application insert update remove notify subscribe ROD Crate DAQ Application ROD Crate DAQ Application ROD Crate DAQ Application DAQ Application DAQ Control Workstation IS Server DAQ Workstation IS Server get_value STOLEN FROM: Serguei Kolos CERN/PNPI ATLAS Trigger/DAQ Online Software group OHP DQMF

Data Storage Event Filter (EF)‏ Event Builder (EB)‏ HLT/DAQ/DCS system Detector Control System (DCS)‏ LVL2 Trigger Read Out Systems (ROSs)‏ Online Software: Configure Control Monitoring PixelTileCalLArMDTCSCSCTTRT Calorimeter Inner Detector Muon Spectrometer RPC TGC RODs STOLEN FROM: Serguei Kolos CERN/PNPI ATLAS Trigger/DAQ Online Software group

Forward SCT TRT Pixel Detectors Barrel SCT 6 m 2 m

Inner Detector – Silicon Pixel detector (Pixel)‏  Cabling is finishing these days – SemiConductor Tracker(SCT)‏  Barrell and endcap fully installed‏ ~1/2year ago – Transition Radiation Tracker (TRT)‏  TRT fully installed ~year ago SCT and TRT were tested together on the surface about a year ago First cosmic test with SCT were last fall

ATLAS Detector Layout Each partition may be operated independently Some partitions may be operated in parallel ~1000 Read-Out Drivers (RODs) in ~100 VME crates 33 sub-detector Partitions PixelTileCalLArMDTCSCSCTTRT Calorimeter Inner Detector Muon Spectrometer RPC TGC STOLEN FROM: Serguei Kolos CERN/PNPI ATLAS Trigger/DAQ Online Software group

Monitoring Services responsible for routing information and histograms IS: Main task: transport monitoring data request from monitoring destination to monitoring sources and to transport monitoring data back from the sources to the destinations Histogramming Service: a specialization of the IS used to transport histograms Allows different applications to exchange histograms Has a user-interface via application called Histogram Display Error Reporting Service Event Monitoring Service: transports physics events or fractions of events Event sampler

TDAQ: Trigger and Data Acquisition System HLT:High Level Trigger EF: Event Filter DF: Data Flow software EDF:Event Data Flow PT:Processing Task ROD: Read Out Driver ROS: Read Out Subsystem RODC:Read Out Driver Crate ROC: Read Out Crate EMON: Atlas Event Monitoring – sampler system ACRONYMS

Atlantis:Atlas Standalone Event Display Athena: Atlas Offline software framework DQM: Data Quality Monitoring DQMF: Atlas Data Quality Framework OH: Online Histogramming service IS: Information Service DCS: Detector Control System SFI: SubFarm Input SFO: SubFarm Output DSP:Digital Signal Processor OKS: Object Kernel Support  Library to support simple in-memoryobject manager. Suitable for real-timeobject manager for e.g. Data Acquisition ACRONYMS