The Run Control and Monitoring System of the CMS Experiment Presented by Andrea Petrucci INFN, Laboratori Nazionali di Legnaro, Italy On behalf of the.

Slides:



Advertisements
Similar presentations
GNAM and OHP: Monitoring Tools for the ATLAS Experiment at LHC GNAM and OHP: Monitoring Tools for the ATLAS Experiment at LHC M. Della Pietra, P. Adragna,
Advertisements

Performance analysis of a Grid-based Instrumentation Device Farm Luca Berruti 1, Franco Davoli 1, Stefano Vignola 1, Sandro Zappatore 1 1 CNIT – University.
CHEP 2012 – New York City 1.  LHC Delivers bunch crossing at 40MHz  LHCb reduces the rate with a two level trigger system: ◦ First Level (L0) – Hardware.
DataGrid is a project funded by the European Union 22 September 2003 – n° 1 EDG WP4 Fabric Management: Fabric Monitoring and Fault Tolerance
CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/27 A Control Software for the ALICE High Level Trigger Timm.
CMS Michele Gulmini, CHEP2003, San Diego USA, March Run Control and Monitor System for the CMS Experiment Michele Gulmini CERN/EP – INFN Legnaro.
Data Quality Monitoring for CMS RPC A. Cimmino, D. Lomidze P. Noli, M. Maggi, P. Paolucci.
First year experience with the ATLAS online monitoring framework Alina Corso-Radu University of California Irvine on behalf of ATLAS TDAQ Collaboration.
Data Acquisition Software for CMS HCAL Testbeams Jeremiah Mans Princeton University CHEP2003 San Diego, CA.
Java Based Run Control for CMS Small DAQ Systems Michele Gulmini CHEP2000 February 2000 M. Bellato (INFN – Padova) L. Berti (INFN – Legnaro) D. Ceccato.
First operational experience with the CMS Run Control System Hannes Sakulin, CERN/PH on behalf of the CMS DAQ group 17 th IEEE Real Time Conference,
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
GRID job tracking and monitoring Dmitry Rogozin Laboratory of Particle Physics, JINR 07/08/ /09/2006.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Calo Piquet Training Session - Xvc1 ECS Overview Piquet Training Session Cuvée 2012 Xavier Vilasis.
CMS Michele Gulmini, Cern, DAQ Weekly 07/05/ RCMS – Plan of work Michele Gulmini DAQ Weekly 7th May 2002.
D.S.I. Venezia L.N.L CMS Error Handling in RCMS An Overview Francesco Lelli
F. Lelli, Advanced Training Workshop on Instruments and Sensors on the Grid,Trieste, April 25, 2007 Instrument Elements at Work Imperial Brunel Elettra.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
GT Components. Globus Toolkit A “toolkit” of services and packages for creating the basic grid computing infrastructure Higher level tools added to this.
JCOP Workshop September 8th 1999 H.J.Burckhart 1 ATLAS DCS Organization of Detector and Controls Architecture Connection to DAQ Front-end System Practical.
IceCube DAQ Mtg. 10,28-30 IceCube DAQ: “DOM MB to Event Builder”
CMS Luigi Zangrando, Cern, 05/03/ RCMS for XDaq based small DAQ Systems M. Gulmini, M. Gaetano, N. Toniolo, S. Ventura, L. Zangrando INFN – Laboratori.
C.Combaret, L.Mirabito Lab & beamtest DAQ with XDAQ tools.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
RPC PAC Trigger system installation and commissioning How we make it working… On-line software Resistive Plate Chambers Link Boxes Optical Links Synchronization.
Contents 1.Introduction, architecture 2.Live demonstration 3.Extensibility.
G. Maron, Agata Week, Orsay, January Agata DAQ Layout Gaetano Maron INFN – Laboratori Nazionali di Legnaro.
André Augustinus 10 September 2001 DCS Architecture Issues Food for thoughts and discussion.
Databases E. Leonardi, P. Valente. Conditions DB Conditions=Dynamic parameters non-event time-varying Conditions database (CondDB) General definition:
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
André Augustinus 10 October 2005 ALICE Detector Control Status Report A. Augustinus, P. Chochula, G. De Cataldo, L. Jirdén, S. Popescu the DCS team, ALICE.
Control in ATLAS TDAQ Dietrich Liko on behalf of the ATLAS TDAQ Group.
Gnam Monitoring Overview M. Della Pietra, D. della Volpe (Napoli), A. Di Girolamo (Roma1), R. Ferrari, G. Gaudio, W. Vandelli (Pavia) D. Salvatore, P.
ALICE, ATLAS, CMS & LHCb joint workshop on
7. CBM collaboration meetingXDAQ evaluation - J.Adamczewski1.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Introduction CMS database workshop 23 rd to 25 th of February 2004 Frank Glege.
Overview of DAQ at CERN experiments E.Radicioni, INFN MICE Daq and Controls Workshop.
Incorporating Scientific Instruments into the Grid Francesco Lelli University of Lugano F. Lelli, Incorporating Scientific Instruments into.
Gaetano Maron, CPT week, CERN, 18 April Run Control and Conditions DB CMS CPT Week, CERN 18 April 2002 G. Maron INFN – Laboratori Nazionali di Legnaro.
DAQ Andrea Petrucci 6 May 2008 – CMS-UCSD meeting OUTLINE Introduction SCX Setup Run Control Current Status of the Tests Summary.
August 2003 At A Glance The IRC is a platform independent, extensible, and adaptive framework that provides robust, interactive, and distributed control.
CMS Luigi Zangrando, Cern, 16/4/ Run Control Prototype Status M. Gulmini, M. Gaetano, N. Toniolo, S. Ventura, L. Zangrando INFN – Laboratori Nazionali.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
Michele de Gruttola 2008 Report: Online to Offline tool for non event data data transferring using database.
Management of the LHCb Online Network Based on SCADA System Guoming Liu * †, Niko Neufeld † * University of Ferrara, Italy † CERN, Geneva, Switzerland.
First CMS Results with LHC Beam
First EELA Conference, Santiago(Cile), 4th September, GridCC: Real-time Instrumentations Grids.
Software for the CMS Cosmic Challenge Giacomo BRUNO UCL, Louvain-la-Neuve, Belgium On behalf of the CMS Collaboration CHEP06, Mumbay, India February 16,
The DCS Databases Peter Chochula. 31/05/2005Peter Chochula 2 Outline PVSS basics (boring topic but useful if one wants to understand the DCS data flow)
ATLAS Distributed Analysis DISTRIBUTED ANALYSIS JOBS WITH THE ATLAS PRODUCTION SYSTEM S. González D. Liko
CMS Luigi Zangrando, Cern, 16/4/ Run Control Prototype Status M. Gulmini, M. Gaetano, N. Toniolo, S. Ventura, L. Zangrando INFN – Laboratori Nazionali.
ATLAS The ConditionDB is accessed by the offline reconstruction framework (ATHENA). COOLCOnditions Objects for LHC The interface is provided by COOL (COnditions.
M. Caprini IFIN-HH Bucharest DAQ Control and Monitoring - A Software Component Model.
Online Software November 10, 2009 Infrastructure Overview Luciano Orsini, Roland Moser Invited Talk at SuperB ETD-Online Status Review.
20OCT2009Calo Piquet Training Session - Xvc1 ECS Overview Piquet Training Session Cuvée 2009 Xavier Vilasis.
EPS 2007 Alexander Oh, CERN 1 The DAQ and Run Control of CMS EPS 2007, Manchester Alexander Oh, CERN, PH-CMD On behalf of the CMS-CMD Group.
The Control and Hardware Monitoring System of the CMS Level-1 Trigger Ildefons Magrans, Computing and Software for Experiments I IEEE Nuclear Science Symposium,
This project is supported under EU FP6 Contract Istituto Nazionale di Fisica Nucleare, Legnaro, Italy Brunel University, Uxbridge, UK IBM, Haifa,
Detector Control System
CMS Trigger Supervisor Framework
CMS High Level Trigger Configuration Management
LHC experiments Requirements and Concepts ALICE
Controlling a large CPU farm using industrial tools
RCMS Internet - Intranet UI 9-1.
Design Principles of the CMS Level-1 Trigger Control and Hardware Monitoring System Ildefons Magrans de Abril Institute for High Energy Physics, Vienna.
Tools for the Automation of large distributed control systems
SDMX IT Tools SDMX Registry
Presentation transcript:

The Run Control and Monitoring System of the CMS Experiment Presented by Andrea Petrucci INFN, Laboratori Nazionali di Legnaro, Italy On behalf of the DAQ Group of CMS collaboration ACAT 2007, April 2007, Amsterdam, Netherlands

2 Outline 2ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN Run Control and Monitor System :  Architecture Logical Layer Services Components Technologies  At the Magnet Test and Cosmic Challenge (MTCC) Control structure Operation Components Results  GRICC Project

3 What is CMS? 3ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN The Compact Muon Solenoid (CMS) experiment is one of two large general-purpose particle physics detectors being built on the proton- proton Large Hadron Collider (LHC) at CERN in Switzerland. to explore physics at the TeV scale to discover the Higgs boson to look for evidence of physics beyond the standard model to be able to study aspects of heavy ion collisions The main goals of the experiment are:

4 Run Control and Monitor System 4ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN The Run Control and Monitor System (RCMS) is responsible for controlling and monitoring the CMS experiment during the data taking. RCMS views the experiment as a set of partition, where a partition is a grouping of entities that can be operated independently. Main operations are configuration, monitoring, error handling, logging and synchronization with other subsystems.

55ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN Baseline DAQ Configuration 512 inputs 2024 outputs CMS Data Acquisition Control and Monitor requirements O(10 4 ) distributed Objects to – control – configure – monitor On-line diagnostics Interactive system

6 RCMS is integrated in the CMS On-line system : It controls the “DAQ component” –Data transport –Event processing It monitors the “Detector Control System” DCS – manages the slow controls of the whole experiment. 6ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN Run Control and Monitor System The SOAP protocol and the Web Services have been adopted as the main means for communication. The online process environment is XDAQ, a C++ framework for a distributed Data Acquisition System.

7 RCMS Logical Structure 7ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN A Session is the allocation of the hardware and software of a CMS partition needed to perform data- taking. Multiple Sessions may coexist and operate concurrently. Each Session is associated with a Top Function Manager, that coordinates all the actions. Top Sub-Detector DAQ Resources Services

8 –SECURITY SERVICE login and user account management; –RESOURCE SERVICE (RS) information about DAQ resources and partitions; –INFORMATION AND MONITOR SERVICE (IMS) Collects messages and monitor data; distributes them to the subscribers; –JOB CONTROL Starts, monitors and stops the software elements of RCMS, including the DAQ components; RCMS Services 8ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN

9 Function Manager 9ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN Input Handler : It handles all the input events of the FM (GUIs or other FMs, errors, states, logs and monitor messages) Event Processor: It handles all the incoming message and decide where to send them. It has processing capability Finite State Machine (FSM): The behavior of the FM is driven by a FSM. Resource Proxy: It handles all the outgoing connections with the resources. The purpose of a Function Manager (FM) is to control a set of resources. Input Handler Event Processor FSM Engine Resource Proxy Function Manager Resources State Flow Error Flow Monitor Flow Control Flow Customizable

10 Resource Service 10ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN RS Manager tool RS DB RS API DAQ Configurator The Resource Service (RS) stores the process configuration of the On-line System. features Flexible data store Java API Configuration documents can be built on the fly from relational schema Versioning system Oracle and MySQL Implementation

11 Log Collector 11ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN Collects log information from log4j compliant applications (i.e. on-line process). … Publish Subscriber System Display System Storage System Log Collector Relational DB Oracle,MySQL Message System Access via JDBC Access via TCP RCMS applications and XDAQ applications Send log information directly to a Display System (Chainsaw). Stores log information in a database and visualizes them (LogDBViewer). Distributes/publishes log information through a message system (Java Message Service).

12 RCMS main components 12ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN Log Collector WebApp Chainsaw Log Viewer JobControl Process Control Firefox JSP/Ajax GUI Config data Conditions data Process Config Log Messages Commands Notifications RCMS RS Manager Manager/Editor DAQ Configurator Configuration Function Manager Framework RS API RunInfo API Hwcfg API RS DB GCK DB RunInfo DB Log DB Hwcfg DB User Interface

13 RCMS Technologies 13ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN Technologies and tools: Web Applications,Java Servlets (Apache Tomcat) WebService (Axis, WSDL, SOAP) Web Tecnologies (Ajax,JSP) Databases –Oracle –MySQL ArchitectureImplementation Resource Service (RS)Resource Service Information and Monitor Service (IMS)LogCollector SubSystem Controllers (FMs)RCMS Framework Top Function ManagerRCMS Framework GUIsDefault JSP GUI - RCMS Framework JobControlXDAQ Framework

14 Magnet Test and Cosmic Challenge 14ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN The main goals of the Cosmic Challenge were: Test Muon alignment systems. Commission the several sub-detectors (Drift Tubes - DT, Hadron Calorimeter – HCAL, Tracker, etc.) and Cosmic Trigger. demonstrate cosmic ray reconstruction with multiple sub-detectors. The Magnet Test and Cosmic Challenge (MTCC) is a milestone of the CMS experiment, it completes the commissioning of the magnet system (coil & yoke) before its lowering into the cavern. Scale MTCC versus CMS Data Sources:20 out of 600 3% Filter Nodes: 14 out of % Trigger rate: 100 Hz out of 100 kHz 0.1% Event size: 200 kB out of 1 MB20% Scale MTCC versus CMS Data Sources:20 out of 600 3% Filter Nodes: 14 out of % Trigger rate: 100 Hz out of 100 kHz 0.1% Event size: 200 kB out of 1 MB20%

15 FMs Control Structure at MTCC I & II 15ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN Web Browser (GUI) Level 0 FM Level 1 FM Level 2 FM User interaction with Web Browser connected to Level 0 FM. Level 0 FM is entry point to Run Control System. Level 2 FMs are sub-system specific custom implementations. Level 1 FM interface to the Level 0 FM and have to implement a standard set of inputs and states. TOP LTC CSCDAQ RPCDT TRK ECAL HCAL FBRBFF Resources FECFED Resources are on-line system components

16 RCMS at MTCC I & II ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN TopDAQ LTC HCAL ECAL DT CSC RPC TRK RCMS Operation Scenario –Sub-system function managers were written using the RCMS software –The run configuration was communicated via a global configuration key –The Run Info DB was used to store end-of- run summary information and status information about the run. It also contained the schema to generate Run Numbers and Run Sequence Numbers. N Sub-Detector controlled8 Function Managers used14 Online resources controlled~ 100

17 RCMS Components at MTCC I & II 17ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN global key local key configuration logmessage TopDAQ LTC-Trg HCAL ECAL DT CSC RPC Global Configuration Keys RPC RS RPC RS LTC-Trg RS LTC-Trg RS EMU RS EMU RS DT RS DT RS DAQ RS DAQ RS global key ECAL RS ECAL RS HCAL RS HCAL RS TRK RS TRK RS LOG DB Top RS Top RS Collector Configuration A Global configuration Key identified a sub- system configuration. The configuration local to the sub-system were decouple from each other and the top configuration.

18 MTCC Data taking 18ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN

19 MTCC result 19ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN RCMS software was stable. Separation of Subsystem installations worked well. Recorded ~ 160 M events on a period of one month

20 RCMS and GRIDCC 20ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN It is a project of 3-years and started in September 2004 Web site: What is GRIDCC ? The Grid enabled Remote Instrumentation with Distributed Control and Computation (GRIDCC) is a project funded by the European community, aimed to provide access to and control of distributed complex instrumentation. The CMS RCMS is one of the main applications for the GRIDCC project. The RCMS software is the core of the Instrument Element of the GRIDCC.

21 ACAT April 2007, AmsterdamAndrea Petrucci - LNL-INFN Thank you for your attention. Any Questions?