CMS Michele Gulmini, CHEP2003, San Diego USA, March 2003 1 Run Control and Monitor System for the CMS Experiment Michele Gulmini CERN/EP – INFN Legnaro.

Slides:



Advertisements
Similar presentations
TeraGrid Deployment Test of Grid Software JP Navarro TeraGrid Software Integration University of Chicago OGF 21 October 19, 2007.
Advertisements

Performance analysis of a Grid-based Instrumentation Device Farm Luca Berruti 1, Franco Davoli 1, Stefano Vignola 1, Sandro Zappatore 1 1 CNIT – University.
CHEP 2012 – New York City 1.  LHC Delivers bunch crossing at 40MHz  LHCb reduces the rate with a two level trigger system: ◦ First Level (L0) – Hardware.
DataGrid is a project funded by the European Union 22 September 2003 – n° 1 EDG WP4 Fabric Management: Fabric Monitoring and Fault Tolerance
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Chapter 9: Moving to Design
March 2003 CHEP Online Monitoring Software Framework in the ATLAS Experiment Serguei Kolos CERN/PNPI On behalf of the ATLAS Trigger/DAQ Online Software.
Institute of Computer Science AGH Performance Monitoring of Java Web Service-based Applications Włodzimierz Funika, Piotr Handzlik Lechosław Trębacz Institute.
First year experience with the ATLAS online monitoring framework Alina Corso-Radu University of California Irvine on behalf of ATLAS TDAQ Collaboration.
Data Acquisition Software for CMS HCAL Testbeams Jeremiah Mans Princeton University CHEP2003 San Diego, CA.
Java Based Run Control for CMS Small DAQ Systems Michele Gulmini CHEP2000 February 2000 M. Bellato (INFN – Padova) L. Berti (INFN – Legnaro) D. Ceccato.
First operational experience with the CMS Run Control System Hannes Sakulin, CERN/PH on behalf of the CMS DAQ group 17 th IEEE Real Time Conference,
PVSS SOAP Interface RCMS  DCS XDAQ  DCS (anyclient  PVSS) Peter Rosinsky, CERN PH/CMD CMS week 15-19/3/2004.
Chapter 9 Elements of Systems Design
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
XCAT Science Portal Status & Future Work July 15, 2002 Shava Smallen Extreme! Computing Laboratory Indiana University.
Framework for Automated Builds Natalia Ratnikova CHEP’03.
CMS Michele Gulmini, Cern, DAQ Weekly 07/05/ RCMS – Plan of work Michele Gulmini DAQ Weekly 7th May 2002.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
The Run Control and Monitoring System of the CMS Experiment Presented by Andrea Petrucci INFN, Laboratori Nazionali di Legnaro, Italy On behalf of the.
Ramiro Voicu December Design Considerations  Act as a true dynamic service and provide the necessary functionally to be used by any other services.
CMS Luigi Zangrando, Cern, 05/03/ RCMS for XDaq based small DAQ Systems M. Gulmini, M. Gaetano, N. Toniolo, S. Ventura, L. Zangrando INFN – Laboratori.
C.Combaret, L.Mirabito Lab & beamtest DAQ with XDAQ tools.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
G. Maron, Agata Week, Orsay, January Agata DAQ Layout Gaetano Maron INFN – Laboratori Nazionali di Legnaro.
André Augustinus 10 September 2001 DCS Architecture Issues Food for thoughts and discussion.
LHCb week, 27 May 2004, CERN1 Using services in DIRAC A.Tsaregorodtsev, CPPM, Marseille 2 nd ARDA Workshop, June 2004, CERN.
Event-Based Hybrid Consistency Framework (EBHCF) for Distributed Annotation Records Ahmet Fatih Mustacoglu Advisor: Prof. Geoffrey.
Control in ATLAS TDAQ Dietrich Liko on behalf of the ATLAS TDAQ Group.
ALICE, ATLAS, CMS & LHCb joint workshop on
7. CBM collaboration meetingXDAQ evaluation - J.Adamczewski1.
9 Systems Analysis and Design in a Changing World, Fourth Edition.
9 Systems Analysis and Design in a Changing World, Fourth Edition.
Introduction CMS database workshop 23 rd to 25 th of February 2004 Frank Glege.
Overview of DAQ at CERN experiments E.Radicioni, INFN MICE Daq and Controls Workshop.
Gaetano Maron, CPT week, CERN, 18 April Run Control and Conditions DB CMS CPT Week, CERN 18 April 2002 G. Maron INFN – Laboratori Nazionali di Legnaro.
CHEP March 2003 Sarah Wheeler 1 Supervision of the ATLAS High Level Triggers Sarah Wheeler on behalf of the ATLAS Trigger/DAQ High Level Trigger.
Gaetano Maron, CPT week, CERN, 16 April RCS Discussion.
Architecture View Models A model is a complete, simplified description of a system from a particular perspective or viewpoint. There is no single view.
DAQ Andrea Petrucci 6 May 2008 – CMS-UCSD meeting OUTLINE Introduction SCX Setup Run Control Current Status of the Tests Summary.
CMS Luigi Zangrando, Cern, 16/4/ Run Control Prototype Status M. Gulmini, M. Gaetano, N. Toniolo, S. Ventura, L. Zangrando INFN – Laboratori Nazionali.
Overview of Grid Webservices in Distributed Scientific Applications Dennis Gannon Aleksander Slominski Indiana University Extreme! Lab.
Status & development of the software for CALICE-DAQ Tao Wu On behalf of UK Collaboration.
TDAQ Experience in the BNL Liquid Argon Calorimeter Test Facility Denis Oliveira Damazio (BNL), George Redlinger (BNL).
Daniele Spiga PerugiaCMS Italia 14 Feb ’07 Napoli1 CRAB status and next evolution Daniele Spiga University & INFN Perugia On behalf of CRAB Team.
G. Anders, G. Avolio, G. Lehmann Miotto, L. Magnoni CERN, Geneva, Switzerland The Run Control System and the Central Hint and Information Processor of.
Site Authorization Service Local Resource Authorization Service (VOX Project) Vijay Sekhri Tanya Levshina Fermilab.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
CMS Luigi Zangrando, Cern, 16/4/ Run Control Prototype Status M. Gulmini, M. Gaetano, N. Toniolo, S. Ventura, L. Zangrando INFN – Laboratori Nazionali.
Maria del Carmen Barandela Pazos CERN CHEP 2-7 Sep 2007 Victoria LHCb Online Interface to the Conditions Database.
M. Caprini IFIN-HH Bucharest DAQ Control and Monitoring - A Software Component Model.
Online Software November 10, 2009 Infrastructure Overview Luciano Orsini, Roland Moser Invited Talk at SuperB ETD-Online Status Review.
MONITORING CMS TRACKER CONSTRUCTION AND DATA QUALITY USING A GRID/WEB SERVICE BASED ON A VISUALIZATION TOOL G. ZITO, M.S. MENNEA, A. REGANO Dipartimento.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
EPS 2007 Alexander Oh, CERN 1 The DAQ and Run Control of CMS EPS 2007, Manchester Alexander Oh, CERN, PH-CMD On behalf of the CMS-CMD Group.
The Control and Hardware Monitoring System of the CMS Level-1 Trigger Ildefons Magrans, Computing and Software for Experiments I IEEE Nuclear Science Symposium,
CMS Trigger Supervisor Framework
CMS High Level Trigger Configuration Management
Controlling a large CPU farm using industrial tools
OO Methodology OO Architecture.
Hannes Sakulin, CERN/EP on behalf of the CMS DAQ group
ProtoDUNE SP DAQ assumptions, interfaces & constraints
RCMS Structure RCMS Security UserDB Service Resource ConfDB
RCMS Internet - Intranet UI 9-1.
M. Gulmini, G, Maron, N. Toniolo, L. Zangrando
Design Principles of the CMS Level-1 Trigger Control and Hardware Monitoring System Ildefons Magrans de Abril Institute for High Energy Physics, Vienna.
The Performance and Scalability of the back-end DAQ sub-system
Pierluigi Paolucci & Giovanni Polese
Tools for the Automation of large distributed control systems
Presentation transcript:

CMS Michele Gulmini, CHEP2003, San Diego USA, March Run Control and Monitor System for the CMS Experiment Michele Gulmini CERN/EP – INFN Legnaro On behalf of the CMS DAQ collaboration CHEP 2003, San Diego USA, March 2003

CMS Michele Gulmini, CHEP2003, San Diego USA, March Outline Run Control and Monitor System : RCMS RCMS Architecture Session Managers Subsystem Controllers Services RCMS Prototypes RCMS for Small DAQ Systems RCMS Demonstrators –Performance and Scalability Tests Plans Summary

CMS Michele Gulmini, CHEP2003, San Diego USA, March Run Control and Monitor System RCMS Internet Intranet Internet Intranet UI The Run Control and Monitor System (RCMS) is the collection of hardware and software components responsible for controlling and monitoring the CMS experiment during the data taking. RCMS enables users to access and control the experiment from any part in the world providing a “virtual counting room”, where physicists and operators can effectively taking shifts from a distance. RCMS views the experiment as a set of partitions, where a partition is a grouping of entities that can be operated independently. Main operations are configuration, monitoring, error handling, logging and synchronization with other subsystems. RCMS Trigger Event Builder Event Filter DCS Computing Services UI RCMS Context

CMS Michele Gulmini, CHEP2003, San Diego USA, March Partitions Example Session Manager-A UI Services Connection Services Services CS Ctrl TRG Ctrl DCS Ctrl EVF Ctrl CS Sub- System Glbl Mu Cal DCS Sub- System EVB Sub-System TRG Sub-System EVF Sub-System Session Manager-B UI EVB Ctrl FED Builder Sub-System RU Builder Sub-System FED-B Ctrl RU-B Ctrl

CMS Michele Gulmini, CHEP2003, San Diego USA, March RCMS Logical Layout The execution of the RCMS is organized on the basis of “Sessions”. A Session is the allocation of the hardware and software of a CMS partition needed to perform data-taking. Multiple Sessions may coexist and operate concurrently Each Session is associated with a Session Manager (SMR), that coordinates all the actions

CMS Michele Gulmini, CHEP2003, San Diego USA, March Sub-System Controller (SSC) A SSC consists of a Function Manager (FM) and a local database (DB) service. There is one FM per partition that receives requests from a Session Manager (SMR) and transforms them into the corresponding requests for actions that are sent to the sub-system. The local DB service can be used as a proxy to the services.

CMS Michele Gulmini, CHEP2003, San Diego USA, March Basic RCMS Services –SECURITY SERVICE login and user account management; –RESOURCE SERVICE (RS) information about DAQ resources and partitions; –INFORMATION AND MONITOR SERVICE (IMS) Collects messages and monitor data; distributes them to the subscribers; –JOB CONTROL Starts, monitors and stops the software elements of RCMS, including the DAQ components –PROBLEM SOLVER Uses information from the RS and IMS to identify mulfunctions and attempts to provide automatic recovery procedures where applicable

CMS Michele Gulmini, CHEP2003, San Diego USA, March Resource Service Block Diagram The Resource Service (RS) handles all the hardware and software components of the DAQ system including its partitions. SS UserDB RS IMS Job Ctrl PS SSC Services Connection Session Manager RCMS UI ConfDB LogDB

CMS Michele Gulmini, CHEP2003, San Diego USA, March Information and Monitor Service Block Diagram The Information and Monitor Service (IMS) collects the information (log, warning, errors, monitoring, etc.) from the sub-systems and provides them to the subscribers. SS UserDB RS IMS Job Ctrl PS SSC Services Connection Session Manager RCMS UI ConfDB LogDB

CMS Michele Gulmini, CHEP2003, San Diego USA, March Time Requirements –Configuration and setup of the system: minutes –Control (state change, execution of commands): seconds –Monitoring: depending on the amount of data required Information and Monitor Service: Tens of subscribers Peak: about 2000 messages (status change, log) Average: Tens to a few hundred messages/s

CMS Michele Gulmini, CHEP2003, San Diego USA, March RCMS Prototypes RCMS for small DAQ Systems –Fully functional RCMS systems targeted to small DAQs (Production systems, Testbeam DAQ systems) –Real-life examples used to check the RCMS functionality. RCMS demonstrators –Partially functional RCMS systems targeted to prove scalability issues. –Test bed systems used to emulate slices or parts of the hierarchical structure of the final DAQ. –Help to confirm the architecture and to evaluate and eventually select the technologies to be used in the final system.

CMS Michele Gulmini, CHEP2003, San Diego USA, March RCMS for small DAQs Current Running Prototype: –Designed to work together with XDAQ CMS online software framework (XDAQ: See Chep2003 J. Gutleber talk - “Using XDAQ in Application Scenarios of the CMS Experiment”) –Available services: Resource Service (RS) Information and Monitor Service (IMS) SubSystem Controllers (Function Managers) Session Managers GUIs Technologies and tools: Java Servlets (Apache Tomcat) Sun “Java Web Services Developer Package” (JWSDP) –JAXP, JAXM, XPath,... SOAP communication protocol Databases –XMLDB interface »eXist native XML database –mySQL

CMS Michele Gulmini, CHEP2003, San Diego USA, March RCMS for Small DAQs – Current Applications CMS Muon Drift Tubes Chamber Production DAQ (Legnaro - Italy) Testbeam (CERN – next May) CMS Tracker “ROD System Tests” (CERN) Testbeam (CERN – next May) CMS TriDAS (CERN) DAQ Column TDR Demonstrator

CMS Michele Gulmini, CHEP2003, San Diego USA, March Session and Function Manager Prototype SS UserDB RS IMS Job Ctrl PS SSC Services Connection Session Manager RCMS UI ConfDB LogDB XML definition JavaImplementation FSM SM/FM servlet Function Managers and Session Manager have a built in Finite State Machine (FSM) to command the controlled components, and to track their state; The FSM is composed of a XML definition and a Java class implementation representing the actions to be performed; The definition and the implementation of the FSMs are managed by the Resource Service; Session Manager and Function Managers are launched when a new “Session” is opened, and can have a hierarchical structure;

CMS Michele Gulmini, CHEP2003, San Diego USA, March RS and IMS Prototype IMS Xpath Filter Engine JAXM XML message Tomcat servlet container NOTIFY PUBLISH Subs Info JDOM FS Java Publisher JAXM Java Subscriber JAXM SUBSCRIBE Tomcat/Jetty Soap DB (eXist, File,mySQL) XDAQ Application XOAP XMLDB Servlet container (TOMCAT) Java Servlet Resource Service XML Java client Java Objs XML Parser (CASTOR) XML Parser C++ client XML Parser XML DB Inter face REL DB XML:DB SOAP

CMS Michele Gulmini, CHEP2003, San Diego USA, March RCMS GUIs Generic GUI: –Insertion and retrieval of resources (PCs, software, partitions, etc.) –Ability to command, set and retrieve parameters from XDAQ applications –Scripting facility –Customisation facilities (plugins) Muon DT TestBeam GUI

CMS Michele Gulmini, CHEP2003, San Diego USA, March Legnaro T2 CMS farm: 136 P GHz processors RCMS Demonstrators

CMS Michele Gulmini, CHEP2003, San Diego USA, March Demonstrator 1 Exploring the ability to command a set of XDAQ executives running “empty” applications The time measured represents the time required to perform a state change of the entire cluster..... XDAQ PC XDAQ FM FM: Function Manager..... XDAQ XDAQ FM..... XDAQ XDAQ FM FM PC SOAP

CMS Michele Gulmini, CHEP2003, San Diego USA, March Demonstrator nodes 100 ms

CMS Michele Gulmini, CHEP2003, San Diego USA, March Demonstrator 2 Simplified version of a log message service based on Web Services technologies (Glue platform) 15 clients and a variable number of Web Services used The performance scales linearly with the number of instances of the service available

CMS Michele Gulmini, CHEP2003, San Diego USA, March IMS Prototype Test (I) PUBLISH IMS IMS IMS Publisher mySQL DB Percistency on eXist XML native DB not plotted – very slow Between 200 and 300 SOAP messages/s handled by the IMS prototype

CMS Michele Gulmini, CHEP2003, San Diego USA, March IMS Prototype Test (2) PUBLISH NOTIFY IMS IMS IMS IMS IMS Publisher IMS IMS Subscriber SUBSCRIBE mySQL DB SOAP Performance improves augmenting the number of service instances Notification mechanism not optimized Test to be completed

CMS Michele Gulmini, CHEP2003, San Diego USA, March IMS hierarchical structure –Performance test done with the present prototypes: –Commanding a cluster of DAQ application fits the requirements –Information and Monitor Service prototype needs further investigation –Notification architecture –Hierarchical structure IMS hierarchical structure:..... IMS IMS proxy XDAQ XDAQ..... XDAQ XDAQ

CMS Michele Gulmini, CHEP2003, San Diego USA, March Future – OGSA??? RCMS architecture is service and web oriented Web services development tools (Apache Axis, Glue) may help to deploy reliable services quickly Open Grid Service Architecture (OGSA) ( is Web Services based An alpha release of the framework is now available First official release foreseen in a few months time OGSA could be adopted for the RCMS services, providing several advantages: RCMS open to the Grid world Well supported and reliable framework Useful built-in services OGSA is under evaluation: The RCMS Resource Service has been successfully ported (Globus 3.0 alpha release) functionality and performance tests in progress

CMS Michele Gulmini, CHEP2003, San Diego USA, March Summary and Plans RCMS architecture defined Prototypes developed aiming: –Control of small DAQs to be used in Testbeam applications: Next May Testbeams (CMS Tracker and Muon DT) will provide important feedbacks on its functionality –Demonstrators aiming the validation of the architecture in terms of performance and scalability Further investigation needed mainly on the IMS Open Grid Software Architecture (OGSA) under evaluation Problem Solver development in progress: –Error detection and recovery Databases studies and evaluation foreseen