Management of the LHCb Online Network Based on SCADA System Guoming Liu * †, Niko Neufeld † * University of Ferrara, Italy † CERN, Geneva, Switzerland.

Slides:



Advertisements
Similar presentations
The Detector Control System – FERO related issues
Advertisements

The Control System for the ATLAS Pixel Detector
Peter Chochula CERN-ALICE ALICE DCS Workshop, CERN September 16, 2002 DCS – Frontend Monitoring and Control.
Experiment Control Systems at the LHC An Overview of the System Architecture An Overview of the System Architecture JCOP Framework Overview JCOP Framework.
André Augustinus ALICE Detector Control System  ALICE DCS is responsible for safe, stable and efficient operation of the experiment  Central monitoring.
CHEP 2012 – New York City 1.  LHC Delivers bunch crossing at 40MHz  LHCb reduces the rate with a two level trigger system: ◦ First Level (L0) – Hardware.
The LHCb Event-Builder Markus Frank, Jean-Christophe Garnier, Clara Gaspar, Richard Jacobson, Beat Jost, Guoming Liu, Niko Neufeld, CERN/PH 17 th Real-Time.
Clara Gaspar, May 2010 The LHCb Run Control System An Integrated and Homogeneous Control System.
LHCb readout infrastructure NA62 TDAQ WG Meeting April 1 st, 2009 Niko Neufeld, PH/LBC.
L. Granado Cardoso, F. Varela, N. Neufeld, C. Gaspar, C. Haen, CERN, Geneva, Switzerland D. Galli, INFN, Bologna, Italy ICALEPCS, October 2011.
Control and monitoring of on-line trigger algorithms using a SCADA system Eric van Herwijnen Wednesday 15 th February 2006.
The LHCb Online System Design, Implementation, Performance, Plans Presentation at the 2 nd TIPP Conference Chicago, 9 June 2011 Beat Jost Cern.
SNMP ( Simple Network Management Protocol ) based Network Management.
Calo Piquet Training Session - Xvc1 ECS Overview Piquet Training Session Cuvée 2012 Xavier Vilasis.
Using PVSS for the control of the LHCb TELL1 detector emulator (OPG) P. Petrova, M. Laverne, M. Muecke, G. Haefeli, J. Christiansen CERN European Organization.
Architecture and Dataflow Overview LHCb Data-Flow Review September 2001 Beat Jost Cern / EP.
Summary DCS Workshop - L.Jirdén1 Summary of DCS Workshop 28/29 May 01 u Aim of workshop u Program u Summary of presentations u Conclusion.
Readout & Controls Update DAQ: Baseline Architecture DCS: Architecture (first round) August 23, 2001 Klaus Honscheid, OSU.
1 ALICE Control System ready for LHC operation ICALEPCS 16 Oct 2007 L.Jirdén On behalf of the ALICE Controls Team CERN Geneva.
JCOP Workshop September 8th 1999 H.J.Burckhart 1 ATLAS DCS Organization of Detector and Controls Architecture Connection to DAQ Front-end System Practical.
TRIGGER-LESS AND RECONFIGURABLE DATA ACQUISITION SYSTEM FOR POSITRON EMISSION TOMOGRAPHY Grzegorz Korcyl 2013.
Clara Gaspar, October 2011 The LHCb Experiment Control System: On the path to full automation.
Cisco S2 C4 Router Components. Configure a Router You can configure a router from –from the console terminal (a computer connected to the router –through.
Status and plans for online installation LHCb Installation Review April, 12 th 2005 Niko Neufeld for the LHCb Online team.
XXVI Workshop on Recent Developments in High Energy Physics and Cosmology Theodoros Argyropoulos NTUA DCS group Ancient Olympia 2008 ATLAS Cathode Strip.
The Joint COntrols Project Framework Manuel Gonzalez Berges on behalf of the JCOP FW Team.
CERN Real Time conference, Montreal May 18 – 23, 2003 Richard Jacobsson 1 Driving the LHCb Front-End Readout TFC Team: Arek Chlopik, IPJ, Poland Zbigniew.
André Augustinus 10 October 2005 ALICE Detector Control Status Report A. Augustinus, P. Chochula, G. De Cataldo, L. Jirdén, S. Popescu the DCS team, ALICE.
© 2008 Cisco Systems, Inc. All rights reserved.Cisco ConfidentialPresentation_ID 1 Chapter 1: Introduction to Scaling Networks Scaling Networks.
Network Architecture for the LHCb DAQ Upgrade Guoming Liu CERN, Switzerland Upgrade DAQ Miniworkshop May 27, 2013.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Data Acquisition Backbone Core J. Adamczewski-Musch, N. Kurz, S. Linev GSI, Experiment Electronics, Data processing group.
Online Software 8-July-98 Commissioning Working Group DØ Workshop S. Fuess Objective: Define for you, the customers of the Online system, the products.
Overview of DAQ at CERN experiments E.Radicioni, INFN MICE Daq and Controls Workshop.
LHCb DAQ system LHCb SFC review Nov. 26 th 2004 Niko Neufeld, CERN.
Clara Gaspar, July 2005 RTTC Control System Status and Plans.
Niko Neufeld, CERN/PH. Online data filtering and processing (quasi-) realtime data reduction for high-rate detectors High bandwidth networking for data.
A Super-TFC for a Super-LHCb (II) 1. S-TFC on xTCA – Mapping TFC on Marseille hardware 2. ECS+TFC relay in FE Interface 3. Protocol and commands for FE/BE.
DAQ interface + implications for the electronics Niko Neufeld LHCb Electronics Upgrade June 10 th, 2010.
14 November 08ELACCO meeting1 Alice Detector Control System EST Fellow : Lionel Wallet, CERN Supervisor : Andre Augustinus, CERN Marie Curie Early Stage.
Clara Gaspar, April 2006 LHCb Experiment Control System Scope, Status & Worries.
LHCb Configuration Database Lana Abadie, PhD student (CERN & University of Pierre et Marie Curie (Paris VI), LIP6.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
DAQ & ConfDB Configuration DB workshop CERN September 21 st, 2005 Artur Barczyk & Niko Neufeld.
Introduction to DAQ Architecture Niko Neufeld CERN / IPHE Lausanne.
Maria del Carmen Barandela Pazos CERN CHEP 2-7 Sep 2007 Victoria LHCb Online Interface to the Conditions Database.
M. Caprini IFIN-HH Bucharest DAQ Control and Monitoring - A Software Component Model.
20OCT2009Calo Piquet Training Session - Xvc1 ECS Overview Piquet Training Session Cuvée 2009 Xavier Vilasis.
The Evaluation Tool for the LHCb Event Builder Network Upgrade Guoming Liu, Niko Neufeld CERN, Switzerland 18 th Real-Time Conference June 13, 2012.
Stefan Koestner IEEE NSS – San Diego 2006 IEEE - Nuclear Science Symposium San Diego, Oct. 31 st 2006 Stefan Koestner on behalf of the LHCb Online Group.
Grzegorz Kasprowicz1 Level 1 trigger sorter implemented in hardware.
Giovanna Lehmann Miotto CERN EP/DT-DI On behalf of the DAQ team
Electronics Trigger and DAQ CERN meeting summary.
LHC experiments Requirements and Concepts ALICE
Enrico Gamberini, Giovanna Lehmann Miotto, Roland Sipos
CMS – The Detector Control System
TELL1 A common data acquisition board for LHCb
Controlling a large CPU farm using industrial tools
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
The LHCb Event Building Strategy
The LHCb Run Control System
Philippe Vannerem CERN / EP ICALEPCS - Oct03
SNMP (Simple Network Management Protocol) based Network Management
IEEE - Nuclear Science Symposium San Diego, Oct. 31st 2006
The LHCb High Level Trigger Software Framework
Network Processors for a 1 MHz Trigger-DAQ System
Throttling: Infrastructure, Dead Time, Monitoring
Tools for the Automation of large distributed control systems
Configuration DB Status report Lana Abadie
Presentation transcript:

Management of the LHCb Online Network Based on SCADA System Guoming Liu * †, Niko Neufeld † * University of Ferrara, Italy † CERN, Geneva, Switzerland

Outline  Introduction to LHCb Online system  LHCb online network  Network management based on SCADA system  Summary Guoming Liu 2 ICALEPCS2009

LHCb online system  LHCb is one of the large particle physics experiments on LHC at CERN  Online system is one of the infrastructures for LHCb, providing IT services for the entire experiment  Three major components:  Data Acquisition (DAQ) Transfers the event data from the detector front-end electronics to the permanent storage  Timing and Fast Control (TFC) Provides fast clock and drives all stages of the data readout of the LHCb detector between the front-end electronics and the online processing farm  Experiment Control System (ECS), Controls and monitors all parts of the experiment Guoming Liu 3 ICALEPCS2009

LHCb online system Guoming Liu 4 ICALEPCS2009 Control and Monitoring data CASTOR SWITCH HLT farm Detector TFC System SWITCH READOUT NETWORK L0 trigger LHC clock MEP Request Event building Front-End CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU Experiment Control System (ECS) VELO STOT RICH ECalHCalMuon L0 Trigger Event data Timing and Fast Control Signals SWITCH MON farm CPUCPU CPUCPU CPUCPU CPUCPU Readout Board FEE Readout Board

LHCb Online Network  Two dedicated networks:  Control network: general purpose network for experiment control system Connects all the Ethernet devices in LHCb  Data network: dedicated to data acquisition Performance critical Guoming Liu 5 ICALEPCS2009

LHCb Online Network  Two geographic parts: surface and underground Connected by two 10G links Guoming Liu 6 ICALEPCS2009

LHCb Online Network Guoming Liu 7 ICALEPCS2009 Core CTRL Routers Core DAQ Router CTRL Access Switches (~100) DAQ Access Switches (~50) On the surface

Network Monitoring System based on SCADA  Motivation  This large network needs sophisticated monitoring  Integration into LHCb ECS coherently  Provides homogeneous interfaces for non-expert shift-crew Commercial network management software?  Expensive  Integration? Guoming Liu 8 ICALEPCS2009

Network Monitoring System: Architecture  Supervisory layer  PVSS II: commercial SCADA system  JCOP: Joint Control Project for LHC experiments  Front–end Processes:  SNMP  sFlow  syslog  Data communication  DIM: Distributed Information Management Guoming Liu 9 ICALEPCS2009 SNMP / sFlow / Syslog DIM

Network Monitoring System: FSM  All behaviors are modeled as Finite State Machines (FSM)  Hierarchical structure: status/command propagated Guoming Liu 10 ICALEPCS2009  Device Units: Device Description Device Access Based on PVSS II datapoint: Alarm Handling, Archiving, Trending etc.  Control Units Abstract behavior modeling Represents the associated sub-tree

Network Monitoring System The major items under monitor  Physical topology  Discovery of the network topology based on the Link Layer Discovery Protocol (LLDP)  Discovery of the network nodes: based on the information in switches (ARP, MAC forwarding table)  Traffic  Octet / packet counters  Discard/Error counters ...  Switch status: CPU/Memory, temperature, power supply,...  Data Paths for DAQ Guoming Liu 11 ICALEPCS2009

Network Monitoring Snapshot(1): Topology Guoming Liu 12 ICALEPCS2009

Network Monitoring Snapshot(2): traffic Guoming Liu 13 ICALEPCS2009

Summary  The network management system has been implemented based on the commercial SCADA system PVSS II and the framework JCOP  It provides sophisticated monitoring of the network which are essential for our operation, i.e. switch status, traffic  It provides the homogenous operation interface and intuitive display as well  Currently only monitoring is provided, some control commands of switches to be integrated Guoming Liu 14 ICALEPCS2009

Thanks for your attention! Guoming Liu 15 ICALEPCS2009

Backup Guoming Liu 16 ICALEPCS2009

NMS Architecture: front-end processes Guoming Liu 17 ICALEPCS2009  SNMP: Simple network management protocol Used for general network monitoring, configuring  sFlow:  A sampling mechanism to capture traffic data  Based on hardware.  Two kinds of sFlow samples: flow samples and counter samples. Used on the core switch to collect traffic counters: SNMP too slow, and consumes high CPU/Memory  Syslog: event notification messages  Three distinct parts: priority, header and message.  The priority part represents both the facility and severity of the message.

Network Monitoring: hardware/system  Syslog can collect some information not covered by SNMP  Syslog server is setup to receive the syslog messages from the network devices and parse the messages. Alarm information:  Hardware: temperature, fan status, power supply status  System: CPU, memory, login authentication etc.  All the messages with the priority higher than warning, will be sent to PVSS for further processing Guoming Liu 18 ICALEPCS2009

Network Monitoring: IP routing Guoming Liu 19 ICALEPCS2009 Control and Monitoring data CASTOR SWITCH HLT farm Detector SWITCH READOUT NETWORK Event building Front-End CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU CPUCPU Readout Board VELO STOT RICH ECalHCalMuon L0 Trigger Event data Timing and Fast Control Signals SWITCH MON farm CPUCPU CPUCPU CPUCPU CPUCPU Readout Board FEE  Monitoring the status of the routing using “ping“/”arping”  Three stages for the DAQ: 1. From readout board to HLT farm 2. From HLT Farm to the LHCb online storage 3. From the online storage to CERN CASTOR