DAQ Overview + selected Topics Beat Jost Cern EP.

Slides:



Advertisements
Similar presentations
LHCb Upgrade Overview ALICE, ATLAS, CMS & LHCb joint workshop on DAQ Château de Bossey 13 March 2013 Beat Jost / Cern.
Advertisements

A Gigabit Ethernet Link Source Card Robert E. Blair, John W. Dawson, Gary Drake, David J. Francis*, William N. Haberichter, James L. Schlereth Argonne.
The LHCb Event-Builder Markus Frank, Jean-Christophe Garnier, Clara Gaspar, Richard Jacobson, Beat Jost, Guoming Liu, Niko Neufeld, CERN/PH 17 th Real-Time.
Remigius K Mommsen Fermilab A New Event Builder for CMS Run II A New Event Builder for CMS Run II on behalf of the CMS DAQ group.
The LHCb DAQ and Trigger Systems: recent updates Ricardo Graciani XXXIV International Meeting on Fundamental Physics.
CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/20 New Experiences with the ALICE High Level Trigger Data Transport.
UNIVERSITY OF BERGEN DEPARTMENT OF PHYSICS 1 UiB DR 2003 High Level API for the TPC-FEE control and configuration.
The LHCb Online System Design, Implementation, Performance, Plans Presentation at the 2 nd TIPP Conference Chicago, 9 June 2011 Beat Jost Cern.
5 Feb 2002Alternative Ideas for the CALICE Backend System 1 Alternative Ideas for the CALICE Back-End System Matthew Warren and Gordon Crone University.
LHCb DAQ Review, September LHCb Timing and Fast Control System TFC Team: Arek Chlopik, Warsaw Zbigniew Guzik, Warsaw Richard Jacobsson, CERN Beat.
Architecture and Dataflow Overview LHCb Data-Flow Review September 2001 Beat Jost Cern / EP.
A TCP/IP transport layer for the DAQ of the CMS Experiment Miklos Kozlovszky for the CMS TriDAS collaboration CERN European Organization for Nuclear Research.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
TFC Partitioning Support and Status Beat Jost Cern EP.
LECC2003 AmsterdamMatthias Müller A RobIn Prototype for a PCI-Bus based Atlas Readout-System B. Gorini, M. Joos, J. Petersen (CERN, Geneva) A. Kugel, R.
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
CERN Real Time conference, Montreal May 18 – 23, 2003 Richard Jacobsson 1 Driving the LHCb Front-End Readout TFC Team: Arek Chlopik, IPJ, Poland Zbigniew.
LHCb Trigger and Data Acquisition System Beat Jost Cern / EP Presentation given at the 11th IEEE NPSS Real Time Conference June 14-18, 1999 Santa Fe, NM.
Network Architecture for the LHCb DAQ Upgrade Guoming Liu CERN, Switzerland Upgrade DAQ Miniworkshop May 27, 2013.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
1 Network Performance Optimisation and Load Balancing Wulf Thannhaeuser.
Overview of DAQ at CERN experiments E.Radicioni, INFN MICE Daq and Controls Workshop.
KIP Ivan Kisel JINR-GSI meeting Nov 2003 High-Rate Level-1 Trigger Design Proposal for the CBM Experiment Ivan Kisel for Kirchhoff Institute of.
Latest ideas in DAQ development for LHC B. Gorini - CERN 1.
1 DAQ System Realization DAQ Data Flow Review Sep th, 2001 Niko Neufeld CERN, EP.
LHCb front-end electronics and its interface to the DAQ.
2003 Conference for Computing in High Energy and Nuclear Physics La Jolla, California Giovanna Lehmann - CERN EP/ATD The DataFlow of the ATLAS Trigger.
LHCb DAQ system LHCb SFC review Nov. 26 th 2004 Niko Neufeld, CERN.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
Modeling PANDA TDAQ system Jacek Otwinowski Krzysztof Korcyl Radoslaw Trebacz Jagiellonian University - Krakow.
Niko Neufeld, CERN/PH. Online data filtering and processing (quasi-) realtime data reduction for high-rate detectors High bandwidth networking for data.
1 Network Processor based RU Implementation, Applicability, Summary Readout Unit Review 24 July 2001 Beat Jost, Niko Neufeld Cern / EP.
Online View and Planning LHCb Trigger Panel Beat Jost Cern / EP.
The CMS Event Builder Demonstrator based on MyrinetFrans Meijers. CHEP 2000, Padova Italy, Feb The CMS Event Builder Demonstrator based on Myrinet.
Future experiment specific needs for LHCb OpenFabrics/Infiniband Workshop at CERN Monday June 26 Sai Suman Cherukuwada Sai Suman Cherukuwada and Niko Neufeld.
Guirao - Frascati 2002Read-out of high-speed S-LINK data via a buffered PCI card 1 Read-out of High Speed S-LINK Data Via a Buffered PCI Card A. Guirao.
DAQ interface + implications for the electronics Niko Neufeld LHCb Electronics Upgrade June 10 th, 2010.
1 Farm Issues L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
Niko Neufeld HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
1 Event Building L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
CODA Graham Heyes Computer Center Director Data Acquisition Support group leader.
KIP Ivan Kisel, Uni-Heidelberg, RT May 2003 A Scalable 1 MHz Trigger Farm Prototype with Event-Coherent DMA Input V. Lindenstruth, D. Atanasov,
Monitoring for the ALICE O 2 Project 11 February 2016.
Introduction to DAQ Architecture Niko Neufeld CERN / IPHE Lausanne.
An Overview over Online Systems at the LHC Invited Talk at NSS-MIC 2012 Anaheim CA, 31 October 2012 Beat Jost, Cern.
DAQ Summary Beat Jost Cern EP. Beat Jost, Cern 2 Agenda (Reminder)
The Evaluation Tool for the LHCb Event Builder Network Upgrade Guoming Liu, Niko Neufeld CERN, Switzerland 18 th Real-Time Conference June 13, 2012.
EPS 2007 Alexander Oh, CERN 1 The DAQ and Run Control of CMS EPS 2007, Manchester Alexander Oh, CERN, PH-CMD On behalf of the CMS-CMD Group.
Giovanna Lehmann Miotto CERN EP/DT-DI On behalf of the DAQ team
Modeling event building architecture for the triggerless data acquisition system for PANDA experiment at the HESR facility at FAIR/GSI Krzysztof Korcyl.
PANDA collaboration meeting FEE session
Electronics Trigger and DAQ CERN meeting summary.
Richard Jacobsson, CERN
Enrico Gamberini, Giovanna Lehmann Miotto, Roland Sipos
TELL1 A common data acquisition board for LHCb
Controlling a large CPU farm using industrial tools
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
The LHCb Event Building Strategy
LHCb Trigger and Data Acquisition System Requirements and Concepts
John Harvey CERN EP/LBC July 24, 2001
Event Building With Smart NICs
Philippe Vannerem CERN / EP ICALEPCS - Oct03
LHCb Trigger, Online and related Electronics
The LHCb High Level Trigger Software Framework
Network Processors for a 1 MHz Trigger-DAQ System
Throttling: Infrastructure, Dead Time, Monitoring
LHCb Online Meeting November 15th, 2000
Use Of GAUDI framework in Online Environment
TELL1 A common data acquisition board for LHCb
Presentation transcript:

DAQ Overview + selected Topics Beat Jost Cern EP

Beat Jost, Cern 2 FE/Controls/DAQ Workshop 2000 Overall Architecture Read-out Network (RN) RU Control & Monitoring RU 2-4 GB/s 4 GB/s 20 MB/s Variable latency L2 ~10 ms L3 ~200 ms LAN Read-out units (RU) Timing & Fast Control Front-End Electronics VDET TRACK ECAL HCAL MUON RICH LHC-B Detector L0 L1 Level 0 Trigger Level 1 Trigger 40 MHz 1 MHz 40 kHz Fixed latency 4.0  s Variable latency <1 ms Data rates 40 TB/s 1 TB/s Front-End Multiplexers (FEM) 1 MHz Front End Links Trigger Level 2 & 3 Event Filter SFC CPU Sub-Farm Controllers (SFC) Storage Throttle

Beat Jost, Cern 3 FE/Controls/DAQ Workshop 2000 Event Builder Architecture

Beat Jost, Cern 4 FE/Controls/DAQ Workshop 2000 Event Building Protocol qPure “push” protocol ãEach source (RU) of the event-builder sends data autonomously to a destination as soon as they are available ãThe destination is assigned through a static algorithm based on the event number. The identical algorithm runs in all sources (belonging to the same partition) -> The RUs belong to only one partition ãThe destinations (SFCs) receive the the fragments of the RUs and assemble them to form complete events. The SFCs will be configured to treat only fragments from the sources of the same partition -> The SFCs belong to only one partition ãNo communication (other than event fragments) between RUs and SFCs

Beat Jost, Cern 5 FE/Controls/DAQ Workshop 2000 Event Fragment Assembly qThe current idea is that the SFCs will be equipped with intelligent NICs (Network Interface Cards) ãThe NICs will handle all the fragment assembly to form complete events. ãThe host computer will only receive fully assembled events (or events flagged as erroneous) -> load on host reduced by ~100 (number of SFCs). SFC CPU MEM Bridge Local Bus PCI Bus SNIC Readout Network Controls Network Subfarm Network NIC

Beat Jost, Cern 6 FE/Controls/DAQ Workshop 2000 SFC Functionality qAssemble complete events qMonitor the state of the CPUs connected to it, implement dynamic load balancing among them and send raw data to ‘free’ CPUs qReceive the accepted and reconstructed events from the CPUs and send them to the storage controller qMonitor the buffer occupancy in the sub-farm and issue L1-Throttle in case buffers start getting full qProvide monitoring information to ECS

Beat Jost, Cern 7 FE/Controls/DAQ Workshop 2000 Data Flow

Beat Jost, Cern 8 FE/Controls/DAQ Workshop 2000 Readout Network Technologies Limit speed of end-node connections to ~1 Gb/s Current Candidate Technologies åATM –‘no’ smart NICs –enough performance –too expensive åMyrinet –smart NICs exist –works in principle –cheap switches, ~expensive interfaces –No buffering in switches –have to cascade switches -> have to build FIFOs to put between levels of switches (scalability) åGb Ethernet –under study –has smart NICs –prob. works –cheap –Prob. have to cascade switches, should improve with 10GbE

Beat Jost, Cern 9 FE/Controls/DAQ Workshop 2000 Other Networks qData Network ãCheapest possible compatible with performance requirements åInput rate into CPU: ~5 MB/s åOutput rate from CPU: ~25 kB/s ÜFast Ethernet should be sufficient qControls Network ãCheapest possible (prob. Fast Ethernet)

Beat Jost, Cern 10 FE/Controls/DAQ Workshop 2000 Open Issues qNon Zero-suppressed readout ãMemory on smart NICs is limited (currently 1 MB, including code) ãPossible solution could be to do fragment assembly in the SFC itself? (Not too nice…)