Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Computer Instrumentation Triggering and DAQ Jos Vermeulen, UvA / NIKHEF Topical.

Slides:



Advertisements
Similar presentations
G ö khan Ü nel / CHEP Interlaken ATLAS 1 Performance of the ATLAS DAQ DataFlow system Introduction/Generalities –Presentation of the ATLAS DAQ components.
Advertisements

Sander Klous on behalf of the ATLAS Collaboration Real-Time May /5/20101.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
Copyright© 2000 OPNET Technologies, Inc. R.W. Dobinson, S. Haas, K. Korcyl, M.J. LeVine, J. Lokier, B. Martin, C. Meirosu, F. Saka, K. Vella Testing and.
Kostas KORDAS INFN – Frascati XI Bruno Touschek spring school, Frascati,19 May 2006 Higgs → 2e+2  O (1/hr) Higgs → 2e+2  O (1/hr) ~25 min bias events.
Remigius K Mommsen Fermilab A New Event Builder for CMS Run II A New Event Builder for CMS Run II on behalf of the CMS DAQ group.
31 May, 2006Stefan Stancu1 Control and Data Networks Architecture Proposal S. Stancu, C. Meirosu, B. Martin.
1 The ATLAS Online High Level Trigger Framework: Experience reusing Offline Software Components in the ATLAS Trigger Werner Wiedenmann University of Wisconsin,
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
A TCP/IP transport layer for the DAQ of the CMS Experiment Miklos Kozlovszky for the CMS TriDAS collaboration CERN European Organization for Nuclear Research.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
+ discussion in Software WG: Monte Carlo production on the Grid + discussion in TDAQ WG: Dedicated server for online services + experts meeting (Thusday.
LECC2003 AmsterdamMatthias Müller A RobIn Prototype for a PCI-Bus based Atlas Readout-System B. Gorini, M. Joos, J. Petersen (CERN, Geneva) A. Kugel, R.
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
Design and Performance of a PCI Interface with four 2 Gbit/s Serial Optical Links Stefan Haas, Markus Joos CERN Wieslaw Iwanski Henryk Niewodnicznski Institute.
Data Acquisition for the 12 GeV Upgrade CODA 3. The good news…  There is a group dedicated to development and support of data acquisition at Jefferson.
The BaBar Event Building and Level-3 Trigger Farm Upgrade S.Luitz, R. Bartoldus, S. Dasu, G. Dubois-Felsmann, B. Franek, J. Hamilton, R. Jacobsen, D. Kotturi,
Network Architecture for the LHCb DAQ Upgrade Guoming Liu CERN, Switzerland Upgrade DAQ Miniworkshop May 27, 2013.
The new CMS DAQ system for LHC operation after 2014 (DAQ2) CHEP2013: Computing in High Energy Physics Oct 2013 Amsterdam Andre Holzner, University.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
The Data Flow System of the ATLAS DAQ/EF "-1" Prototype Project G. Ambrosini 3,9, E. Arik 2, H.P. Beck 1, S. Cetin 2, T. Conka 2, A. Fernandes 3, D. Francis.
The ATLAS Trigger: High-Level Trigger Commissioning and Operation During Early Data Taking Ricardo Gonçalo, Royal Holloway University of London On behalf.
Online-Offsite Connectivity Experiments Catalin Meirosu *, Richard Hughes-Jones ** * CERN and Politehnica University of Bucuresti ** University of Manchester.
Latest ideas in DAQ development for LHC B. Gorini - CERN 1.
2003 Conference for Computing in High Energy and Nuclear Physics La Jolla, California Giovanna Lehmann - CERN EP/ATD The DataFlow of the ATLAS Trigger.
LHCb DAQ system LHCb SFC review Nov. 26 th 2004 Niko Neufeld, CERN.
CHEP March 2003 Sarah Wheeler 1 Supervision of the ATLAS High Level Triggers Sarah Wheeler on behalf of the ATLAS Trigger/DAQ High Level Trigger.
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
Chapter 13 – I/O Systems (Pgs ). Devices  Two conflicting properties A. Growing uniformity in interfaces (both h/w and s/w): e.g., USB, TWAIN.
LNL 1 SADIRC2000 Resoconto 2000 e Richieste LNL per il 2001 L. Berti 30% M. Biasotto 100% M. Gulmini 50% G. Maron 50% N. Toniolo 30% Le percentuali sono.
Stefan StancuCHEP06 Mumbai, India1 Networks for ATLAS Trigger and Data Acquisition S. Stancu *†, C. Meirosu †‡, M. Ciobotaru *†, L. Leahu †‡, B. Martin.
Experience with multi-threaded C++ applications in the ATLAS DataFlow Szymon Gadomski University of Bern, Switzerland and INP Cracow, Poland on behalf.
Kostas KORDAS INFN – Frascati 10th Topical Seminar on Innovative Particle & Radiation Detectors (IPRD06) Siena, 1-5 Oct The ATLAS Data Acquisition.
1 Farm Issues L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Pierre VANDE VYVRE ALICE Online upgrade October 03, 2012 Offline Meeting, CERN.
LECC2004 BostonMatthias Müller The final design of the ATLAS Trigger/DAQ Readout-Buffer Input (ROBIN) Device B. Gorini, M. Joos, J. Petersen, S. Stancu,
Remigius K Mommsen Fermilab CMS Run 2 Event Building.
ROD Activities at Dresden Andreas Glatte, Andreas Meyer, Andy Kielburg-Jeka, Arno Straessner LAr Electronics Upgrade Meeting – LAr Week September 2009.
The Evaluation Tool for the LHCb Event Builder Network Upgrade Guoming Liu, Niko Neufeld CERN, Switzerland 18 th Real-Time Conference June 13, 2012.
Recent experience with PCI-X 2.0 and PCI-E network interfaces and emerging server systems Yang Xia Caltech US LHC Network Working Group October 23, 2006.
Jos VermeulenTopical lectures, Computer Instrumentation, Introduction, June Computer Instrumentation Introduction Jos Vermeulen, UvA / NIKHEF Topical.
The ALICE Data-Acquisition Read-out Receiver Card C. Soós et al. (for the ALICE collaboration) LECC September 2004, Boston.
HTCC coffee march /03/2017 Sébastien VALAT – CERN.
Giovanna Lehmann Miotto CERN EP/DT-DI On behalf of the DAQ team
Chapter 1: Introduction to the Personal Computer
MPD Data Acquisition System: Architecture and Solutions
M. Bellato INFN Padova and U. Marconi INFN Bologna
LHCb and InfiniBand on FPGA
5/14/2018 The ATLAS Trigger and Data Acquisition Architecture & Status Benedetto Gorini CERN - Physics Department on behalf of the ATLAS TDAQ community.
Institute of Nuclear Physics Polish Academy of Sciences
Local Area Networks, 3rd Edition David A. Stamper
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
CMS DAQ Event Builder Based on Gigabit Ethernet
DAQ upgrades at SLHC S. Cittolin CERN/CMS, 22/03/07
Operating the ATLAS Data-Flow System with the First LHC Collisions
Evolution of S-LINK to PCI interfaces
Instructor Materials Chapter 1: Introduction to the Personal Computer
The LHCb Event Building Strategy
Example of DAQ Trigger issues for the SoLID experiment
Web Server Administration
12/3/2018 The ATLAS Trigger and Data Acquisition Architecture & Status Benedetto Gorini CERN - Physics Department on behalf of the ATLAS TDAQ community.
TDAQ commissioning and status Stephen Hillier, on behalf of TDAQ
John Harvey CERN EP/LBC July 24, 2001
1/2/2019 The ATLAS Trigger and Data Acquisition Architecture & Status Benedetto Gorini CERN - Physics Department on behalf of the ATLAS TDAQ community.
Windows Virtual PC / Hyper-V
Network Processors for a 1 MHz Trigger-DAQ System
The Performance and Scalability of the back-end DAQ sub-system
TELL1 A common data acquisition board for LHCb
Chapter 13: I/O Systems.
Presentation transcript:

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Computer Instrumentation Triggering and DAQ Jos Vermeulen, UvA / NIKHEF Topical lectures, 29 June 2007

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Event data ≤ 100 kHz, 1600 fragments of ~ 1 kByte each ATLAS detector Read- Out Drivers ( RODs ) First- level trigger Read-Out Subsystems ( ROSs ) LVL2 Super- visor UX15 USA15 SDX1 CERN computer centre SDX1 USA15 UX15 Dedicated links Timing Trigger Control (TTC) 1600 Read- Out Links 10 Gigabit Ethernet ATLAS Trigger / DAQ DataFlow Overview RoI Builder DataFlow Manager Event Filter (EF) pROS ~ 870~1500 Regions Of Interest VME Data of events accepted by first-level trigger Event data requests Delete commands Requested event data stores LVL2 output dual 1, 2 or 4-core CPU nodes ~100~30 Network switches Event data pulled: partial ≤ 100 kHz, full ~ 3 kHz Event rate ~ 200 Hz Data storage Local Storage SubFarm Outputs (SFOs) LVL2 Farm + switches Network switches Event Builder SubFarm Inputs (SFIs) Second- level trigger ~40x10 ~320x1 Gbit/s ~20 switches ~150 PCs Gigabit Ethernet

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Event data ≤ 100 kHz, 1600 fragments of ~ 1 kByte each ATLAS detector Read- Out Drivers ( RODs ) First- level trigger Read-Out Subsystems ( ROSs ) UX15 USA15 Dedicated links Timing Trigger Control (TTC) 1600 Read- Out Links RoI Builder VME Data of events accepted by first-level trigger RODs, ROS PCs and ROBINs 10 Gigabit Ethernet Read-Out Drivers (ROD): subdetector-specific, collect and process data (no event selection) output via Read-Out Links (ROL, 200 MByte/s optical fibers) to buffers on ROBIN cards in Read-Out Subsystem (ROS) PCs Same type of ROLs, ROBINs and ROS PCs used for all sub-detectors ROBINs: 64-bit 66 MHz PCI cards 3 ROL inputs ROS PCs: 4U rack-mounted PCs with 4 ROBINs => 12 ROLs per ROS PC ~40x10 ~320x1 Gbit/s ~20 switches ~150 PCs

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June ROBIN : about 700 produced Read-Out Link channels (1) (200 MByte/s per channel), 64 MByte buffer memory per ROL, electrical Gigabit Ethernet (2), PowerPC processor (466 MHz) (3), 128 MByte program and data memory, Xilinx XC2V2000 FPGA (4), 66 MHz PCI-64 interface (5) 12 layer PCB, 220*106mm, surface Mounted Devices on both sides Power consumption ~ 15W (operational)

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June PCI bus: Peripheral Interconnect Bus: parallel multi-master bus, used on PC motherboards. The ROS PCs use the 66 MHz, 64 bit variety, desktop machines used in most cases the 32 MHz, 32 bit variety, but this is phased out and replaced by PCI Express, which makes use of fast serial links, up to 16 in parallel for a single slot. More information:

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June ROBIN designers (Gerard Kieft (NIKHEF), Barry Green (RHUL) and Andreas Kugel (Mannheim) ) at work at NIKHEF ROBIN

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June ROS system integrators in USA-15 (Gokhan Ünel, Markus Joos, Benedetto Gorini, Louis Tremblet)

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June ROS: PC properties  4U, 19" rack mountable PC  Motherboard: Supermicro X6DHE-XB  CPU: One 3.4 GHz Xeon  Hyper threading not used  uni-processor kernel  RAM: 512 MB  Network:  2 GB onboard 1 used for control network  4 GB on PCI-Express card 2 used + 2 free (upgrade) for LVL2 RoI collection & for Event Building  Redundant power supply  Network booted (no local hard disk)  Remote management via IPMI HP Beck - LHEP BernTDAQ weekCERN May21 – May 25,

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June DataFlow in a ROS PC ROS Application PPC Event store Data from RODs ROBIN FPGA PCI bus... PPC Event store ROBIN FPGA ROBIN request queues Multi-threaded C++ program running under Linux (SLC3 / SLC4) PowerPC processor in ROBIN runs C program booted from FLASH memory The application retrieves data fragments from the ROBINs, combines them in a single fragment and sends this to the requester Fragments have to be requested for each ROL individually Data requests from LVL2 or EB ROS request queue Request handlers 1 event per thread = DAQ thread = Linux process = Scheduler = Control thread Data to LVL2 or EB Data fragment Control message Request receiver ("trigger I/F") ROS GHz Xeon PC (SuperMicro X6DHE-XB Motherboard)

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June PCI Measurements … PCI ROBIN NIC DOLAR ROBIN DOLAR Test PCData SourceROS PC NIC Setup Test PC generates following messages: “data request”, 2 types: "LVL2" -> data requested from only few ROBIN channels "EB" -> data requested from all ROBIN channel of the ROS "event delete" (or "clear") containing a list of 100 events XOFF of link throttles data rate when ROBIN data buffer full Test data generator (FPGA based), emulates RODs, free running at max. speed Gigabit Ethernet Philosophy Let the system run as fast as possible in different ATLAS-like conditions, defined by different LVL2 accept and EB fractions Measure the sustained ”event delete" rate, this is the maximum first-level accept rate compatible with the LVL2 and EB rates following from the fractions chosen

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Running the system at NIKHEF 21.5 kHz event rate, EB fraction 100%, no LVL2 requests, 7 ROLs active

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June IPMI: Intelligent Platform Management Interface: remote control and monitoring of server via Ethernet.

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June The Event Builder Node: SFI HP Beck - LHEP BernTDAQ week  32 SFI PCs installed  Final system ~100 SFIs  1U, 19" rack mountable PC  Motherboard: Supermicro H8DSR-i  CPU: AMD Opteron GHz  SMP kernel  RAM: 2 GB  Network:  2 GB onboard 1 used for control network 1 used for data-in  1 GB on PCI-Express card used for data-out  1 dedicated IPMI port  Cold-swappable power supply  Network booted  Local hard disk to store event data; only used for commissioning  Remote management via IPMI CERN May21 – May 25,

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June  Perfect scaling observed up to 29 Event size of 1.5 MB  3.3 GB/s aggregated bandwidth  2.2 kHz EB rate  Single SFIs close to GE speed  114 MB/s per SFI  78 Hz per SFI Event Building scaling properties HP Beck - LHEP BernTDAQ weekCERN May21 – May 25,

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Compute server: Dell Poweredge * 4 cores, 1.86 GHz, 8 GByte memory, Dual Gigabit Ethernet (copper) 130 machines installed in SDX1 as “interchangeable processing unit” or “XPU” (can be used as SFI as well as EF processor)

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Interchangeable processing power Standard processor rack with up- links to both FrontEnd and BackEnd networks. The processing power migration between L2 and EF is achieved by software enabling/disabling of the appropriate up-links. From presentation Stefan Stancu, CHEP06, Mumbai, India switch for control network

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June XPUs HP Beck - LHEP BernTDAQ week  130 XPU PCs installed  Can act as L2P  Can act as EFP  1U, 19" rack mountable PC  Motherboard: Intel  CPU: 2 x Intel E5320 quad core 1.86 GHz  SMP kernel  RAM: 8 GB i.e., 1 GB per core  Network:  2 GB onboard 1 used for control and IPMI 1 used for data  VLAN to allow connecting to both DataCollection & BackEnd n/w’s  Cold-swappable power supply  Network booted  Local hard disk – recovery from crashes; pre-load events CERN May21 – May 25,

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June SFOs HP Beck - LHEP BernTDAQ weekCERN May21 – May 25,  6 SFO PCs installed  live spare  5U, 19" rack mountable PC  Motherboard: Intel  CPU: 2 x Intel E5130 dual core 2.0 GHz  SMP kernel  RAM: 4 GB  Network:  2 GB onboard 1 used for control and IPMI 1 used for data  3 GB on PCIe cards  used for data  3 SATA RAID controllers  24 SATA disks à 500GB each  Quadruple cold-swappable power supplies  Network booted

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Network architecture From presentation Stefan Stancu, CHEP06, Mumbai, India

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Two central switches One VLAN per switch (V1 and V2) Fault tolerant (tolerate one switch failure) Network architecture From presentation Stefan Stancu, CHEP06, Mumbai, India

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Ethernet is the dominant technology for LANs –TDAQ’s choice for networks multi-vendor, long term support, commodity (on-board GE adapters), etc. –Gigabit and TenGigabit Ethernet Use GE for end-nodes 10GE whenever the bandwidth requirements exceed 1Gbit/s Multi-vendor Ethernet switches/routers available on the market: –Chassis-based devices ( ~320 Gbit/s switching) GE line-cards: typically ~40 ports 10GE line-cards: typically 4 ports –Pizza-box devices (~60 Gbit/s switching) 24/48 GE ports Optional 10GE module with 2 up-links Networking From presentation Stefan Stancu, CHEP06, Mumbai, India

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June From presentation Stefan Stancu, CHEP06, Mumbai, India What happens if a switch or link fails? –Phone call, but nothing critical should happen after a single failure. Networks are made resilient by introducing redundancy: –Component-level redundancy: deployment of devices with built-in redundancy (PSU, supervision modules, switching fabric) –Network-level redundancy: deployment of additional devices/links in order to provide alternate paths between communicating nodes. Protocols are needed to correctly (and efficiently) deal with multiple paths in the network NB: Webster: characterized or marked by resilience : as a : capable of withstanding shock without permanent deformation or rupture b : tending to recover from or adjust easily to misfortune or change

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June see Cisco whitepaper, “Understanding Multiple Spanning Tree Protocol (802.1s)” Cluster Core switch Conc. sw VLAN X VLAN Y 2 x loop From presentation Stefan Stancu, TDAQ workshop, Mainz, October 2005 Problem! Multiple Spanning Tree protocol and virtual LANs (VLANs)

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Cluster Core switch Conc. sw VLAN X VLAN Y MST OK Multiple Spanning Tree protocol and virtual LANs (VLANs) From presentation Stefan Stancu, TDAQ workshop, Mainz, October 2005

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Cluster Core switch Conc. sw VLAN X VLAN Y failure OK Multiple Spanning Tree protocol and virtual LANs (VLANs) From presentation Stefan Stancu, TDAQ workshop, Mainz, October 2005

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Queue starts to form here if all senders send at the same time Switch Senders Requester Traffic shaping By limiting the number of outstanding requests the building up of queues can be avoided and the associated risk of packet loss be minimized

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June ~2000 end-nodes One core device with built-in redundancy Rack level concentration with use of link aggregation for redundant up-links to the core. BackEnd network From presentation Stefan Stancu, CHEP06, Mumbai, India

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Control network ~3000 end nodes Design assumption: the instantaneous traffic does not exceed 1 Gbit/s on any segment, including up-link. One device suffices for the core layer, but better redundancy is achieved by deploying 2 devices. A rack level concentration switch can be deployed for all units except for critical services.

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Characteristics of the software running in the system: Multi-threading: necessary for optimizing system performance, ROSs, LVL2 processors, SFIs: requests for data are generated and transmitted, while waiting for the data other processing can be done, like e.g. inputting data requested earlier. By assigning the various tasks to different threads and also allowing several threads for each task processor utilization can be optimized. Also machines with multi multi-core CPUs provide a natural environment for multi-threaded software. Thread-safety is of course a must! Communication between processes on different machines is essential for the system. TCP is preferred, but scalability is an issue, UDP in conjunction with explicitly programmed time-outs in any case will provide the necessary communication. Remote process invocation is also essential for bringing up the system Basic functionality is provided by ssh, which provides support for remote command execution, in conjunction with specification of all environment variables required

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Configuration data base

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Modelling of the dataflow system: Paper model: started as "back on the envelope" calculation, now a small C++ program, which, starting from a trigger menu and rejection factors, allows to calculate average rates in the system. Computer model: C++ program based on discrete event simulation, allowing to simulate the behavior of the system with respect to data flow aspects, again starting from a trigger menu and rejection factors. Continuous effort to keep up with new hardware and software introduced in the system and to help in understanding results of test measurements.

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Example output “computer model”, design luminosity, 100 kHz LVL1 rate LVL2 decision time (µs) (histogram selected in list) Scrolling list with histogram titles # of LVL1 triggers Monitoring of simulation Scrolling list with statistics Menu index vs. LVL2 decision time (µs) Scrolling list with histogram titles (~ 1 hr running on 1.8 GHz G5 Mac)

Jos VermeulenTopical Lectures, Computer Instrumentation, TDAQ, June Example output “computer model”, design luminosity, 100 kHz LVL1 rate Time interval between LVL1 accept and completion of event building (µs) (histogram selected in list) Scrolling list with histogram titles # of LVL1 triggers Monitoring of simulation Scrolling list with statistics Number of events buffered in em calo ROBIn Scrolling list with histogram titles