LHCb Upgrade Architecture Review BE DAQ Interface Rainer Schwemmer.

Slides:



Advertisements
Similar presentations
Copyright© 2000 OPNET Technologies, Inc. R.W. Dobinson, S. Haas, K. Korcyl, M.J. LeVine, J. Lokier, B. Martin, C. Meirosu, F. Saka, K. Vella Testing and.
Advertisements

Remigius K Mommsen Fermilab A New Event Builder for CMS Run II A New Event Builder for CMS Run II on behalf of the CMS DAQ group.
LHCb readout infrastructure NA62 TDAQ WG Meeting April 1 st, 2009 Niko Neufeld, PH/LBC.
PCIe based readout U. Marconi, INFN Bologna CERN, May 2013.
Connecting LANs, Backbone Networks, and Virtual LANs
The LHCb Online System Design, Implementation, Performance, Plans Presentation at the 2 nd TIPP Conference Chicago, 9 June 2011 Beat Jost Cern.
Niko Neufeld, CERN/PH Many stimulating, fun discussions with my event-building friends in ALICE, ATLAS, CMS and LHCb, and with a lot of smart people in.
Architecture and Dataflow Overview LHCb Data-Flow Review September 2001 Beat Jost Cern / EP.
A TCP/IP transport layer for the DAQ of the CMS Experiment Miklos Kozlovszky for the CMS TriDAS collaboration CERN European Organization for Nuclear Research.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
Input/Output 2 What is I/O? How we get the CPU to communicate with devices From the computer’s point of view, it’s just 1’s and 0’s Gets interpreted.
Network Architecture for the LHCb DAQ Upgrade Guoming Liu CERN, Switzerland Upgrade DAQ Miniworkshop May 27, 2013.
DAQ and Trigger upgrade U. Marconi, INFN Bologna Firenze, March
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Federico Alessio Zbigniew Guzik Richard Jacobsson TFC Team: A Super-TFC for a Super-LHCb - Top-down approach -
Latest ideas in DAQ development for LHC B. Gorini - CERN 1.
McGraw-Hill©The McGraw-Hill Companies, Inc., 2004 Connecting Devices CORPORATE INSTITUTE OF SCIENCE & TECHNOLOGY, BHOPAL Department of Electronics and.
LHCb front-end electronics and its interface to the DAQ.
LHCb DAQ system LHCb SFC review Nov. 26 th 2004 Niko Neufeld, CERN.
Guido Haefeli CHIPP Workshop on Detector R&D Geneva, June 2008 R&D at LPHE/EPFL: SiPM and DAQ electronics.
Why it might be interesting to look at ARM Ben Couturier, Vijay Kartik Niko Neufeld, PH-LBC SFT Technical Group Meeting 08/10/2012.
Niko Neufeld, CERN/PH. ALICE – “A Large Ion Collider Experiment” Size: 26 m long, 16 m wide, 16m high; weight: t 35 countries, 118 Institutes Material.
Interconnect Networks Basics. Generic parallel/distributed system architecture On-chip interconnects (manycore processor) Off-chip interconnects (clusters.
Niko Neufeld, CERN/PH. Online data filtering and processing (quasi-) realtime data reduction for high-rate detectors High bandwidth networking for data.
Management of the LHCb Online Network Based on SCADA System Guoming Liu * †, Niko Neufeld † * University of Ferrara, Italy † CERN, Geneva, Switzerland.
Links from experiments to DAQ systems Jorgen Christiansen PH-ESE 1.
A Super-TFC for a Super-LHCb (II) 1. S-TFC on xTCA – Mapping TFC on Marseille hardware 2. ECS+TFC relay in FE Interface 3. Protocol and commands for FE/BE.
Niko Neufeld Electronics Upgrade Meeting, April 10 th 2014.
Future experiment specific needs for LHCb OpenFabrics/Infiniband Workshop at CERN Monday June 26 Sai Suman Cherukuwada Sai Suman Cherukuwada and Niko Neufeld.
1 Level 1 Pre Processor and Interface L1PPI Guido Haefeli L1 Review 14. June 2002.
DAQ interface + implications for the electronics Niko Neufeld LHCb Electronics Upgrade June 10 th, 2010.
1 Farm Issues L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
Niko Neufeld HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Pierre VANDE VYVRE ALICE Online upgrade October 03, 2012 Offline Meeting, CERN.
Niko Neufeld LHCC Detector Upgrade Review, June 3 rd 2014.
DAQ Overview + selected Topics Beat Jost Cern EP.
Update on Optical Fibre issues Rainer Schwemmer. Key figures Minimum required bandwidth: 32 Tbit/s –# of 100 Gigabit/s links > 320, # of 40 Gigabit/s.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Introduction to DAQ Architecture Niko Neufeld CERN / IPHE Lausanne.
LECC2004 BostonMatthias Müller The final design of the ATLAS Trigger/DAQ Readout-Buffer Input (ROBIN) Device B. Gorini, M. Joos, J. Petersen, S. Stancu,
ROM. ROM functionalities. ROM boards has to provide data format conversion. – Event fragments, from the FE electronics, enter the ROM as serial data stream;
The Evaluation Tool for the LHCb Event Builder Network Upgrade Guoming Liu, Niko Neufeld CERN, Switzerland 18 th Real-Time Conference June 13, 2012.
Connectors, Repeaters, Hubs, Bridges, Switches, Routers, NIC’s
DHH at DESY Test Beam 2016 Igor Konorov TUM Physics Department E18 19-th DEPFET workshop May Kloster Seeon Overview: DHH system overview DHE/DHC.
Grzegorz Kasprowicz1 Level 1 trigger sorter implemented in hardware.
The ALICE Data-Acquisition Read-out Receiver Card C. Soós et al. (for the ALICE collaboration) LECC September 2004, Boston.
Eric Hazen1 Ethernet Readout With: E. Kearns, J. Raaf, S.X. Wu, others... Eric Hazen Boston University.
WPFL General Meeting, , Nikhef A. Belias1 Shore DAQ system - report on studies A.Belias NOA-NESTOR.
HTCC coffee march /03/2017 Sébastien VALAT – CERN.
Giovanna Lehmann Miotto CERN EP/DT-DI On behalf of the DAQ team
LHCb and InfiniBand on FPGA
Modeling event building architecture for the triggerless data acquisition system for PANDA experiment at the HESR facility at FAIR/GSI Krzysztof Korcyl.
Challenges in ALICE and LHCb in LHC Run3
Niko Neufeld LHCb Upgrade Online Computing Challenges CERN openlab Workshop on Data Center Technologies and Infrastructures, Mar 2017.
Electronics Trigger and DAQ CERN meeting summary.
Enrico Gamberini, Giovanna Lehmann Miotto, Roland Sipos
TELL1 A common data acquisition board for LHCb
Electronics, Trigger and DAQ for SuperB
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
The Train Builder Data Acquisition System for the European-XFEL
VELO readout On detector electronics Off detector electronics to DAQ
John Harvey CERN EP/LBC July 24, 2001
Event Building With Smart NICs
LHCb Trigger, Online and related Electronics
Network Processors for a 1 MHz Trigger-DAQ System
Throttling: Infrastructure, Dead Time, Monitoring
The LHCb Front-end Electronics System Status and Future Development
TELL1 A common data acquisition board for LHCb
Connectors, Repeaters, Hubs, Bridges, Switches, Routers, NIC’s
Presentation transcript:

LHCb Upgrade Architecture Review BE DAQ Interface Rainer Schwemmer

Sketch of the upgrade DAQ

GBT: custom radiation- hard link over MMF, 3.2 – 4.8 Gbit/s (~ ) Input into DAQ network (10/40 Gigabit Ethernet or FDR IB) (1000 to 4000) Output from DAQ network into compute unit clusters (100 Gbit Ethernet / EDR IB) (200 to 400 links) LHCb Upgrade DAQ 3 Detector DAQ network 100 m rock Readout Units Compute Units LHCb FE-review DAQ BE, Rainer Schwemmer

Key figures Minimum required bandwidth: > 32 Tbit/s –# of 100 Gigabit/s links > 320, # of 40 Gigabit/s links > 800, # of 10 Gigabit/s links > 3200 # of compute units: many (depends on CPU power, trigger algorithms, GPU or not, etc…)  design should scale “easily” between An event (“snapshot of a collision”) is about 100 kB of data  for the DAQ only the bandwidth counts # of events processed every second: 10 to 40 millions # of events retained after filtering: to per second (data reduction of at least a factor 1000) LHCb FE-review DAQ BE, Rainer Schwemmer 4

The network options Two local area network technologies are investigated: Ethernet and InfiniBand Speed-grades –Ethernet 10 Gbit/s, 40 Gbit/s 100 Gbit/s (only for up-links and interconnects, not from the BE-board) –InfiniBand QDR 32 Gbit/s FDR 54 Gbit/ s (in the future: EDR ~ 100 Gbit/s) LHCb FE-review DAQ BE, Rainer Schwemmer 5

How to minimise cost in the DAQ 1.Minimise distances between network devices (switches) and end-nodes (TELL40s, compute-units)  location of BE (later) 2.Try to use high-bandwidth LAN devices with shallow / no buffering (like InfiniBand or “cut-through” Ethernet) 3.Minimise the use of optics - use copper links (industry standard, largest market volume) –copper = SFP+ twinax cable or 10 GBaseT (twisted pair) LHCb FE-review DAQ BE, Rainer Schwemmer 6

Network costs (based on recent offers 2012) Cost of switch ports excluding optics based on 2012 offers (not list- prices) “Edge” ports are ports with shallow buffers (cut-through) and without high-speed uplinks, typically found in switches build around chip-sets (Broadcom) by OEMs “Core” ports are ports with deep buffers (store and forward) and/or including interconnects and chassis overhead costs LHCb FE-review DAQ BE, Rainer Schwemmer 7 Price per port (USD) 10 GbE40 GbEIB FDR (52 Gb) edge core

Why worry about BE location? Using 10 Gig links from UX to SX requires a lot of links (cost in ports and optics!) Faster links (40 G / 100 G) do not have the required range –(40G-SR4 and 100G-SR10 are specified over 100 m (OM3) 125 m (OM4) Want to use long-range GBT  bring the BE to SX Even if the BE cannot use copper (which is already costly) at least between core-network and compute units (farm) we want to maximize the use of copper  network and farm have to be co- located The closer everything is the more we can use data-centre technology (volume!) LHCb FE-review DAQ BE, Rainer Schwemmer 8

GBT: custom radiation- hard link over MMF, 4.8 Gbit/s (about 10000) Input into DAQ network (10/40 Gigabit Ethernet or FDR IB) (1000 to 4000) Output from DAQ network into compute unit clusters (100 Gbit Ethernet / EDR IB) (200 to 400 links) Link technology – location of back-end 9 Detector DAQ network 100 m rock Readout Units Compute Units Long distance covered by low-speed links from detector to Readout Units. Cheap and optical links required anyhow ~400 m cable equivalent length LHCb FE-review DAQ BE, Rainer Schwemmer

BE  DAQ Implementation 10

Farm Example 1: FPGA DAQ Core (Ethernet / Push) 11 TFC Info / Destination Assignment Memory D_IN from preprocessing Event Scheduler MEP Assembly Ethernet Core(s) Packetizer (UDP/IP) Data TFC Inter- face TFC Inter- face DAQ Core Mem Writer Evt Descriptors Not needed for pure push

Buffering Why do we need it? Simplest case: –Push, no resend -> current scheme –Problems: See current network issues Current DAQ edge switches: 1.5GB memory / 48 Ports = MB/Port “Theoretically” necessary: 3 * Event size * MEP factor = 2 MB Still drops packets due to full buffers –In the future: Switches with large buffers tend to be very expensive (mostly telco-market) Need at least Buffer for scheduled sending + re-send –Needs to buffer Bandwidth x 1 RTT + several MEPs + contingency O(10 MBs per link) Buffering also needed for advanced flow-control protocols in Ethernet (QCN, etc…) Safest traffic-shaping method: pull protocol –Data is buffered on the BE board, the sinks collect it when they are ready for processing –Full control over output buffer depth (depends on protocol) –O(50 MBs per link) 12

Farm Example 2: FPGA DAQ Core (Ethernet / Push – with scheduling and acknowledgment) 13 TFC Info / Destination Assignment Memory D_IN from preprocessing Event Scheduler MEP Assembly Ethernet Core(s) Packetizer (UDP/IP) Acknowledgments DataAck TFC Inter- face TFC Inter- face DAQ Core Mem Writer Evt Descriptors

Farm Example 3: FPGA DAQ Core (Pull) 14 TFC Inter- face TFC Inter- face TFC Info Event Scheduler MEP Assembly Infiniband Core(s) Packetizer (RDMA/IB) Destination Assignment DataRDMA Req/Ack RDMA Engine Memory D_IN from preprocessing Mem Writer Evt Descriptors DAQ Core

FPGA Interface Core interface consists of 3 components –Streaming Data Interface for receiving the data from pre-processing (Avalon-ST) –Memory I/O Interface: Write once, Read once + ε (Avalon MM) –TFC Input data (Avalon ??) Additional Interface for ECS –To be specified 15

Debugging Essential: –Counters and status registers of all sub algorithms on the board –Operational experience shows that it is often useless to have counters if not synchronized => snapshot trigger + latched please –If your algorithm is blocking the data train, we need to know! Necessary for post mortem analysis –Need history of input data If we want to know why something is throttling –Core dump like functionality: –Upon configurable trigger, the DAQ is blocked and all current data in I/O buffers of sub algorithms + N available previous events are sent to the DAQ core –Data is sent as high priority packet to HLT which can not be rejected –Obviously disabled during production phases 16

Debugging 17 Main Data Path FIFO 1 FIFO 2 FIFO N Core Dump Decision Logic To DAQ Core ECS Data Path Status Shadow Memory To ECS Interface To DAQ Core Throttle/Buffer Full Info Status Latch Logic Algo N Algo 2 Algo 1 Use Ring Buffers for History analysis

Summary DAQ will use exclusively COTS technology (Ethernet or InfiniBand) Minimising distances between BE, network and farm allows minimising of cost Buffering on the BE allows using cheaper network equipment DAQ FPGA core should hide the interconnect technology and protocol from the rest of the firmware 18

Backup slides 19

Example of IB based read-out TELL40 using IB FPGA core (Bologna, CERN, …)  –2 FDR IB ports (52 Gb/s) –need 16 serializers for about 100 Gbit/s output (can absorb 30 versatile links) LHCb FE-review DAQ BE, Rainer Schwemmer 20