LHCb and InfiniBand on FPGA

Slides:



Advertisements
Similar presentations
Remigius K Mommsen Fermilab A New Event Builder for CMS Run II A New Event Builder for CMS Run II on behalf of the CMS DAQ group.
Advertisements

CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/20 New Experiences with the ALICE High Level Trigger Data Transport.
LHCb readout infrastructure NA62 TDAQ WG Meeting April 1 st, 2009 Niko Neufeld, PH/LBC.
1 Tell10 Guido Haefeli, Lausanne Electronics upgrade meeting 10.February 2011.
PCIe based readout U. Marconi, INFN Bologna CERN, May 2013.
Niko Neufeld CERN PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.
The LHCb Online System Design, Implementation, Performance, Plans Presentation at the 2 nd TIPP Conference Chicago, 9 June 2011 Beat Jost Cern.
Niko Neufeld, CERN/PH Many stimulating, fun discussions with my event-building friends in ALICE, ATLAS, CMS and LHCb, and with a lot of smart people in.
Niko Neufeld, CERN/PH-Department
Towards a Common Communication Infrastructure for Clusters and Grids Darius Buntinas Argonne National Laboratory.
Silicon Building Blocks for Blade Server Designs accelerate your Innovation.
A TCP/IP transport layer for the DAQ of the CMS Experiment Miklos Kozlovszky for the CMS TriDAS collaboration CERN European Organization for Nuclear Research.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
GBT Interface Card for a Linux Computer Carson Teale 1.
InfiniSwitch Company Confidential. 2 InfiniSwitch Agenda InfiniBand Overview Company Overview Product Strategy Q&A.
Network Architecture for the LHCb DAQ Upgrade Guoming Liu CERN, Switzerland Upgrade DAQ Miniworkshop May 27, 2013.
DAQ and Trigger upgrade U. Marconi, INFN Bologna Firenze, March
The new CMS DAQ system for LHC operation after 2014 (DAQ2) CHEP2013: Computing in High Energy Physics Oct 2013 Amsterdam Andre Holzner, University.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Data Acquisition Backbone Core J. Adamczewski-Musch, N. Kurz, S. Linev GSI, Experiment Electronics, Data processing group.
LHCb Upgrade Architecture Review BE DAQ Interface Rainer Schwemmer.
Latest ideas in DAQ development for LHC B. Gorini - CERN 1.
LHCb DAQ system LHCb SFC review Nov. 26 th 2004 Niko Neufeld, CERN.
Guido Haefeli CHIPP Workshop on Detector R&D Geneva, June 2008 R&D at LPHE/EPFL: SiPM and DAQ electronics.
Why it might be interesting to look at ARM Ben Couturier, Vijay Kartik Niko Neufeld, PH-LBC SFT Technical Group Meeting 08/10/2012.
Niko Neufeld, CERN/PH. ALICE – “A Large Ion Collider Experiment” Size: 26 m long, 16 m wide, 16m high; weight: t 35 countries, 118 Institutes Material.
LHCb ideas and path toward vectorization and parallelization Niko Neufeld CERN, Wednesday, Nov 28 th 2012 Fourth International Workshop for Future Challenges.
Niko Neufeld, CERN/PH. Online data filtering and processing (quasi-) realtime data reduction for high-rate detectors High bandwidth networking for data.
Niko Neufeld Electronics Upgrade Meeting, April 10 th 2014.
Future experiment specific needs for LHCb OpenFabrics/Infiniband Workshop at CERN Monday June 26 Sai Suman Cherukuwada Sai Suman Cherukuwada and Niko Neufeld.
GBT-FPGA Interface Carson Teale. GBT New radiation tolerant ASIC for bidirectional 4.8 Gb/s optical links to replace current timing, trigger, and control.
DAQ interface + implications for the electronics Niko Neufeld LHCb Electronics Upgrade June 10 th, 2010.
1 Farm Issues L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
Niko Neufeld HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Niko Neufeld, CERN. Trigger-free read-out – every bunch-crossing! 40 MHz of events to be acquired, built and processed in software 40 Tbit/s aggregated.
Prototyping electronics and data acquisition for sTGC
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Pierre VANDE VYVRE ALICE Online upgrade October 03, 2012 Offline Meeting, CERN.
PCIe based readout for the LHCb upgrade U. Marconi, INFN Bologna On behalf of the LHCb Online Group (Bologna-CERN-Padova) CHEP2013, Amsterdam, 15 th October.
Update on Optical Fibre issues Rainer Schwemmer. Key figures Minimum required bandwidth: 32 Tbit/s –# of 100 Gigabit/s links > 320, # of 40 Gigabit/s.
Introduction to DAQ Architecture Niko Neufeld CERN / IPHE Lausanne.
LECC2004 BostonMatthias Müller The final design of the ATLAS Trigger/DAQ Readout-Buffer Input (ROBIN) Device B. Gorini, M. Joos, J. Petersen, S. Stancu,
ROM. ROM functionalities. ROM boards has to provide data format conversion. – Event fragments, from the FE electronics, enter the ROM as serial data stream;
R&D on data transmission FPGA → PC using UDP over 10-Gigabit Ethernet Domenico Galli Università di Bologna and INFN, Sezione di Bologna XII SuperB Project.
Remigius K Mommsen Fermilab CMS Run 2 Event Building.
The Evaluation Tool for the LHCb Event Builder Network Upgrade Guoming Liu, Niko Neufeld CERN, Switzerland 18 th Real-Time Conference June 13, 2012.
Grzegorz Kasprowicz1 Level 1 trigger sorter implemented in hardware.
HTCC coffee march /03/2017 Sébastien VALAT – CERN.
Giovanna Lehmann Miotto CERN EP/DT-DI On behalf of the DAQ team
Balazs Voneki CERN/EP/LHCb Online group
M. Bellato INFN Padova and U. Marconi INFN Bologna
Use of FPGA for dataflow Filippo Costa ALICE O2 CERN
Niko Neufeld  (quasi) real-time connectivity requirements ”CERN openlab workshop.
Challenges in ALICE and LHCb in LHC Run3
Niko Neufeld LHCb Upgrade Online Computing Challenges CERN openlab Workshop on Data Center Technologies and Infrastructures, Mar 2017.
Enrico Gamberini, Giovanna Lehmann Miotto, Roland Sipos
TELL1 A common data acquisition board for LHCb
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
CMS DAQ Event Builder Based on Gigabit Ethernet
GBT-FPGA Interface Carson Teale.
The LHCb Event Building Strategy
Low Latency Analytics HPC Clusters
VELO readout On detector electronics Off detector electronics to DAQ
Example of DAQ Trigger issues for the SoLID experiment
Event Building With Smart NICs
LHCb Trigger, Online and related Electronics
The LHCb High Level Trigger Software Framework
Network Processors for a 1 MHz Trigger-DAQ System
TELL1 A common data acquisition board for LHCb
Connectors, Repeaters, Hubs, Bridges, Switches, Routers, NIC’s
Presentation transcript:

LHCb and InfiniBand on FPGA Niko Neufeld

Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA Motivation The LHC (large hadron collider) collides protons every 25 ns (40 MHz) Each collision produces about 100 kB of data in the (upgraded) detector Currently a pre-selection in custom electronics rejects 97.5% of these events  unfortunately a lot of them contain interesting physics In 2017 the detector will be changed so that all events can be read-out into a standard compute platform for detailed inspection Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA

Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA Key figures Minimum required bandwidth: > 32 Tbit/s # of 100 Gigabit/s links > 320 # of compute units > 1500 An event (“snapshot of a collision”) is about 100 kB of data # of events processed every second: 10 to 40 millions # of events retained after filtering: 20000 to 30000 (data reduction of at least a factor 1000) Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA

Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA Dataflow Detector GBT: custom radiation- hard link over MMF, 3.2 Gbit/s (about 12000) Input into DAQ network (10/40 Gigabit Ethernet or FDR IB) (1000 to 4000) Output from DAQ network into compute unit clusters (100 Gbit Ethernet / EDR IB) (200 to 400 links) Readout Units DAQ network 100 m rock Compute Units Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA

Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA Event-building Detector DAQ network 100 m rock Readout Units Compute Units Readout Units send to Compute Units Compute Units receive passively “Push-architecture” GBT: custom radiation- hard link over MMF, 3.2 Gbit/s (about 10000) Input into DAQ network (10/40 Gigabit Ethernet or FDR IB) (1000 to 4000) Output from DAQ network into compute unit clusters (100 Gbit Ethernet / EDR IB) (200 to 400 links) Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA

Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA Deployment System needs to be ready for detector commissioning in 2018 Acquisition and deployment of 1/3 of network + 1/6 of compute-nodes in 2017 Additional compute nodes acquisitioned and installed by end of 2018 More compute nodes as money comes in and accelerator ramps up Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA

Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA Compute Unit A compute unit is a destination for the event-data fragments from the readout units It assembles the fragments into a complete “event” and runs various selection algorithms on this event About 0.1 % of events is retained A compute unit will be a high-density server platform (mainboard with standard CPUs), probably augmented with a co-processor card (like Intel MIC or GPU) Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA

Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA Readout Unit Readout Unit needs to collect custom-links Some pre-processing Buffering Coalescing of data-fragment  reduce message-rate / transport overheads Needs an FPGA Sends data using standard network protocol (IB, Ethernet) Sending of data can be done directly from the FPGA or via a standard network silicon Works together with Compute Units to build events Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA

Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA Challenges LHCb compute unit needs to handle about 100 Gbit/s (to keep number reasonably low) Data-transport of 32 Tbit/s over 200 m Current 100 Gbit/s PHY standards either too short (SR-10) or too expensive (LR-4) 40 and 10 Gigabit standards require a lot of optical links and switch ports Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA

The case for InfiniBand Cost of switch ports excluding optics Edge ports are ports with shallow buffers (cut-through) and without high-speed uplinks Core ports are ports with deep buffers (store and forward) and/or including interconnects and chassis overhead costs 10 GbE 40 GbE IB FDR (52 Gb) edge 150 400 300 core 1500 4800 1000 Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA

Example of IB based read-out TELL40 using IB FPGA core (Bologna, CERN, …)  2 FDR IB ports (52 Gb/s) need16 serializers for about 100 Gbit/s output (can absorb 30 versatile links) PC using dual-port IB card (Mellanox CX 3) and AMCPCIe converted GBT card needs PCIe3 16x (1 U PCs with 4 slots exist) Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA

Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA InfiniBand: summary At the bandwidths we need, if it can be shown to work, it will be significantly cheaper ( the core network will be about 33% at today’s best prices) All arguments for optics / cabling apply in the same way to InfiniBand and Ethernet  FDR Infiniband can transport about 25% more data / fibre and hence needs overall less links  additional cost reduction Definitely important to establish the feasibility of this option (including on FPGA) Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA

Basline: Ethernet option Use 12 x 10 Gigabit Ethernet IP or 3 x 40 Gigabit IP Effective output bandwidth 120 Gb/s 24 x GBT Stratix V 12 x 10 Gbit/s DDR3 Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA

Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA InfiniBand IP Option 24 x GBT Stratix V 12 x 10 Gbit/s DDR3 Polybus IP = 3 x QDR InfiniBand Linklayer IP Made by Polybus Inc. Currently QDR only Effective output bandwidth 96 Gbit/s Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA

Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA ASIC option 24 x GBT Stratix V 12 x 10 Gbit/s DDR3 Stratix V 24 x GBT ConnectX3 16x 8 Gbit/s (PCIe Gen 3) 2 x FDR = 8 x 14 Gbit/s Larger board needed Needs more (slower) serializers Need PCIe bridge or x16 softcore Effective output bandwidth 104 Gbit/s DDR3 Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA

InfiniBand on custom electronics: summary Can use Mellanox ASIC or Polybus IP Both options have advantages and disadvantages from a board-design point of view In any case need a Subnet Manager Agent (SMA) and a some additional software (available in open-source)  most likely easiest to implement using a micro-controller (NIOS) Cost advantage of IB must offset additional cost and complexity in Readout Unit layer Technical WS on InfiniBand in Trigger&DAQ - LHCb and IB on FPGA