Niko Neufeld HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.

Slides:



Advertisements
Similar presentations
LHCb Upgrade Overview ALICE, ATLAS, CMS & LHCb joint workshop on DAQ Château de Bossey 13 March 2013 Beat Jost / Cern.
Advertisements

The LHCb Event-Builder Markus Frank, Jean-Christophe Garnier, Clara Gaspar, Richard Jacobson, Beat Jost, Guoming Liu, Niko Neufeld, CERN/PH 17 th Real-Time.
Remigius K Mommsen Fermilab A New Event Builder for CMS Run II A New Event Builder for CMS Run II on behalf of the CMS DAQ group.
CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/20 New Experiences with the ALICE High Level Trigger Data Transport.
LHCb readout infrastructure NA62 TDAQ WG Meeting April 1 st, 2009 Niko Neufeld, PH/LBC.
Niko Neufeld PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.
PCIe based readout U. Marconi, INFN Bologna CERN, May 2013.
Niko Neufeld CERN PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.
The LHCb Online System Design, Implementation, Performance, Plans Presentation at the 2 nd TIPP Conference Chicago, 9 June 2011 Beat Jost Cern.
Niko Neufeld, CERN/PH-Department
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
Status and plans for online installation LHCb Installation Review April, 12 th 2005 Niko Neufeld for the LHCb Online team.
CERN Real Time conference, Montreal May 18 – 23, 2003 Richard Jacobsson 1 Driving the LHCb Front-End Readout TFC Team: Arek Chlopik, IPJ, Poland Zbigniew.
Network Architecture for the LHCb DAQ Upgrade Guoming Liu CERN, Switzerland Upgrade DAQ Miniworkshop May 27, 2013.
DAQ and Trigger upgrade U. Marconi, INFN Bologna Firenze, March
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
1 Network Performance Optimisation and Load Balancing Wulf Thannhaeuser.
Niko Neufeld PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.
LHCb Upgrade Architecture Review BE DAQ Interface Rainer Schwemmer.
Latest ideas in DAQ development for LHC B. Gorini - CERN 1.
1 DAQ System Realization DAQ Data Flow Review Sep th, 2001 Niko Neufeld CERN, EP.
LHCb front-end electronics and its interface to the DAQ.
LHCb DAQ system LHCb SFC review Nov. 26 th 2004 Niko Neufeld, CERN.
Guido Haefeli CHIPP Workshop on Detector R&D Geneva, June 2008 R&D at LPHE/EPFL: SiPM and DAQ electronics.
Why it might be interesting to look at ARM Ben Couturier, Vijay Kartik Niko Neufeld, PH-LBC SFT Technical Group Meeting 08/10/2012.
Niko Neufeld, CERN/PH. ALICE – “A Large Ion Collider Experiment” Size: 26 m long, 16 m wide, 16m high; weight: t 35 countries, 118 Institutes Material.
LNL 1 SADIRC2000 Resoconto 2000 e Richieste LNL per il 2001 L. Berti 30% M. Biasotto 100% M. Gulmini 50% G. Maron 50% N. Toniolo 30% Le percentuali sono.
Niko Neufeld, CERN/PH. Online data filtering and processing (quasi-) realtime data reduction for high-rate detectors High bandwidth networking for data.
Management of the LHCb Online Network Based on SCADA System Guoming Liu * †, Niko Neufeld † * University of Ferrara, Italy † CERN, Geneva, Switzerland.
Links from experiments to DAQ systems Jorgen Christiansen PH-ESE 1.
Niko Neufeld Electronics Upgrade Meeting, April 10 th 2014.
Future experiment specific needs for LHCb OpenFabrics/Infiniband Workshop at CERN Monday June 26 Sai Suman Cherukuwada Sai Suman Cherukuwada and Niko Neufeld.
DAQ interface + implications for the electronics Niko Neufeld LHCb Electronics Upgrade June 10 th, 2010.
1 Farm Issues L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
Niko Neufeld, CERN. Trigger-free read-out – every bunch-crossing! 40 MHz of events to be acquired, built and processed in software 40 Tbit/s aggregated.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Pierre VANDE VYVRE ALICE Online upgrade October 03, 2012 Offline Meeting, CERN.
Niko Neufeld LHCC Detector Upgrade Review, June 3 rd 2014.
PCIe based readout for the LHCb upgrade U. Marconi, INFN Bologna On behalf of the LHCb Online Group (Bologna-CERN-Padova) CHEP2013, Amsterdam, 15 th October.
LHCbComputing Computing for the LHCb Upgrade. 2 LHCb Upgrade: goal and timescale m LHCb upgrade will be operational after LS2 (~2020) m Increase significantly.
PCIe40 — a Tell40 implementation on PCIexpress Beat Jost DAQ Mini Workshop 27 May 2013.
DAQ Overview + selected Topics Beat Jost Cern EP.
Update on Optical Fibre issues Rainer Schwemmer. Key figures Minimum required bandwidth: 32 Tbit/s –# of 100 Gigabit/s links > 320, # of 40 Gigabit/s.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Introduction to DAQ Architecture Niko Neufeld CERN / IPHE Lausanne.
ROM. ROM functionalities. ROM boards has to provide data format conversion. – Event fragments, from the FE electronics, enter the ROM as serial data stream;
Remigius K Mommsen Fermilab CMS Run 2 Event Building.
The Evaluation Tool for the LHCb Event Builder Network Upgrade Guoming Liu, Niko Neufeld CERN, Switzerland 18 th Real-Time Conference June 13, 2012.
Grzegorz Kasprowicz1 Level 1 trigger sorter implemented in hardware.
Event Building for the LHCb Upgrade 30 MHz Rate (32 Tb/s Aggregate Throughput) Domenico Galli INFN Bologna and Università di Bologna Workshop INFN CCR.
HTCC coffee march /03/2017 Sébastien VALAT – CERN.
Giovanna Lehmann Miotto CERN EP/DT-DI On behalf of the DAQ team
Balazs Voneki CERN/EP/LHCb Online group
M. Bellato INFN Padova and U. Marconi INFN Bologna
LHCb and InfiniBand on FPGA
Joint Genome Institute
Niko Neufeld  (quasi) real-time connectivity requirements ”CERN openlab workshop.
Challenges in ALICE and LHCb in LHC Run3
Niko Neufeld LHCb Upgrade Online Computing Challenges CERN openlab Workshop on Data Center Technologies and Infrastructures, Mar 2017.
Electronics Trigger and DAQ CERN meeting summary.
Enrico Gamberini, Giovanna Lehmann Miotto, Roland Sipos
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
CMS DAQ Event Builder Based on Gigabit Ethernet
The LHCb Event Building Strategy
VELO readout On detector electronics Off detector electronics to DAQ
Example of DAQ Trigger issues for the SoLID experiment
John Harvey CERN EP/LBC July 24, 2001
LHCb Trigger, Online and related Electronics
Network Processors for a 1 MHz Trigger-DAQ System
LHCb Online Meeting November 15th, 2000
Presentation transcript:

Niko Neufeld HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014

Substantial increase in physics reach only possible with massive increase in read-out rate Geometry (spectrometer) and comparatively small event-size make it possible – and the easiest solution – to run trigger-free, reading every bunch-crossing Note: Any increase beyond 1 MHz requires change of all front-end electronics To keep data-size reasonable, all detectors must zero-suppress at the front-end HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 2

Event rate 40 MHz of which ~ 30 MHz have protons Mean nominal event size 100 kBytes Readout board bandwidth up to 100 Gbits/s to match DAQ links of 2018 CPU nodes up to 4000 actual requirements are probably less, but provide for sufficient power, cooling and connectivity to accommodate a wide range of implementations Output rate to permanent storage 20 to 100 kHz HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 3

8800 (# VL) * 4.48 Gbit/s (wide mode)  40 Tbit/s HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 4

By Gbit/s technologies will be well established in the data-centre. Currently we see three candidates: 100 G Ethernet (data-centre links probably 2015) InfiniBand EDR (available end of 2014) Intel OmniScale Fabric (available ~ 2015) The event-builder will use 100 Gbit/s links. Add 20% safety margin for protocol overheads etc…  need Gbit/s links Start study with InfiniBand (because it’s already available) HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 5

Want to be able to use data-centre switches  need lots of buffering in event-builder units Want to decide on network technology and manufacturer as late as possible  use COTS network interfaces (i.e. PC) Trigger processing in any imaginable “compute unit” will be CPU-bound not I/O-bound  optimal match by combination of “high-speed” network (event-building) and “low-speed” network (event-filtering) Keep distances short  minimize cost of individual links HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 6

PCIe Gen4 expected Chelsio T5 (40 GbE) Mellanox FDR Mellanox 40 GbE NIC PCIe Gen3 available EDR (100 Gb/s) HCA expected 100 GbE NIC expected 7

Detector front-end electronics Eventbuilder network Eventbuilder PCs + software LLT Eventfilter Farm ~ 80 subfarms Eventfilter Farm ~ 80 subfarms UX85B Point 8 surface subfarm switch TFC x 100 Gbit/s subfarm switch Online storage Clock & fast commands 8800 Versatile Link 8800 Versatile Link throttle from PCIe40 Clock & fast commands 6 x 100 Gbit/s

200 Gbit/s full duplex in PC (including opportunistic use of idle CPU resources) 100 Gbit/s FPGA receiver card 300 m operation of Versatile Link 40 Tbit/s event-building network HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 9

Most compact system achieved by locating all Online components in a single location Power, space and cooling constraints allow such an arrangement only on the surface: containerized data- centre Versatile links connecting detector to readout-boards need to cover 300 m Test installation will start tomorrow 5/9/14 in collaboration and with the help of EN/MEF and EN/EL HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 10

Various optical fibres tested show good optical power margin and very low bit error rates For critical ECS and TFC signals Forward Error Correction ( standard option in GBT) gives additional margin On DAQ links expect < 0.25 bit errors / day / link in 24/7 operation HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 11

HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 12

Message Size[byte] lat[μs] Latency measurements for single threaded client/server Average value over 100 repetitions Low latencies are good for RDMA / pull protocols measurements by A. Falabella et al. (UNIBO & INFN) on Qlogic IB HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 13

Gbps stable on I/O Opportunistic CPU usage on event- builder nodes possible Can be used for High Level Trigger and/or Low Level Trigger measurements by D. Campora et al. (CERN) on Mellanox IB HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld

Core network will require a 500 port 100 Gbit/s device  this will be available Internally probably a Clos (like) topology  need to carefully verify blocking factors and protocol Large scale tests require large system Can test opportunistically in HPC sites HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 15

HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld LHCb Run1 & 2LHCb Run 3 Max. inst. luminosity4 x 10^322 x 10^33 Event-size (mean – zero-suppressed) [kB]~ 60 (L0 accepted) ~ 100 Event-building rate [MHz]1 40 # read-out boards~ link speed from detector [Gbit/s] output data-rate / read-out board [Gbit/s]4100 # detector-links / readout-boardup to 24up to 48 # farm-nodes~ 1000 (+ 500 in 2015) # links 100 Gbit/s (from event-builder PCs)n/a final output rate to tape [kHz]

HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 17 Business / year 2900 PB Facebook upload / year 182 PB Youtube upload / year 15 PB LHC / year 16 PB Google search index 97 PB 2900 PB © Wired Data processed by the LHCb software trigger per year from PB

The trigger-free readout of the LHCb detector requires new, zero-suppressing front-end electronics a 40 Tbit/s DAQ system This will be realized by a single, high performance, custom-designed FPGA card (PCIe40) A PC based event-builder using 100 Gbit/s technology and data centre-switches We are confident that all inherent challenges can be met at a reasonable cost HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 18

Only major discussion: performance evolution and/or number of instances per node HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 20

Cost of the Online System Cost [kCHF] Event builder (network and PCs) 3600 Optical Fibres 1700 Controls network 905 Controls system (ECS) 930 Event-filter farm 2800 Infrastructure 775 Timing and Fast Control (TFC) 500 HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld 21

22 QPI link Gbps Dual-port IB FDR – 109 Gbps GPUgen – 110 Gbps Memory throughpu t – 200 Gbps Dual- port IB FDR – 109 Gbps Memory throughpu t – 200 Gbps HL-LHC Trigger, Online and Offline Computing WG Topical Workshop Niko Neufeld