Challenges in ALICE and LHCb in LHC Run3

Slides:



Advertisements
Similar presentations
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Advertisements

Remigius K Mommsen Fermilab A New Event Builder for CMS Run II A New Event Builder for CMS Run II on behalf of the CMS DAQ group.
CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/20 New Experiences with the ALICE High Level Trigger Data Transport.
5 th LHCb Computing Workshop, May 19 th 2015 Niko Neufeld, CERN/PH-Department
Niko Neufeld PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.
PCIe based readout U. Marconi, INFN Bologna CERN, May 2013.
Niko Neufeld CERN PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.
Chapter 6 High-Speed LANs Chapter 6 High-Speed LANs.
Online Systems Status Review of requirements System configuration Current acquisitions Next steps... Upgrade Meeting 4-Sep-1997 Stu Fuess.
Data Acquisition Backbone Core DABC J. Adamczewski, H.G. Essel, N. Kurz, S. Linev GSI, Darmstadt The new Facility for Antiproton and Ion Research at GSI.
Niko Neufeld, CERN/PH-Department
Architecture and Dataflow Overview LHCb Data-Flow Review September 2001 Beat Jost Cern / EP.
A TCP/IP transport layer for the DAQ of the CMS Experiment Miklos Kozlovszky for the CMS TriDAS collaboration CERN European Organization for Nuclear Research.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
GBT Interface Card for a Linux Computer Carson Teale 1.
Status and plans for online installation LHCb Installation Review April, 12 th 2005 Niko Neufeld for the LHCb Online team.
The ALICE DAQ: Current Status and Future Challenges P. VANDE VYVRE CERN-EP/AID.
Network Architecture for the LHCb DAQ Upgrade Guoming Liu CERN, Switzerland Upgrade DAQ Miniworkshop May 27, 2013.
The new CMS DAQ system for LHC operation after 2014 (DAQ2) CHEP2013: Computing in High Energy Physics Oct 2013 Amsterdam Andre Holzner, University.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Niko Neufeld PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.
Latest ideas in DAQ development for LHC B. Gorini - CERN 1.
LHCb front-end electronics and its interface to the DAQ.
LHCb DAQ system LHCb SFC review Nov. 26 th 2004 Niko Neufeld, CERN.
Why it might be interesting to look at ARM Ben Couturier, Vijay Kartik Niko Neufeld, PH-LBC SFT Technical Group Meeting 08/10/2012.
Niko Neufeld, CERN/PH. ALICE – “A Large Ion Collider Experiment” Size: 26 m long, 16 m wide, 16m high; weight: t 35 countries, 118 Institutes Material.
Niko Neufeld, CERN/PH. Online data filtering and processing (quasi-) realtime data reduction for high-rate detectors High bandwidth networking for data.
Management of the LHCb Online Network Based on SCADA System Guoming Liu * †, Niko Neufeld † * University of Ferrara, Italy † CERN, Geneva, Switzerland.
XLV INTERNATIONAL WINTER MEETING ON NUCLEAR PHYSICS Tiago Pérez II Physikalisches Institut For the PANDA collaboration FPGA Compute node for the PANDA.
Future experiment specific needs for LHCb OpenFabrics/Infiniband Workshop at CERN Monday June 26 Sai Suman Cherukuwada Sai Suman Cherukuwada and Niko Neufeld.
ICHEC Presentation ESR2: Reconfigurable Computing and FPGAs ICE-DIP Srikanth Sridharan 9/2/2015.
DAQ interface + implications for the electronics Niko Neufeld LHCb Electronics Upgrade June 10 th, 2010.
1 Farm Issues L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
Niko Neufeld HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Niko Neufeld, CERN. Trigger-free read-out – every bunch-crossing! 40 MHz of events to be acquired, built and processed in software 40 Tbit/s aggregated.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Pierre VANDE VYVRE ALICE Online upgrade October 03, 2012 Offline Meeting, CERN.
Niko Neufeld LHCC Detector Upgrade Review, June 3 rd 2014.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Introduction to DAQ Architecture Niko Neufeld CERN / IPHE Lausanne.
ROM. ROM functionalities. ROM boards has to provide data format conversion. – Event fragments, from the FE electronics, enter the ROM as serial data stream;
COMPASS DAQ Upgrade I.Konorov, A.Mann, S.Paul TU Munich M.Finger, V.Jary, T.Liska Technical University Prague April PANDA DAQ/FEE WS Игорь.
The Evaluation Tool for the LHCb Event Builder Network Upgrade Guoming Liu, Niko Neufeld CERN, Switzerland 18 th Real-Time Conference June 13, 2012.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
IRFU The ANTARES Data Acquisition System S. Anvar, F. Druillole, H. Le Provost, F. Louis, B. Vallage (CEA) ACTAR Workshop, 2008 June 10.
HTCC coffee march /03/2017 Sébastien VALAT – CERN.
Giovanna Lehmann Miotto CERN EP/DT-DI On behalf of the DAQ team
Balazs Voneki CERN/EP/LHCb Online group
MPD Data Acquisition System: Architecture and Solutions
M. Bellato INFN Padova and U. Marconi INFN Bologna
Use of FPGA for dataflow Filippo Costa ALICE O2 CERN
LHCb and InfiniBand on FPGA
Niko Neufeld  (quasi) real-time connectivity requirements ”CERN openlab workshop.
Niko Neufeld LHCb Upgrade Online Computing Challenges CERN openlab Workshop on Data Center Technologies and Infrastructures, Mar 2017.
LHC experiments Requirements and Concepts ALICE
Enrico Gamberini, Giovanna Lehmann Miotto, Roland Sipos
ETD/Online Report D. Breton, U. Marconi, S. Luitz
Controlling a large CPU farm using industrial tools
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
Bernd Panzer-Steindel, CERN/IT
CMS DAQ Event Builder Based on Gigabit Ethernet
DAQ upgrades at SLHC S. Cittolin CERN/CMS, 22/03/07
Introduction to Networks
The LHCb Event Building Strategy
Event Building With Smart NICs
Characteristics of Reconfigurable Hardware
LHCb Trigger, Online and related Electronics
Network Processors for a 1 MHz Trigger-DAQ System
Cluster Computers.
Presentation transcript:

Challenges in ALICE and LHCb in LHC Run3 Niko Neufeld (including material kindly provided by Pierre vande Vyvre) CERN/EP

The Large Hadron Collider High Throughput Computing Project Status Report 11/16/16 - Niko Neufeld

LHC long-term planning ALICE & LHCb upgrade High Throughput Computing Project Status Report 11/16/16 - Niko Neufeld

IT infrastructure for LHCb experiment after 2019 Driven by upgrade of the experiment data acquisition and trigger to read all(!) of the 40 Million bunch-crossings in the LHC into a computer farm Requires 2 MW of computing power Will consist of two separate, connected clusters with different characteristics: predominantly I/O and predominantly CPU I/O-cluster: Requires a lot of external I/O (50 Tbit/s) which will come on about ~ 10000 optical fibres: current plan to enter as MPO12 CPU-cluster: Requires an estimated 4000 dual-socket xeon severs of the 2020 type (or equivalent CPU power)

LHCb Data Acquisition & Trigger 2018 Detector front-end electronics UX85B Clock & fast commands 10000 Versatile Link Clock & fast commands I/O cluster 500 Eventbuilder PCs (software LLT) TFC throttle from PCIe40 100 Gbit/s Eventbuilder network 100 Gbit/s Versatile Link: proprietary 4.8 Gbit/s link connecting the detector in the radiation, strong B-field area to the commodity hardware of the I/O cluster Eventbuilder PC: element of the I/O cluster Sub-farm: logical unit in the CPU-farm (flexible size) TFC: timing and fast control: synchronous, hard-real-time distribution mechanism which is interfaced to the I/O cluster The 6x100 Gbit/s in the picture are just indicative of a certain network topology used for modelling – any other topology providing the at least the required 40 Tbit/s is possible. Point 8 surface CPU cluster subfarm switch subfarm switch Online storage Eventfilter Farm ~ 4000 dual socket nodes

Network facts Total bandwidth about 40 Tbit/s full duplex Target is about 500 nodes with 100 Gbit/s interfaces: Each node is sending and receiving round-robin to and from all other nodes (full duplex!) Candidate technologies: IB EDR, Intel OPA, 100 G Ethernet “Event-builder” PCs in I/O cluster have two interfaces one for event-building, one to feed completely assembled data to filter units (workers) Can be, but need not be, the same technologies

Anatomy of a server in the I/O cluster PCIe40: custom-made 16-lane PCIe Gen3 card (FPGA based using an Altera Arria 10) – puts about 100 Gigabit/s into server memory Server: needs to be able to sustain at least 400 Gigabit/s I/O per PCIe40 card (baseline: one card per server)

Requirements Lots of high-speed (>= 100 Gb/s) connections  short distances for economical cabling At least 1000 100 Gbit/s ports I/O cluster needs to house custom-made PCIe cards Consists of lower-density servers (GPGPU type) Need lot of I/O ports I/O cluster needs only moderate room for growth Full system deployed from day 1 (2019) I/O precludes high density (1U / node) CPU cluster will be upgraded and extended regularly Can be a mixture of CPU and accelerators Dense deployment (as dense as possible) Server in the CPU-cluster will be CPU-bound Dense server packing possible (0.5 U / server) Up to 2600 U needed

Workload aspects I/O cluster: CPU cluster: needs efficient I/O (low CPU overhead) Ideally want to use remaining CPU cycle parasitically (similar to CPU cluster) but need to watch out for memory bandwidth and I/O constraints CPU cluster: current work-loads mostly sequential work on better threading on-going (complex) use of accelerators (GPGPUs, XeonPhis, FPGAs) under study

Facility aspects New DC to be constructed (2 MW) Require about 4000 U (or equivalent) Power available (not battery backed) Location fixed (arrival of fibres from underground) Cooling available (water @ 25 C) or Outside air average high throughout the year < 26 C Average low throughout the year > -3 C Annual average T = 10 C

Long-distance data-transport option The CPU cluster can be hosted elsewhere. In that case the 32 Tbit/s raw data need to be transported off-site Distance: about 10 km Amount of SMF available dictates the use of DWDM Using 25 Ghz wavelength about 1200 lambdas are needed in the maximum configuration. The solution should be compact, does not need redundancy (non-critical data), it should be scalable (starting smaller to be ramped up to max later) Traffic is essentially uni-directional  should exploit this to reduce cost of optics (I think I know how to do this at layer 3 (in IP)) Would prefer a Layer 1 option (protocol agnostic), but are willing to consider Layer 2 or Layer 3, if lower cost

Tentative planning (subject to change) Decision on data-centre until Q2 2017 Decision on network technology until Q1 2019, procurement during 2019 Decision on I/O server during until Q1 2019, procurement during 2019 CPU server procurement Q4 2020, followed by procurement in early 2021

Acceleration with FPGA in ALICE

High efficiency LHC up-time integrated over the year is "only" about 30% Classical re-construction step (which requires re-reading the full data to create an intermediate format used then for analysis) is being phased out in favour of "streaming" DAQ which provides data directly at the source in the "best" possible state, aligned and calibrated Tight integration of off- and online facilities needed.  Idle cycles used for analysis, simulation, etc...

Flexible infrastructure Ideally could seamlessly run batch-type (simulation, analysis) and (near) real-time workloads. Require easy access to disk-storage High speed network between (some) servers Housing of custom I/O cards (PCIe) Flexible amount of accelerators (FPGA, GPGPU, Xeon/Phi), shoudl also be flexibly be assigned to different work-loads  and all this ideally in the same infrastructure, easily reconfigurable rack-level and data-centre oriented desgin might be the way to go?

Backup

ALICE O2 Technical Design Report https://cds.cern.ch/record/2011297/files/ALICE-TDR-019.pdf