Niko Neufeld, CERN. Trigger-free read-out – every bunch-crossing! 40 MHz of events to be acquired, built and processed in software 40 Tbit/s aggregated.

Slides:



Advertisements
Similar presentations
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Advertisements

CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/20 New Experiences with the ALICE High Level Trigger Data Transport.
Niko Neufeld PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.
PCIe based readout U. Marconi, INFN Bologna CERN, May 2013.
Niko Neufeld CERN PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
Niko Neufeld, CERN/PH-Department
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
+ discussion in Software WG: Monte Carlo production on the Grid + discussion in TDAQ WG: Dedicated server for online services + experts meeting (Thusday.
HEAnet Centralised NAS Storage Justin Hourigan, Senior Network Engineer, HEAnet Limited.
Status and plans for online installation LHCb Installation Review April, 12 th 2005 Niko Neufeld for the LHCb Online team.
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Data Acquisition for the 12 GeV Upgrade CODA 3. The good news…  There is a group dedicated to development and support of data acquisition at Jefferson.
The ALICE DAQ: Current Status and Future Challenges P. VANDE VYVRE CERN-EP/AID.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
CERN Real Time conference, Montreal May 18 – 23, 2003 Richard Jacobsson 1 Driving the LHCb Front-End Readout TFC Team: Arek Chlopik, IPJ, Poland Zbigniew.
6-10 Oct 2002GREX 2002, Pisa D. Verkindt, LAPP 1 Virgo Data Acquisition D. Verkindt, LAPP DAQ Purpose DAQ Architecture Data Acquisition examples Connection.
Fermi National Accelerator Laboratory SC2006 Fermilab Data Movement & Storage Multi-Petabyte tertiary automated tape store for world- wide HEP and other.
Network Architecture for the LHCb DAQ Upgrade Guoming Liu CERN, Switzerland Upgrade DAQ Miniworkshop May 27, 2013.
DAQ and Trigger upgrade U. Marconi, INFN Bologna Firenze, March
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Niko Neufeld PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015.
LHCb Upgrade Architecture Review BE DAQ Interface Rainer Schwemmer.
Latest ideas in DAQ development for LHC B. Gorini - CERN 1.
Monte Carlo Data Production and Analysis at Bologna LHCb Bologna.
VLVnT09A. Belias1 The on-shore DAQ system for a deep-sea neutrino telescope A.Belias NOA-NESTOR.
LHCb DAQ system LHCb SFC review Nov. 26 th 2004 Niko Neufeld, CERN.
Why it might be interesting to look at ARM Ben Couturier, Vijay Kartik Niko Neufeld, PH-LBC SFT Technical Group Meeting 08/10/2012.
Niko Neufeld, CERN/PH. ALICE – “A Large Ion Collider Experiment” Size: 26 m long, 16 m wide, 16m high; weight: t 35 countries, 118 Institutes Material.
Niko Neufeld, CERN/PH. Online data filtering and processing (quasi-) realtime data reduction for high-rate detectors High bandwidth networking for data.
Management of the LHCb Online Network Based on SCADA System Guoming Liu * †, Niko Neufeld † * University of Ferrara, Italy † CERN, Geneva, Switzerland.
Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015.
The Past... DDL in ALICE DAQ The DDL project ( )  Collaboration of CERN, Wigner RCP, and Cerntech Ltd.  The major Hungarian engineering contribution.
Future experiment specific needs for LHCb OpenFabrics/Infiniband Workshop at CERN Monday June 26 Sai Suman Cherukuwada Sai Suman Cherukuwada and Niko Neufeld.
ICHEC Presentation ESR2: Reconfigurable Computing and FPGAs ICE-DIP Srikanth Sridharan 9/2/2015.
DAQ interface + implications for the electronics Niko Neufeld LHCb Electronics Upgrade June 10 th, 2010.
Common meeting of CERN DAQ teams CERN May 3 rd 2006 Niko Neufeld PH/LBC for the LHCb Online team.
1 Farm Issues L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
Predrag Buncic ALICE Status Report LHCC Referee Meeting CERN
Niko Neufeld HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Pierre VANDE VYVRE ALICE Online upgrade October 03, 2012 Offline Meeting, CERN.
Niko Neufeld LHCC Detector Upgrade Review, June 3 rd 2014.
1 Event Building L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
CODA Graham Heyes Computer Center Director Data Acquisition Support group leader.
11/01/20081 Data simulator status CCRC’08 Preparatory Meeting Radu Stoica, CERN* 11 th January 2007 * On leave from IFIN-HH.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Introduction to DAQ Architecture Niko Neufeld CERN / IPHE Lausanne.
An Overview over Online Systems at the LHC Invited Talk at NSS-MIC 2012 Anaheim CA, 31 October 2012 Beat Jost, Cern.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
The Evaluation Tool for the LHCb Event Builder Network Upgrade Guoming Liu, Niko Neufeld CERN, Switzerland 18 th Real-Time Conference June 13, 2012.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Storage for Run 3 Rainer Schwemmer, LHCb Computing Workshop 2015.
High throughput computing collaboration (HTCC) Jon Machen, Network Software Specialist DCG IPAG, EU Exascale Labs INTEL Switzerland.
HTCC coffee march /03/2017 Sébastien VALAT – CERN.
Giovanna Lehmann Miotto CERN EP/DT-DI On behalf of the DAQ team
LHCb and InfiniBand on FPGA
Niko Neufeld  (quasi) real-time connectivity requirements ”CERN openlab workshop.
WP18, High-speed data recording Krzysztof Wrona, European XFEL
Challenges in ALICE and LHCb in LHC Run3
Niko Neufeld LHCb Upgrade Online Computing Challenges CERN openlab Workshop on Data Center Technologies and Infrastructures, Mar 2017.
Enrico Gamberini, Giovanna Lehmann Miotto, Roland Sipos
Controlling a large CPU farm using industrial tools
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
Bernd Panzer-Steindel, CERN/IT
ALICE Computing Model in Run3
ALICE Computing Upgrade Predrag Buncic
VELO readout On detector electronics Off detector electronics to DAQ
Characteristics of Reconfigurable Hardware
Presentation transcript:

Niko Neufeld, CERN

Trigger-free read-out – every bunch-crossing! 40 MHz of events to be acquired, built and processed in software 40 Tbit/s aggregated throughput, about 500 data-sources with 100 Gigabit/s each More than optical fibres from the detector At least 2000 servers openlab technical WS 6/11/15 - Niko Neufeld 2

Detector front-end electronics Eventbuilder network Eventbuilder PCs/PCIe40 Eventfilter Farm ~ 80 subfarms Eventfilter Farm ~ 80 subfarms UX85B Point 8 surface subfarm switch TFC x 100 Gbit/s subfarm switch Online storage Clock & fast commands ~ 9000 Versatile Links for DAQ ~ 9000 Versatile Links for DAQ throttle from PCIe40 Clock & fast commands 6 x 100 Gbit/s ECS openlab technical WS 6/11/15 - Niko Neufeld m

Arria10 FPGA PCIe Gen3 x 16 == 100 Gbit/s Up to 48 optical input links Will have > 500 in experiment Used also by ALICE, and … maybe … who knows… 4 openlab technical WS 6/11/15 - Niko Neufeld

6.4 PB net storage / 12.8 raw openlab technical WS 6/11/15 - Niko Neufeld 5

optical fibres (on 40 – 80 MPO12) 10 2 U I/O servers with 2 x 100 Gbit/s interface 36 compute servers taking between 20 to 40 Gbit/s (each) 1 – 2 PB of storage ~ 40Tbit/s network I/O to network (full duplex) openlab technical WS 6/11/15 - Niko Neufeld 7

Vendor neutral Public tender every time Long lived facility > 10 years Has to grow “adiabatically” – unlike a super-computer we can’t throw away things after 4 years Upgradeable Cost, Cost, Cost Tight cost-efficient integration of compute, storage and network Should be flexible to also accommodate accelerators (Xeon/Phi, FPGA) if they prove efffective Power  electricity at CERN is cheap, but we want to be green and reduce running costs openlab technical WS 6/11/15 - Niko Neufeld 8

Need temporary storage to wait for calibration and alignment – and to profit from no-beam time Current model: completely local storage as a software RAID1 of 4 TB on each node File management by scripts and control software No common name-space 100% overhead Capacity oriented Streaming I/O only, single reader / single writer, typically max 4 streams / RAID set, aggregated I/O low 10 – 20 MB/s openlab technical WS 6/11/15 - Niko Neufeld 10

Operational: No common name-space Disk-failure during data-taking can cause several problems Controller or both disks failed  node needs to be excluded from data-taking Disk does not actually fail but becomes “slow” because of errors  node accumulates backlog of unprocessed data Rebuild can affect performance Inaccessible data (even temporary) block all data from further processing (because offline data-sets are treated as a “whole”) openlab technical WS 6/11/15 - Niko Neufeld 11

Basically disk and I/O requirements / node go up by 10x Need cost-efficient solution Still looks attractive to have disks in each node vs NAS / rack, disaggregated shelves  see challenge 1 Can we have better efficiency RAID5,6,7? Would love to have common name-space, posix or not? openlab technical WS 6/11/15 - Niko Neufeld 12