The new CMS DAQ system for LHC operation after 2014 (DAQ2) CHEP2013: Computing in High Energy Physics 2013 14-18 Oct 2013 Amsterdam Andre Holzner, University.

Slides:



Advertisements
Similar presentations
LHCb Upgrade Overview ALICE, ATLAS, CMS & LHCb joint workshop on DAQ Château de Bossey 13 March 2013 Beat Jost / Cern.
Advertisements

Remigius K Mommsen Fermilab A New Event Builder for CMS Run II A New Event Builder for CMS Run II on behalf of the CMS DAQ group.
CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/20 New Experiences with the ALICE High Level Trigger Data Transport.
Operating Systems.
PCIe based readout U. Marconi, INFN Bologna CERN, May 2013.
Router Architectures An overview of router architectures.
Trigger and online software Simon George & Reiner Hauser T/DAQ Phase 1 IDR.
CERN - IT Department CH-1211 Genève 23 Switzerland t The High Performance Archiver for the LHC Experiments Manuel Gonzalez Berges CERN, Geneva.
MSS, ALICE week, 21/9/041 A part of ALICE-DAQ for the Forward Detectors University of Athens Physics Department Annie BELOGIANNI, Paraskevi GANOTI, Filimon.
Local Area Networks (LAN) are small networks, with a short distance for the cables to run, typically a room, a floor, or a building. - LANs are limited.
S. Cittolin CERN/CMS, 22/03/07 DAQ architecture. TDR-2003 DAQ evolution and upgrades DAQ upgrades at SLHC.
Artdaq Introduction artdaq is a toolkit for creating the event building and filtering portions of a DAQ. A set of ready-to-use components along with hooks.
A TCP/IP transport layer for the DAQ of the CMS Experiment Miklos Kozlovszky for the CMS TriDAS collaboration CERN European Organization for Nuclear Research.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
+ discussion in Software WG: Monte Carlo production on the Grid + discussion in TDAQ WG: Dedicated server for online services + experts meeting (Thusday.
GBT Interface Card for a Linux Computer Carson Teale 1.
InfiniSwitch Company Confidential. 2 InfiniSwitch Agenda InfiniBand Overview Company Overview Product Strategy Q&A.
C.Combaret, L.Mirabito Lab & beamtest DAQ with XDAQ tools.
LECC2003 AmsterdamMatthias Müller A RobIn Prototype for a PCI-Bus based Atlas Readout-System B. Gorini, M. Joos, J. Petersen (CERN, Geneva) A. Kugel, R.
LAN Switching and Wireless – Chapter 1
Design and Performance of a PCI Interface with four 2 Gbit/s Serial Optical Links Stefan Haas, Markus Joos CERN Wieslaw Iwanski Henryk Niewodnicznski Institute.
CCNA 2 Week 1 Routers and WANs. Copyright © 2005 University of Bolton Welcome Back! CCNA 2 deals with routed networks You will learn how to configure.
VLVNT Amsterdam 2003 – J. Panman1 DAQ: comparison with an LHC experiment J. Panman CERN VLVNT workshop 7 Oct 2003 Use as example CMS (slides taken from.
Nov 3, 2009 RN - 1 Jet Propulsion Laboratory California Institute of Technology Current Developments for VLBI Data Acquisition Equipment at JPL Robert.
Network Architecture for the LHCb DAQ Upgrade Guoming Liu CERN, Switzerland Upgrade DAQ Miniworkshop May 27, 2013.
Virtualization for the LHCb Online system CHEP Taipei Dedicato a Zio Renato Enrico Bonaccorsi, (CERN)
Dynamic configuration of the CMS Data Acquisition Cluster Hannes Sakulin, CERN/PH On behalf of the CMS DAQ group Part 1: Configuring the CMS DAQ Cluster.
Upgrade of the CMS Event Builder Andrea Petrucci - CERN (PH/CMD) on behalf of the CMS DAQ group 19 th International Conference on Computing in High Energy.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Overview of DAQ at CERN experiments E.Radicioni, INFN MICE Daq and Controls Workshop.
Latest ideas in DAQ development for LHC B. Gorini - CERN 1.
LHCb front-end electronics and its interface to the DAQ.
2003 Conference for Computing in High Energy and Nuclear Physics La Jolla, California Giovanna Lehmann - CERN EP/ATD The DataFlow of the ATLAS Trigger.
Advanced Computer Networks Lecturer: E EE Eng. Ahmed Hemaid Office: I 114.
LNL 1 SADIRC2000 Resoconto 2000 e Richieste LNL per il 2001 L. Berti 30% M. Biasotto 100% M. Gulmini 50% G. Maron 50% N. Toniolo 30% Le percentuali sono.
Niko Neufeld, CERN/PH. Online data filtering and processing (quasi-) realtime data reduction for high-rate detectors High bandwidth networking for data.
Management of the LHCb Online Network Based on SCADA System Guoming Liu * †, Niko Neufeld † * University of Ferrara, Italy † CERN, Geneva, Switzerland.
The CMS Event Builder Demonstrator based on MyrinetFrans Meijers. CHEP 2000, Padova Italy, Feb The CMS Event Builder Demonstrator based on Myrinet.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
1 Farm Issues L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
Niko Neufeld HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Pierre VANDE VYVRE ALICE Online upgrade October 03, 2012 Offline Meeting, CERN.
CODA Graham Heyes Computer Center Director Data Acquisition Support group leader.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
ROM. ROM functionalities. ROM boards has to provide data format conversion. – Event fragments, from the FE electronics, enter the ROM as serial data stream;
Remigius K Mommsen Fermilab CMS Run 2 Event Building.
DAQ 1000 Tonko Ljubicic, Mike LeVine, Bob Scheetz, John Hammond, Danny Padrazo, Fred Bieser, Jeff Landgraf.
The Evaluation Tool for the LHCb Event Builder Network Upgrade Guoming Liu, Niko Neufeld CERN, Switzerland 18 th Real-Time Conference June 13, 2012.
EPS 2007 Alexander Oh, CERN 1 The DAQ and Run Control of CMS EPS 2007, Manchester Alexander Oh, CERN, PH-CMD On behalf of the CMS-CMD Group.
CHEP 2010, October 2010, Taipei, Taiwan 1 18 th International Conference on Computing in High Energy and Nuclear Physics This research project has.
29/05/09A. Salamon – TDAQ WG - CERN1 LKr calorimeter L0 trigger V. Bonaiuto, L. Cesaroni, A. Fucci, A. Salamon, G. Salina, F. Sargeni.
Eric Hazen1 Ethernet Readout With: E. Kearns, J. Raaf, S.X. Wu, others... Eric Hazen Boston University.
HTCC coffee march /03/2017 Sébastien VALAT – CERN.
MPD Data Acquisition System: Architecture and Solutions
Use of FPGA for dataflow Filippo Costa ALICE O2 CERN
LHCb and InfiniBand on FPGA
Challenges in ALICE and LHCb in LHC Run3
Niko Neufeld LHCb Upgrade Online Computing Challenges CERN openlab Workshop on Data Center Technologies and Infrastructures, Mar 2017.
Electronics Trigger and DAQ CERN meeting summary.
Enrico Gamberini, Giovanna Lehmann Miotto, Roland Sipos
TELL1 A common data acquisition board for LHCb
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
CMS DAQ Event Builder Based on Gigabit Ethernet
DAQ upgrades at SLHC S. Cittolin CERN/CMS, 22/03/07
Trigger, DAQ, & Online: Perspectives on Electronics
Event Building With Smart NICs
LHCb Trigger, Online and related Electronics
SVT detector electronics
TELL1 A common data acquisition board for LHCb
Presentation transcript:

The new CMS DAQ system for LHC operation after 2014 (DAQ2) CHEP2013: Computing in High Energy Physics Oct 2013 Amsterdam Andre Holzner, University California at San Diego On behalf of the CMS collaboration 1

Overview  DAQ2 Motivation  Requirements  Layout / Data path  Frontend Readout Link  Event builder core  Performance considerations  Infiniband  File based filter farm and storage  DAQ2 test setup and results  Summary/Outlook 2

DAQ2 motivation  Aging equipment:  Run1 DAQ uses some technologies which are disappearing  PCI-X cards, Myrinet  Almost all equipment reached the end of the 5 year lifecycle  CMS detector upgrades  Some subsystems move to new front-end drivers  Some subsystems will add more channels  LHC performance  Expect higher instantaneous luminosity after LS1 → higher number of interactions per bunch crossing (‘pileup’) → larger event size, higher data rate  Physics  Higher centre-of-mass energy and more pileup imply:  either raise trigger thresholds, or  more intelligent decisions at Higher Level Trigger → requires more CPU power 3

DAQ2 requirements 4 RequirementDAQ1DAQ2 Readout rate100 kHz Front end drivers (FEDs)640: 1 ~ 2 kByte ~50: 2 ~ 8 kByte Total readout bandwidth 100 GByte/s200 GByte/s Interface to FEDs 1) SLink64Slink64/Slink Express Coupling Event builder software/HLT software 2) no requirementdecoupled Lossless event building HLT capacityextendable High availabilty/ fault tolerance 3) Cloud facility for offline processing 4) originally not required Subdetector local runs See talks of 1) P. Žejdl 2) R. Mommsen 3) H.Sakulin 4) J.A.Coarasa

DAQ2 data path 5 FED ~640 (legacy) + 50 ( μ TCA) Front End Drivers FEROL ~576 Front End Readout Optical Links (FEROLs) 10 GBit/s Ethernet → 40 GBit/s Ethernet switches, 8/12/16 →1 concentration RU 72 Readout Unit PCs (superfragment assembly) Infiniband switch (full 72 x 48 connectivity, 2.7 TBit/s) IB 56 GBit/s BU 48 Builder units (full event assembly) Eth 40 GBit/s Slink64/ SlinkExpress IB 56 GBit/s Eth 40 GBit/s Eth 10 GBit/s Ethernet switches 40 GBit/s → 10 GBit/s (→ 1 GBit/s), 1 →M distribution FU Filter units (~ 13’000 cores) storage custom hardware commercial hardware

DAQ2 layout 6 underground surface

FrontEnd Readout Link (FEROL) 7 Slink64 from FEDs see P. Žejdl’s talk for more details Slink Express from μ TCA FEDs 10 GBit/s Ethernet  Replace Myrinet card (upper half) by a new custom card  PCI-X interface to legacy slink receiver card (lower half)  10 GBit/s Ethernet output to central event builder  Restricted TCP/IP protocol engine inside the FPGA  Additional optical links (inputs) for future μ TCA based Front End Drivers (6-10 GBit/s; custom, simple point to point protocol)  Allows to use industry standard 10 GBit/s transceivers, cables and switches/routers  Only commercially available hardware further downstream

Event Builder Core  Two stage event building:  72 Readout units (RU) aggregate 8-16 fragments (4 kByte average) into superfragments  Larger buffers compared to FEROLs  48 Builder Units (BU) build the entire event from superfragments  InifiniBand (or 40 Gbit/s Ethernet) as interconnect  Works in a 15 x 15 system, need to scale to 72 x 48  Fault tolerance:  FEROLs can be routed to different RU (adding a second switching layer improves flexibility)  Builder Units can be excluded from running 8

 Number of DAQ2 elements is an order of magnitude smaller than for DAQ1  Consequently, bandwidth per PC is an order of magnitude higher  CPU frequency did not increase since DAQ1 but number of cores did  Need to pay attention to performance tuning  TCP socket buffers  Interrupt affinities  Non-uniform memory access Performance considerations DAQ1DAQ2 # readout units (RU)64048 RU max. bandwidth3 Gbit/s40 Gbit/s # builder units (BU)> BU max. bandwidth2 Gbit/s56 Gbit/s 9 CPU0 CPU1 QPI PCIe Memory Bus

Infiniband  Advantages:  Designed as a High Performance Computing interconnect over short distances (within datacenters)  Protocol is implemented in the network card silicon → low CPU load  56 GBit/s per link (copper or optical)  Native support for Remote Direct Memory Access (RDMA)  No copying of bulk data between user space and kernel (‘true zero- copy’)  affordable  Disadvantages:  Less widely known, API significantly differs from BSD sockets for TCP/IP  Fewer vendors than Ethernet  Niche market 10 Top500.org share by Interconnect family Infiniband DAQ1 TDR (2002) Myrinet 1 Gbit/s Ethernet 10 Gbit/s Ethernet 2013

 In DAQ1, high level trigger process was running inside a DAQ application → introduces dependencies between online (DAQ) and offline (event selection) software which have different release cycles, compilers, state machines etc.  Decoupling these needs a common, simple interface  files (no special common code required to write and read them)  Builder unit stores events in files in a RAM disk  Builder Unit acts as a NFS server, exports event files to Filter Unit PCs  Baseline: 2 Gbyte/s bandwidth  ‘Local’ within a rack  Filter units write out selected events (~ 1 in 100) back to a global (CMS DAQ wide) filesystem (e.g. Lustre) for transfer to Tier0 computing centre File based filter farm and Storage 11 see R. Mommsen’s talk for more details BU FU

DAQ2 test setup 12 Copper R720 Copper C6220 1U 10/40 Gbit/s Ethernet switch C Gbit Copper 1-10 GBit/s router x2 4 4 x16 BU RG45 FEROL emulator FED Builder RU RU Builder RU/BU Emulators FU R720 1U 40 Gbit/s Ethernet switch 40 Gbit Copper 1U 40 Gbit/s Ethernet switch 40 Gbit Copper R310 A C6220 B 1U Infiniband FDR switch x2 x13 x8 C6220 B R720 C6100 D B 10 Gbit fiber 40 Gbit Cupper 10 Gbit Fiber B RG45 R310 A B C6220 B x8 10 Gbit Fiber C x3 Copper FEROL/RU/BU/FU Emulators x3 x2 x3 x13 x8 X8 10 Gbit Fiber x8 x3 40 Gbit Copper x2 x8 FRL/FEROL x16 x8 x2 10 Gbit Fiber 10 Gbit Fiber x2 C’ 2.67GHz B/C/C’Dual 2.60GHz DDual 2.67GHz

BU RU InfiniBand Measurements 13 FED FEROL RU BU FU 15 RU 15 BU RU BU working range

FEROL Test setup results 14 FEROL RU BU FU 12 FEROLs 1 RU 4 BU working range FEROL BU

Test setup: DAQ1 vs. DAQ2 15 Comparison of throughput per Readout Unit

Summary / Outlook  CMS has designed a central data acquisition system for post-LS1 data taking  replacing outdated standards by modern technology  ~ twice the event building capacity than DAQ system for Run1  accomodating a large dynamic range of up to 8 kByte fragments, flexible configuration  Increase in networking bandwidth was faster than increase in event sizes  Number of event builder PCs reduced by a factor ~10  Each PC handles a factor ~10 more bandwidth  Requires performance related fine-tuning  Performed various performance tests with a small scale demonstrator  First installation activities for DAQ2 have started already  Full deployment foreseen for mid 2014  Looking forward to recording physics data after the Long Shutdown 1 ! 16