A TCP/IP transport layer for the DAQ of the CMS Experiment Miklos Kozlovszky for the CMS TriDAS collaboration CERN European Organization for Nuclear Research.

Slides:



Advertisements
Similar presentations
G ö khan Ü nel / CHEP Interlaken ATLAS 1 Performance of the ATLAS DAQ DataFlow system Introduction/Generalities –Presentation of the ATLAS DAQ components.
Advertisements

Remigius K Mommsen Fermilab A New Event Builder for CMS Run II A New Event Builder for CMS Run II on behalf of the CMS DAQ group.
Timm Morten Steinbeck, Computer Science/Computer Engineering Group Kirchhoff Institute f. Physics, Ruprecht-Karls-University Heidelberg A Framework for.
CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/20 New Experiences with the ALICE High Level Trigger Data Transport.
High Performance Communication using MPJ Express 1 Presented by Jawad Manzoor National University of Sciences and Technology, Pakistan 29 June 2015.
Embedded Transport Acceleration Intel Xeon Processor as a Packet Processing Engine Abhishek Mitra Professor: Dr. Bhuyan.
Connecting LANs, Backbone Networks, and Virtual LANs
Protocol-Dependent Message-Passing Performance on Linux Clusters Dave Turner – Xuehua Chen – Adam Oline This work is funded by the DOE MICS office.
Sven Ubik, Petr Žejdl CESNET TNC2008, Brugges, 19 May 2008 Passive monitoring of 10 Gb/s lines with PC hardware.
Architectural Software Support for Processing Clusters Johannes Gutleber, Luciano Orsini European Organization for Nuclear Research Div. EP/CMD, The CMS.
LNL CMS G. MaronCPT Week CERN, 23 April Legnaro Event Builder Prototypes Luciano Berti, Gaetano Maron Luciano Berti, Gaetano Maron INFN – Laboratori.
Data Acquisition Backbone Core DABC J. Adamczewski, H.G. Essel, N. Kurz, S. Linev GSI, Darmstadt The new Facility for Antiproton and Ion Research at GSI.
Towards a Common Communication Infrastructure for Clusters and Grids Darius Buntinas Argonne National Laboratory.
High Performance User-Level Sockets over Gigabit Ethernet Pavan Balaji Ohio State University Piyush Shivam Ohio State University.
Lecture 3 Review of Internet Protocols Transport Layer.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
The NE010 iWARP Adapter Gary Montry Senior Scientist
LECC2003 AmsterdamMatthias Müller A RobIn Prototype for a PCI-Bus based Atlas Readout-System B. Gorini, M. Joos, J. Petersen (CERN, Geneva) A. Kugel, R.
Transport Layer: TCP and UDP. Overview of TCP/IP protocols Comparing TCP and UDP TCP connection: establishment, data transfer, and termination Allocation.
Towards a Homogeneous Software Environment for DAQ Applications Luciano Orsini Johannes Gutleber CERN EP/CMD.
4 Dec 2006 Testing the machine (X7DBE-X) with 6 D-RORCs 1 Evaluation of the LDC Computing Platform for Point 2 SuperMicro X7DBE-X Andrey Shevel CERN PH-AID.
VLVNT Amsterdam 2003 – J. Panman1 DAQ: comparison with an LHC experiment J. Panman CERN VLVNT workshop 7 Oct 2003 Use as example CMS (slides taken from.
Network Architecture for the LHCb DAQ Upgrade Guoming Liu CERN, Switzerland Upgrade DAQ Miniworkshop May 27, 2013.
D ata A cquisition B ackbone C ore DABCDABC , Huelva J.Adamczewski, H.G.Essel, N.Kurz, S.Linev 1 Work.
Dynamic configuration of the CMS Data Acquisition Cluster Hannes Sakulin, CERN/PH On behalf of the CMS DAQ group Part 1: Configuring the CMS DAQ Cluster.
Upgrade of the CMS Event Builder Andrea Petrucci - CERN (PH/CMD) on behalf of the CMS DAQ group 19 th International Conference on Computing in High Energy.
7. CBM collaboration meetingXDAQ evaluation - J.Adamczewski1.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
1 Network Performance Optimisation and Load Balancing Wulf Thannhaeuser.
An Architecture and Prototype Implementation for TCP/IP Hardware Support Mirko Benz Dresden University of Technology, Germany TERENA 2001.
IT/EE Palaver FAIR DAQ - J.Adamczewski, S.Linev1.
McGraw-Hill©The McGraw-Hill Companies, Inc., 2004 Connecting Devices CORPORATE INSTITUTE OF SCIENCE & TECHNOLOGY, BHOPAL Department of Electronics and.
2003 Conference for Computing in High Energy and Nuclear Physics La Jolla, California Giovanna Lehmann - CERN EP/ATD The DataFlow of the ATLAS Trigger.
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
LNL 1 SADIRC2000 Resoconto 2000 e Richieste LNL per il 2001 L. Berti 30% M. Biasotto 100% M. Gulmini 50% G. Maron 50% N. Toniolo 30% Le percentuali sono.
The CMS Event Builder Demonstrator based on MyrinetFrans Meijers. CHEP 2000, Padova Italy, Feb The CMS Event Builder Demonstrator based on Myrinet.
1 Farm Issues L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
SRB data transmission Vito Palladino CERN 2 June 2014.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
L1/HLT trigger farm Bologna setup 0 By Gianluca Peco INFN Bologna Genève,
Artur BarczykRT2003, High Rate Event Building with Gigabit Ethernet Introduction Transport protocols Methods to enhance link utilisation Test.
LECC2004 BostonMatthias Müller The final design of the ATLAS Trigger/DAQ Readout-Buffer Input (ROBIN) Device B. Gorini, M. Joos, J. Petersen, S. Stancu,
Remigius K Mommsen Fermilab CMS Run 2 Event Building.
The Evaluation Tool for the LHCb Event Builder Network Upgrade Guoming Liu, Niko Neufeld CERN, Switzerland 18 th Real-Time Conference June 13, 2012.
EPS 2007 Alexander Oh, CERN 1 The DAQ and Run Control of CMS EPS 2007, Manchester Alexander Oh, CERN, PH-CMD On behalf of the CMS-CMD Group.
CHEP 2010, October 2010, Taipei, Taiwan 1 18 th International Conference on Computing in High Energy and Nuclear Physics This research project has.
Network Processing Systems Design
Eric Hazen1 Ethernet Readout With: E. Kearns, J. Raaf, S.X. Wu, others... Eric Hazen Boston University.
LISA Linux Switching Appliance Radu Rendec Ioan Nicu Octavian Purdila Universitatea Politehnica Bucuresti 5 th RoEduNet International Conference.
HTCC coffee march /03/2017 Sébastien VALAT – CERN.
Balazs Voneki CERN/EP/LHCb Online group
LHCb and InfiniBand on FPGA
High Rate Event Building with Gigabit Ethernet
CALICE TDAQ Application Network Protocols 10 Gigabit Lab
Module 12: I/O Systems I/O hardware Application I/O Interface
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
CMS DAQ Event Builder Based on Gigabit Ethernet
DAQ upgrades at SLHC S. Cittolin CERN/CMS, 22/03/07
The LHCb Event Building Strategy
Internetworking: Hardware/Software Interface
Operating System Concepts
CS703 - Advanced Operating Systems
Event Building With Smart NICs
Network Processors for a 1 MHz Trigger-DAQ System
Chapter 13: I/O Systems I/O Hardware Application I/O Interface
Chapter 13: I/O Systems I/O Hardware Application I/O Interface
Chapter 13: I/O Systems.
Achieving reliable high performance in LFNs (long-fat networks)
Module 12: I/O Systems I/O hardwared Application I/O Interface
Cluster Computers.
Presentation transcript:

A TCP/IP transport layer for the DAQ of the CMS Experiment Miklos Kozlovszky for the CMS TriDAS collaboration CERN European Organization for Nuclear Research ACAT03 - December 2003

CMS & Data Acquisition Collision rate 40 MHz Level-1 Maximum trigger rate 100 kHz Average event size ­ 1 Mbyte No. of In-Out units 1000 Readout network bandwidth ­ 1 Terabit/s Event filter computing power ­ MIPS Data production ­ Tbyte/day CMS Data

Event builder : Physical system interconnecting data sources with data destinations. It has to move each event data fragments into a same destination Event fragments : Event data fragments are stored in separated physical memory systems Full events : Full event data are stored into one physical memory system associated to a processing unit Data sources for 1 MByte events ~1000sHTL processing nodes NxM EVB Building the events

Distributed DAQ framework developed within CMS. Construct homogeneous applications for heterogeneous processing clusters. Multi-threaded (important to take advantage of SMP efficiently). Zero copy message passing for the event data. Peer to peer communication between the applications. I 2 O for data transport, and SOAP for configuration and control. Hardware and transport independency. OS and Device Drivers HTTP Ethernet Myrinet XDAQ Util/DDM Processing Sensor readout TCP PCI Subject of presentation XDAQ Framework

Reuse old, “cheap” Ethernet for DAQ Transport layer requirements –Reliable communication –Hide the complexity of TCP –Efficient implementation –Simplex communication via sockets –Configurable Support of blocking and non-blocking I/O TCP/IP Peer Transport Requirements

Pending Queues –Thread safe PQ management –One PQ for each destination –Independent sending through sockets Only one “Select” function call both to receive the packet and send the blocked data. Implementation of the non-blocking mode 12345n 12345n #2 Pending Queues XDAQ Application Framesen d 12345n #n Select

Receiver Object(s) OS XDAQ Executive Peer Transport Layer ptATCP Applications (XDAQ) ptATCPPort(s) XDAQ Framework Sender Object(s) Input SAP(s)Output SAP(s) Driver(s) NIC (10GE)NIC (FE)NIC (GE) = Creation of object = Sending = Receiving = other communication Communication via the transport layer

Throughput optimisation Single railMulti-rail App 1 App 2 App 1 Operating System tuning (kernel options+buffers) Jumbo Frames Transport protocol options Communication techniques –Blocking vs. Non-Blocking I/O –Single/Multi-rail –Single/Multi-thread –TCP options (e.g.:Nagle algorithm) –….

Test network Cluster size:8x8 CPU: 2x Intel Xeon (2.4 GHz), 512KB Cache I/O system:PCI-X: 4 buses (max 6). Memory: Two-way interleaved DDR: 3.2 GB/s (512 MB) NICs: 1 Intel 82540EM GE 1 Broadcom NeXtreme BCM 5703x GE 1 Intel Pro 2546EB GE (2port) OS: Linux RedHat (SMP) Switches: 1 BATM- T6 Multi Layer Gigabit Switch (medium range) 2 Dell Power Connect 5224 (medium range)

Conditions: XDAQ+Event Builder –No Readout Unit inputs –No Builder Unit outputs –No Event Manager PC: dual P4 Xeon Linux NIC: e-1000 Switch: Powerconnect 5224 Standard MTU (1500 Bytes) Each BU builds 128 events Fixed fragment sizes Result: For fragment size > 4 kB: Thru /node ~100 MB/s i.e. 80% utilisation Working point Event Building on the cluster

Two Rail Event Builder measurements Test case: Bare Event Builder (2x2) No RU inputs No BU outputs No Event Manager Options: Non blocking TCP Jumbo frames (mtu 8000) Two rail One thread RU working point (16 kB) Throughput/node = 240 MB/ s i.e. 95% bandwidth

Achieved 100 MB/s per node in 8x8 configuration (1rail). Improvements seen with the use of two rail, non-blocking I/O, with Jumbo frames. In 2x2 configuration over 230 MB/s obtained. High CPU load. We are also studying other networking and traffic shaping options. Conclusions