LNL CMS G. MaronCPT Week CERN, 23 April 2001 1 Legnaro Event Builder Prototypes Luciano Berti, Gaetano Maron Luciano Berti, Gaetano Maron INFN – Laboratori.

Slides:



Advertisements
Similar presentations
Interconnection Networks: Flow Control and Microarchitecture.
Advertisements

CCNA3: Switching Basics and Intermediate Routing v3.0 CISCO NETWORKING ACADEMY PROGRAM Switching Concepts Introduction to Ethernet/802.3 LANs Introduction.
Chabot College Chapter 2 Review Questions Semester IIIELEC Semester III ELEC
Protocols and software for exploiting Myrinet clusters Congduc Pham and the main contributors P. Geoffray, L. Prylli, B. Tourancheau, R. Westrelin.
1 Version 3 Module 8 Ethernet Switching. 2 Version 3 Ethernet Switching Ethernet is a shared media –One node can transmit data at a time More nodes increases.
Copyright© 2000 OPNET Technologies, Inc. R.W. Dobinson, S. Haas, K. Korcyl, M.J. LeVine, J. Lokier, B. Martin, C. Meirosu, F. Saka, K. Vella Testing and.
Remigius K Mommsen Fermilab A New Event Builder for CMS Run II A New Event Builder for CMS Run II on behalf of the CMS DAQ group.
Nov-03 ©Cisco Systems CCNA Semester 1 Version 3 Comp11 Mod8 – St. Lawrence College – Cornwall Campus, ON, Canada – Clark slide 1 Cisco Systems CCNA Version.
Service Providers & Data Link & Physical layers Week 4 Lecture 1.
1 K. Salah Module 4.3: Repeaters, Bridges, & Switches Repeater Hub NIC Bridges Switches VLANs GbE.
Can Google Route? Building a High-Speed Switch from Commodity Hardware Guido Appenzeller, Matthew Holliman Q2/2002.
Switches in Networking B. Konkoth. Network Traffic  Scalability  Ability to handle growing amount of work  Capability of a system to increase performance.
Connecting LANs, Backbone Networks, and Virtual LANs
Interconnect Network Topologies
High Performance Embedded Computing © 2007 Elsevier Lecture 16: Interconnection Networks Embedded Computing Systems Mikko Lipasti, adapted from M. Schulte.
Introduction to IT and Communications Technology Justin Champion C208 – 3292 Ethernet Switching CE
© 1999, Cisco Systems, Inc. 3-1 Chapter 10 Controlling Campus Device Access Chapter 3 Connecting the Switch Block © 1999, Cisco Systems, Inc. 3-1.
Semester 1 Module 8 Ethernet Switching Andres, Wen-Yuan Liao Department of Computer Science and Engineering De Lin Institute of Technology
Interconnect Networks
Brierley 1 Module 4 Module 4 Introduction to LAN Switching.
1 Module 15: Network Structures n Topology n Network Types n Communication.
A TCP/IP transport layer for the DAQ of the CMS Experiment Miklos Kozlovszky for the CMS TriDAS collaboration CERN European Organization for Nuclear Research.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
S3C2 – LAN Switching Addressing LAN Problems. Congestion is Caused By Multitasking, Faster operating systems, More Web-based applications Client-Server.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
LECC2003 AmsterdamMatthias Müller A RobIn Prototype for a PCI-Bus based Atlas Readout-System B. Gorini, M. Joos, J. Petersen (CERN, Geneva) A. Kugel, R.
Module 8: Ethernet Switching
A Measurement Based Memory Performance Evaluation of High Throughput Servers Garba Isa Yau Department of Computer Engineering King Fahd University of Petroleum.
CCNA 3 Week 4 Switching Concepts. Copyright © 2005 University of Bolton Introduction Lan design has moved away from using shared media, hubs and repeaters.
1 © 2003, Cisco Systems, Inc. All rights reserved. CCNA 3 v3.0 Module 4 Switching Concepts.
Cisco 3 - Switching Perrine. J Page 16/4/2016 Chapter 4 Switches The performance of shared-medium Ethernet is affected by several factors: data frame broadcast.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
1 Network Performance Optimisation and Load Balancing Wulf Thannhaeuser.
Sem1 - Module 8 Ethernet Switching. Shared media environments Shared media environment: –Occurs when multiple hosts have access to the same medium. –For.
Packet switching network Data is divided into packets. Transfer of information as payload in data packets Packets undergo random delays & possible loss.
Interconnect simulation. Different levels for Evaluating an architecture Numerical models – Mathematic formulations to obtain performance characteristics.
Interconnect simulation. Different levels for Evaluating an architecture Numerical models – Mathematic formulations to obtain performance characteristics.
CMS week, June 2002, CERN 1 First P2P Measurements on Infiniband Luciano Berti INFN Laboratori Nazionali di Legnaro.
LAN Switching Concepts. Overview Ethernet networks used to be built using repeaters. When the performance of these networks began to suffer because too.
McGraw-Hill©The McGraw-Hill Companies, Inc., 2004 Connecting Devices CORPORATE INSTITUTE OF SCIENCE & TECHNOLOGY, BHOPAL Department of Electronics and.
Chapter 13 – I/O Systems (Pgs ). Devices  Two conflicting properties A. Growing uniformity in interfaces (both h/w and s/w): e.g., USB, TWAIN.
LNL 1 SADIRC2000 Resoconto 2000 e Richieste LNL per il 2001 L. Berti 30% M. Biasotto 100% M. Gulmini 50% G. Maron 50% N. Toniolo 30% Le percentuali sono.
Interconnect Networks Basics. Generic parallel/distributed system architecture On-chip interconnects (manycore processor) Off-chip interconnects (clusters.
The CMS Event Builder Demonstrator based on MyrinetFrans Meijers. CHEP 2000, Padova Italy, Feb The CMS Event Builder Demonstrator based on Myrinet.
CISCO NETWORKING ACADEMY Chabot College ELEC Ethernet Switches.
CCNA3 Module 4 Brierley Module 4. CCNA3 Module 4 Brierley Topics LAN congestion and its effect on network performance Advantages of LAN segmentation in.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
1 Event Building L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
Simple LAN Topologies  Terminators stop signals after they have reached their destination  Signal bounce –Phenomenon in which signals travel endlessly.
Artur BarczykRT2003, High Rate Event Building with Gigabit Ethernet Introduction Transport protocols Methods to enhance link utilisation Test.
+ Lecture#2: Ethernet Asma ALOsaimi. + Objectives In this chapter, you will learn to: Describe the operation of the Ethernet sublayers. Identify the major.
TCP/IP1 Address Resolution Protocol Internet uses IP address to recognize a computer. But IP address needs to be translated to physical address (NIC).
Remigius K Mommsen Fermilab CMS Run 2 Event Building.
Grzegorz Korcyl - Jagiellonian University, Kraków Grzegorz Korcyl – PANDA TDAQ Workshop, Giessen April 2010.
Chapter 3: Link Aggregation
Youngstown State University Cisco Regional Academy
Introduction to Networks v6.0
Deterministic Communication with SpaceWire
High Rate Event Building with Gigabit Ethernet
Instructor Materials Chapter 5: Ethernet
CMS DAQ Event Builder Based on Gigabit Ethernet
Event Building With Smart NICs
Network Processors for a 1 MHz Trigger-DAQ System
Network Systems and Throughput Preservation
Chapter 13: I/O Systems I/O Hardware Application I/O Interface
CS 6290 Many-core & Interconnect
Routing and the Network Layer (ref: Interconnections by Perlman
Myrinet 2Gbps Networks (
LHCb Online Meeting November 15th, 2000
Multiprocessors and Multi-computers
Presentation transcript:

LNL CMS G. MaronCPT Week CERN, 23 April Legnaro Event Builder Prototypes Luciano Berti, Gaetano Maron Luciano Berti, Gaetano Maron INFN – Laboratori Nazionali di Legnaro

LNL CMS G. MaronCPT Week CERN, 23 April GE Event Builder Test conditions: No command or event aggregation (each packet transports a command or data frame relative to a single event) No command or event aggregation (each packet transports a command or data frame relative to a single event) full data transfer from/to PC memory full data transfer from/to PC memory recovery from packet loss recovery from packet loss fixed fragment sizes are varied bytes fixed fragment sizes are varied bytes Components:Hardware: switch: FoundryNet FastIron switch: FoundryNet FastIron NIC: SysKonnect SK9821 NIC: SysKonnect SK9821 PC: Supermicro PIII (i840) PC: Supermicro PIII (i840) Software: vxWorks based 15 x 15

LNL CMS G. MaronCPT Week CERN, 23 April Event builder layout EVM RUs BUs Slot 1 Slot 2 Slot 3 Slot 4 RUs and BUs distributed in all switch slots: – Part of the traffic localized within the slot – Reduces switch backplane utilization RU performance problem found with problem found with this configuration this configuration

LNL CMS G. MaronCPT Week CERN, 23 April Modified Event Builder layout EVM RUs BUs Slot 1 Slot 2 Slot 3 Slot 4 Request data commands Fast Ethernet Slot Request data commands - RU fast control message over FE (PCI 32/33) - RU data transfer on GE (PCI 64/66)

LNL CMS G. MaronCPT Week CERN, 23 April The GE Event Builder

LNL CMS G. MaronCPT Week CERN, 23 April EB protocol BU allocate EVM confirm 1 23n RUs send cache

LNL CMS G. MaronCPT Week CERN, 23 April Concurrent building threads in the same BU BU EVM 1 23n RUs BU thread 1 BU thread 2 BU thread 3

LNL CMS G. MaronCPT Week CERN, 23 April BU allocate EVM confirm Sequential reading 1 23n RUs BU allocate EVM confirm Random reading 1 23n RUs 45 send cache send cache Sequential vs Random reading

LNL CMS G. MaronCPT Week CERN, 23 April BU allocate EVM confirm 1 23n RUs 45 send cache “Sliding Window” multiple send to Rusmultiple send to Rus reduce the total rebuilding timereduce the total rebuilding time less events in the Busless events in the Bus not yet testednot yet tested

LNL CMS G. MaronCPT Week CERN, 23 April No difference on performanceNo difference on performance But more allocated event needed on BUs,But more allocated event needed on BUs, All the measurements with random readingAll the measurements with random reading Sequential reading Random reading Sequential - random reading comparison

LNL CMS G. MaronCPT Week CERN, 23 April Recovery from Packets loss BU Req. Data RU timer timer start EvtData Req. Data ( retry ) start cancel timeout EvtData BU Req. EvtId EVM timer timer start EvtId Req. EvtId (retry) start cancel timeout EvtId BU – EVM communication BU – RU communication Timeouts ms

LNL CMS G. MaronCPT Week CERN, 23 April EVB 15x15 performance - Throughput Throughput up to 116 MB/s, ie 93% link speed Throughput up to 116 MB/s, ie 93% link speed no packet loss observed (as expected) no packet loss observed (as expected) 15 x 15

LNL CMS G. MaronCPT Week CERN, 23 April EVB Scaling

LNL CMS G. MaronCPT Week CERN, 23 April EVB Performance – Event Rate Nominal fragment size 2kbytes: event rate = 52 kHz 15 x 15

LNL CMS G. MaronCPT Week CERN, 23 April Conic Event Builder Event Manager Manager Builder Network RU BU FU FU FU FU Event Manager Manager RU FUFUFUFUFUFUFUFU Builder Network FUFUFUFU symmetric EVB conic EVB faster ports at Rus faster ports at Rus slower ports at BUs slower ports at BUs

LNL CMS G. MaronCPT Week CERN, 23 April Conic Event Builder: Layout EVM RUs FUs GE Slot 1 Request Data Command FE Slot 1 FE Slot 2

LNL CMS G. MaronCPT Week CERN, 23 April EVB throughput – Conic vs Symmetric conic EVB: no performance degradation vs symmetric

LNL CMS G. MaronCPT Week CERN, 23 April EVB Conic – Scaling

LNL CMS G. MaronCPT Week CERN, 23 April Conic: RU/FU Throughput ratio

LNL CMS G. MaronCPT Week CERN, 23 April To be done and test variable size eventsvariable size events EB performances with the new implemented “ window” mechanismEB performances with the new implemented “ window” mechanism latency times measurementslatency times measurements Fault generation with the new implemented Random Error Generator to check the error recovery procedureFault generation with the new implemented Random Error Generator to check the error recovery procedure

LNL CMS G. MaronCPT Week CERN, 23 April Multistage Event Builder All our results have been obtained with a single switch event builder configurationAll our results have been obtained with a single switch event builder configuration We propose to extend our tests to a multistage ethernet switches topology and to study the behavior of this configuration.We propose to extend our tests to a multistage ethernet switches topology and to study the behavior of this configuration.

LNL CMS G. MaronCPT Week CERN, 23 April Plain Topology In the Event Builder application data flows in only one directionIn the Event Builder application data flows in only one direction The inter-switch Gigabit Ethernet links are full-duplexThe inter-switch Gigabit Ethernet links are full-duplex Result : half of the inter-switch bandwidth available is wastedResult : half of the inter-switch bandwidth available is wasted RUs BUs

LNL CMS G. MaronCPT Week CERN, 23 April Full Mesh Topology Full Mesh Topology RU and BU distributed in all the switchesRU and BU distributed in all the switches Inter-switch links are used in both directionInter-switch links are used in both direction Same number of ports of the plain topologySame number of ports of the plain topology Twice of the bandwidth of the plain topology in the inter-switch linksTwice of the bandwidth of the plain topology in the inter-switch links RUs BUs

LNL CMS G. MaronCPT Week CERN, 23 April Plain and mesh topology limits Plain and mesh topology limits Each couple of switches is connected by a single linkEach couple of switches is connected by a single link This is a bottleneck if the traffic is not uniformly randomThis is a bottleneck if the traffic is not uniformly random The network is blocking for certain traffic patternsThe network is blocking for certain traffic patterns

LNL CMS G. MaronCPT Week CERN, 23 April Traffic with patterns If traffic has patterns (for example this could happens in the case the event builder is performed in steps) it could make sense to introduce an artificial mechanism that randomise the traffic.If traffic has patterns (for example this could happens in the case the event builder is performed in steps) it could make sense to introduce an artificial mechanism that randomise the traffic. This mechanism exist and it is called Universal RoutingThis mechanism exist and it is called Universal Routing

LNL CMS G. MaronCPT Week CERN, 23 April Universal Routing Reference Discovered by L.G. Valiant in 1980Discovered by L.G. Valiant in 1980 See: M.D May, P.W. Thompson, P.H. Welch NETWORKS,ROUTERS & TRASPUTER available on : M.D May, P.W. Thompson, P.H. Welch NETWORKS,ROUTERS & TRASPUTER available on : Those papers describe the Universal Routing applied to Transputers Networks, a wormhole routing based networkThose papers describe the Universal Routing applied to Transputers Networks, a wormhole routing based network We adapted the same concept to a packet switched network like Gigabit EthernetWe adapted the same concept to a packet switched network like Gigabit Ethernet

LNL CMS G. MaronCPT Week CERN, 23 April Universal Routing with GigaEthernet Based on Clos topologyBased on Clos topology Multiple path available between each couple of switchesMultiple path available between each couple of switches Every packets is sent to a randomly chosen intermediate switchEvery packets is sent to a randomly chosen intermediate switch The intermediate switch send the packet to the final destinationThe intermediate switch send the packet to the final destination Full bandwidth between each couple of switches and uniform buffer utilizationFull bandwidth between each couple of switches and uniform buffer utilization

LNL CMS G. MaronCPT Week CERN, 23 April Universal Routing Transformation of the CLOS topology to a folded CLOSTransformation of the CLOS topology to a folded CLOS The resulting number of ports is the same of the plain topologyThe resulting number of ports is the same of the plain topology Half duplex links Full duplex links RUs BUs RUs BUs

LNL CMS G. MaronCPT Week CERN, 23 April Large (500x500) multistage GE network (1) BUs RUs 20 Ports 25 Ports 40 Ports 25 switches with 60 x 1Gb ports25 switches with 60 x 1Gb ports 20 switches with 25 x 1 Gb ports20 switches with 25 x 1 Gb ports

LNL CMS G. MaronCPT Week CERN, 23 April Large (500x500) multistage GE network (2) 25 switches with 40 x 1Gb ports + 2 x 10 Gb uplinks25 switches with 40 x 1Gb ports + 2 x 10 Gb uplinks 2 switches with 25 x 10 Gb ports2 switches with 25 x 10 Gb ports BUs 20 RUs 2 Ports 10G 25 Ports 10G 40 Ports

LNL CMS G. MaronCPT Week CERN, 23 April Proposal for a multistage event builder demonstrator Multistage event builders can be emulated using the much cheaper fast ethernet connections and switches. The GE speed is not needed in these topological investigationsMultistage event builders can be emulated using the much cheaper fast ethernet connections and switches. The GE speed is not needed in these topological investigations The proposal is to have prototypes for:The proposal is to have prototypes for: –Full Mesh Topology –Folded CLOS topology with (and without) Universal Routing mechanism

LNL CMS G. MaronCPT Week CERN, 23 April Full Mesh 64x64 Event Builder Prototype RUs 8 BUs - 1 host node 4 Rus / 4 BUs or a mix of them a mix of them - 32 hosts FE NICS ( ) FE ports switch Missing components

LNL CMS G. MaronCPT Week CERN, 23 April Folded CLOS 64x64 Event Builder Prototype BUs RUs 16 Ports 4 Ports 32 Ports 3 FastIron with 3 24 FE ports mods - 1 host node 4 Rus / 4 BUs or a mix of them a mix of them - 32 hosts FE NICS ( ) FE ports switch FE ports FastIron module FE ports FastIron module Missing components - 1 host node 3 Rus / 3 BUs or a mix of them a mix of them - 32 hosts - 96 FE NICS ( ) FE ports switch FE ports FastIron module 64x64 48x48

LNL CMS G. MaronCPT Week CERN, 23 April Folded CLOS 80x80 Event Builder Prototype BUs 20 RUs 2 GE Ports 4 Ports 40 FE Ports 3 FastIron with 8 GE ports module (1000 BaseT or 1000 BaseSX) - 1 host node 4 Rus / 4 BUs or a mix of them a mix of them - 40 hosts (32+8) FE NICS ( ) FE ports + 2 GE links switch GE (Base SX) ports FastIron module Missing components 80x80 2

LNL CMS G. MaronCPT Week CERN, 23 April x80 Conic Event Builder Prototype FUs 2 GE Ports 20 FE Ports 3 RU1 RU2 RU3 RU4 RU5 RU6 RU7 RU8 FastIron with 2 8 GE ports modules (1000 BaseT or 1000 BaseSX) - 1 host node 4 FUs - 20 hosts - 80 FE NICS ( ) (48) FE ports + 2 GE up links switch GE (Base SX) ports FastIron module Missing components 8x80

LNL CMS G. MaronCPT Week CERN, 23 April Material for the event builder multistage prototypes Mesh 64x FE NICs FE ports Folded CLOS 80x PCs FE NICS FE ports with 2 GE uplinks 1000 baseT if the 1000 baseT uplinks are not available: 1) Folded CLOS 64x64: 72 FE NICs 4 48 FE ports switch 1 24 FE ports FastIron module 2) Folded CLOS 48x48: 40 FE NICs 4 >36 FE ports switch