R.Divià, CERN/ALICE Challenging the challenge Handling data in the Gigabit/s range.

Slides:



Advertisements
Similar presentations
ARP Traffic Study Jim Rees, Manish Karir Research and Development Merit Network Inc.
Advertisements

CHEP 2012 – New York City 1.  LHC Delivers bunch crossing at 40MHz  LHCb reduces the rate with a two level trigger system: ◦ First Level (L0) – Hardware.
CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000.
Copyright© 2000 OPNET Technologies, Inc. R.W. Dobinson, S. Haas, K. Korcyl, M.J. LeVine, J. Lokier, B. Martin, C. Meirosu, F. Saka, K. Vella Testing and.
1P. Vande Vyvre - CERN/PH ALICE DAQ Technical Design Report DAQ TDR Task Force Tome ANTICICFranco CARENA Wisla CARENA Ozgur COBANOGLU Ervin DENESRoberto.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Remigius K Mommsen Fermilab A New Event Builder for CMS Run II A New Event Builder for CMS Run II on behalf of the CMS DAQ group.
CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/20 New Experiences with the ALICE High Level Trigger Data Transport.
Timm M. Steinbeck - Kirchhoff Institute of Physics - University Heidelberg - DPG 2005 – HK New Test Results for the ALICE High Level Trigger.
CHEP03 - UCSD - March 24th-28th 2003 T. M. Steinbeck, V. Lindenstruth, H. Tilsner, for the Alice Collaboration Timm Morten Steinbeck, Computer Science.
March 2003 CHEP Online Monitoring Software Framework in the ATLAS Experiment Serguei Kolos CERN/PNPI On behalf of the ATLAS Trigger/DAQ Online Software.
LHCb readout infrastructure NA62 TDAQ WG Meeting April 1 st, 2009 Niko Neufeld, PH/LBC.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
ALICE Data Challenge V P. VANDE VYVRE – CERN/PH LCG PEB - CERN March 2004.
MSS, ALICE week, 21/9/041 A part of ALICE-DAQ for the Forward Detectors University of Athens Physics Department Annie BELOGIANNI, Paraskevi GANOTI, Filimon.
A TCP/IP transport layer for the DAQ of the CMS Experiment Miklos Kozlovszky for the CMS TriDAS collaboration CERN European Organization for Nuclear Research.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
System performance monitoring in the ALICE Data Acquisition System with Zabbix Adriana Telesca October 15 th, 2013 CHEP 2013, Amsterdam.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
1 Alice DAQ Configuration DB
The ALICE DAQ: Current Status and Future Challenges P. VANDE VYVRE CERN-EP/AID.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
20-22 September 1999 HPSS User Forum, Santa Fe CERN IT/PDP 1 History  Test system HPSS 3.2 installation in Oct 1997 IBM AIX machines with IBM 3590 drives.
The ALICE Data-Acquisition Software Framework DATE V5 F. Carena, W. Carena, S. Chapeland, R. Divià, I. Makhlyueva, J-C. Marin, K. Schossmaier, C. Soós,
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
The BaBar Event Building and Level-3 Trigger Farm Upgrade S.Luitz, R. Bartoldus, S. Dasu, G. Dubois-Felsmann, B. Franek, J. Hamilton, R. Jacobsen, D. Kotturi,
ALICE Computing Model The ALICE raw data flow P. VANDE VYVRE – CERN/PH Computing Model WS – 09 Dec CERN.
Roberto Divià, CERN/ALICE 1 CHEP 2009, Prague, March 2009 The ALICE Online Data Storage System Roberto Divià (CERN), Ulrich Fuchs (CERN), Irina Makhlyueva.
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
Dynamic configuration of the CMS Data Acquisition Cluster Hannes Sakulin, CERN/PH On behalf of the CMS DAQ group Part 1: Configuring the CMS DAQ Cluster.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Fast Crash Recovery in RAMCloud. Motivation The role of DRAM has been increasing – Facebook used 150TB of DRAM For 200TB of disk storage However, there.
AFFAIR – fabric monitoring ROOT 2005 Tome Antičić Ruđer Bošković Institute, Zagreb,Croatia ALICE,CERN Tome Antičić Ruđer Bošković.
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
Latest ideas in DAQ development for LHC B. Gorini - CERN 1.
2003 Conference for Computing in High Energy and Nuclear Physics La Jolla, California Giovanna Lehmann - CERN EP/ATD The DataFlow of the ATLAS Trigger.
LHCb DAQ system LHCb SFC review Nov. 26 th 2004 Niko Neufeld, CERN.
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
An Efficient Gigabit Ethernet Switch Model for Large-Scale Simulation Dong (Kevin) Jin.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
Management of the LHCb Online Network Based on SCADA System Guoming Liu * †, Niko Neufeld † * University of Ferrara, Italy † CERN, Geneva, Switzerland.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Pierre VANDE VYVRE ALICE Online upgrade October 03, 2012 Offline Meeting, CERN.
CODA Graham Heyes Computer Center Director Data Acquisition Support group leader.
R.Divià, CERN/ALICE 1 ALICE off-line week, CERN, 9 September 2002 DAQ-HLT software interface.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
COMPASS DAQ Upgrade I.Konorov, A.Mann, S.Paul TU Munich M.Finger, V.Jary, T.Liska Technical University Prague April PANDA DAQ/FEE WS Игорь.
Remigius K Mommsen Fermilab CMS Run 2 Event Building.
The Evaluation Tool for the LHCb Event Builder Network Upgrade Guoming Liu, Niko Neufeld CERN, Switzerland 18 th Real-Time Conference June 13, 2012.
Barthélémy von Haller CERN PH/AID For the ALICE Collaboration The ALICE data quality monitoring system.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
The ALICE Data-Acquisition Read-out Receiver Card C. Soós et al. (for the ALICE collaboration) LECC September 2004, Boston.
ALICE Computing Data Challenge VI
NL Service Challenge Plans
PC Farms & Central Data Recording
LHC experiments Requirements and Concepts ALICE
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
CMS DAQ Event Builder Based on Gigabit Ethernet
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
ALICE Data Challenges On the way to 1 GB/s
ITS combined test seen from DAQ and ECS F.Carena, J-C.Marin
Example of DAQ Trigger issues for the SoLID experiment
ALICE Data Challenges Fons Rademakers Click to add notes.
Presentation transcript:

R.Divià, CERN/ALICE Challenging the challenge Handling data in the Gigabit/s range

R.Divià, CERN/ALICE 2 24 March 2003, CHEP03, La Jolla

R.Divià, CERN/ALICE 3 24 March 2003, CHEP03, La Jolla

R.Divià, CERN/ALICE 4 24 March 2003, CHEP03, La Jolla ALICE Data Acquisition architecture InnerTrackingSystem Local Data Concentrator Readout Receiver Card Front-EndElectronics TimeProjectionChamber Muon PhotonSpectrometerParticleIdentification Trigger Level 1 Trigger Level 2 Trigger Level 0 Triggerdata TriggerDetectors Trigger decisions Event Building switch  3 GB/s Global Data Collector Event Destination Manager Storage switch 1.25 GB/s Perm.DataStorage Detector Data Link

R.Divià, CERN/ALICE 5 24 March 2003, CHEP03, La Jolla ALICE running parameters u Two different running modes: m Heavy Ion (HI): 10 6 seconds/year m Proton: 10 7 seconds/year u One Data Acquisition system (DAQ): Data Acquisition and Test Environment (DATE) u Many triggers classes each providing events at different rates, sizes and sources u HI data rates:  3 GB/s  1.25 GB/s  ~1 PB/year to mass storage u Proton run: ~ 0.5 PB/year to mass storage u Staged DAQ installation plan (20%  30%  100%): m 85  300 LDCs, 10  40 Global Data Collectors (GDC) u Different recording options: m Local/remote disks m Permanent Data Storage (PDS): CERN Advanced Storage Manager (CASTOR)

R.Divià, CERN/ALICE 6 24 March 2003, CHEP03, La Jolla History of ALICE Data Challenges u Started in 1998 to put together a high-bandwidth DAQ/recording chain u Continued as a periodic activity to: m Validate interoperability of all existing components m Assess and validate developments, trends and options  commercial products  in-house developments m Provide guidelines for ALICE & IT development and installation u Continuously expand up to ALICE requirement at LHC startup

R.Divià, CERN/ALICE 7 24 March 2003, CHEP03, La Jolla Performance goals MBytes/s

R.Divià, CERN/ALICE 8 24 March 2003, CHEP03, La Jolla Data volume goals TBytes to Mass Storage

R.Divià, CERN/ALICE 9 24 March 2003, CHEP03, La Jolla TheALICE Data Challenge IV

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla RAWEVENTS Objectifier RAWDATAOBJECTS LDCemulator Components & modes ALICEDAQCASTOR PDS AFFAIR CASTOR monitor Private network CERN backbone CASTOR FE

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Targets u DAQ system scalability tests u Single peer-to-peer tests: m Evaluate the behavior of the DAQ system components with the available HW m Preliminary tuning u Multiple LDC/GDC tests: m Add the full Data Acquisition (DAQ) functionality m Verify the objectification process m Validate & benchmark the CASTOR I/F u Evaluate the performance of new hardware components: m New generation of tapes m 10 Gb Ethernet u Achieve a stable production period: m Minimum 200 MB/s sustained m 7 days non stop m 200 TB data to PDS

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Software components u Configurable LDC Emulator (COLE) u Data Acquisition and Test Environment (DATE) 4.2 u A Fine Fabric and Application Information Recorder (AFFAIR) V1 u ALICE Mock Data Challenge objectifier (ALIMDC) u ROOT (Object-Oriented Data Analysis Framework) v3.03 u Permanent Data Storage (PDS): CASTOR V u Linux RedHat 7.2, kernel 2.2 and 2.4 m Physical pinned memory driver (PHYSMEM) m Standard TCP/IP library

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Hardware setup u ALICE DAQ: infrastructure & benchmarking m NFS & DAQ servers m SMP HP Netserver (4 CPUs): setup & benchmarking u LCG testbed (lxshare): setup & production m 78 CPU servers on GE  Dual ~1GHz Pentium III, 512 MB RAM  Linux kernel 2.2 and 2.4  NFS (installation, distribution) and AFS (unused) m [ ] DISK servers (IDE-based) on GE u Mixed FE/GE/trunk GE, private & CERN backbone m 2 * Extreme Networks Summit 7i switches (32 GE ports) m 12 * 3COM 4900 switches (16 GE ports) m CERN backbone: Enterasys SSR8600 routers (28 GE ports) u PDS: 16 * 9940B tape drives in two different buildings m STK linear tapes, 30 MB/s, 200 GB/cartridge

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Networking TAPE servers (distributed) Backbone (4 Gbps) 6 CPU servers on FE LDCs & GDCs DISK servers

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Scalability test u Put together as many hosts as possible to verify the scalability of: m run control m state machines m control and data channels m DAQ services m system services m hardware infrastructure u Connect/control/disconnect plus simple data transfers u Data patterns, payloads and throughputs uninteresting u Keywords: usable, reliable, scalable, responsive

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Scalability test

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Single peer-to-peer u Compare: m Architectures m Network configurations m System and DAQ parameters u Exercise: m DAQ system network modules m DAQ system clients and daemons m Linux system calls, system libraries and network drivers u Benchmark and tune: m Linux parameters m DAQ processes, libraries and network components m DAQ data flow

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Single peer-to-peer

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Full test runtime options u Different trigger classes for different traffic patterns u Several recording options m NULL m GDC disk m CASTOR disks m CASTOR tapes u Raw data vs. ROOT objects u We concentrated on two major traffic patterns: m Flat traffic: all LDCs send the same event m ALICE-like traffic: periodic sequence of different events distributed according to forecasted ALICE raw data

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Performance Goals MBytes/s 650 MB/s

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Flat data traffic u 40 LDCs * 38 GDCs u 1 MB/event/LDC  NULL u Occupancies: m LDCs: 75% m GDCs: 50% u Critical item: load balancing over the GE trunks (2/3 nominal)

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Load distribution on trunks # LDCs Distributed Same switch MB/s

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla ALICE-like traffic u LDCs: m rather realistic simulation m partitioned in detectors m no hardware trigger m simulated readout, no “real” input channels u GDCs acting as: m event builder m CASTOR front-end u Data traffic: m Realistic event sizes and trigger classes m Partial detector readout u Networking & nodes’ distribution scaled down & adapted

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Challenge setup & outcomes u ~ 25 LDCs m TPC: 10 LDCs m others detectors: [ ] LDCs u ~ 50 GDCs u Each satellite switch: 12 LDCs/GDCs (distributed) u [ ] (+1) tape servers on the CERN backbone u [ ] (+1) tape drives attached to a tape server u No objectification m named pipes too slow and too heavy m upgraded to avoid named pipes:  ALIMDC/CASTOR not performing well

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Impact of traffic pattern FLAT/CASTORALICE/NULL ALICE/CASTOR

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Performance Goals MBytes/s 200 MB/s

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Production run u 8 LDCs*16 GDCs, 1 MB/event/LDC (FLAT traffic) u [ ] tape server and tape units u 7 days at ~300 MB/s sustained, > 350 MB/s peak, ~ 180 TB to tape u 9 Dec: too much input data u 10 Dec: HW failures on tape drives & reconfiguration u Despite the failures, always exceeded the performance goals

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla System reliability u Hosts: m ~ 10% Dead On Installation m ~ 25% Failed On Installation u Long period of short runs (tuning): m occasional problems (recovered) with:  name server  network & O.S. m in average [1.. 2 ] O.S. failures per week (on 77 hosts)  unrecoverable occasional failures on GE interfaces u Production run: m one tape unit failed and had to be excluded

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Outcomes u DATE m 80 hosts/160+ roles with one run control m Excellent reliability and performance m Scalable and efficient architecture u Linux m Few hiccups here and there but rather stable and fast m Excellent network performance/CPU usage m Some components are too slow (e.g. named pipes) m More reliability needed from the GE interfaces

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Outcomes u FARM installation and operation: not to be underestimated! u CASTOR m Reliable and effective m Improvements needed on:  Overloading  Parallelizing tape resources u Tapes m One DOA and one DOO u Network: silent but very effective partner m Layout made for a farm, not optimized for ALICE DAQ u 10 GB Ethernet tests: m failure at first m problem “fixed” too late for the Data Challenge m reconfiguration: transparent to DAQ and CASTOR

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla Future ALICE Data Challenges u Continue the planned progression u ALICE-like pattern u Record ROOT objects u New technologies m CPUs m Servers m Network  NICs  Infrastructure  Beyond 1 GbE u Insert online algorithms u Provide some “real” input channels u Challenge the challenge! MBytes/s

R.Divià, CERN/ALICE March 2003, CHEP03, La Jolla