Pierre VANDE VYVRE for the O 2 project 15-Oct-2013 – CHEP – Amsterdam, Netherlands.

Slides:



Advertisements
Similar presentations
High Level Trigger (HLT) for ALICE Bergen Frankfurt Heidelberg Oslo.
Advertisements

LHCb Upgrade Overview ALICE, ATLAS, CMS & LHCb joint workshop on DAQ Château de Bossey 13 March 2013 Beat Jost / Cern.
Torsten Alt - KIP Heidelberg IRTG 28/02/ ALICE High Level Trigger The ALICE High-Level-Trigger.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/20 New Experiences with the ALICE High Level Trigger Data Transport.
PCIe based readout U. Marconi, INFN Bologna CERN, May 2013.
Trigger and online software Simon George & Reiner Hauser T/DAQ Phase 1 IDR.
Quality Control B. von Haller 8th June 2015 CERN.
Tom Dietel University of Cape Town for the ALICE Collaboration Computing for ALICE at the LHC.
FLP data flow – FLP prototype Filippo Costa ALICE O 2 9 th June 2015 WUHAN, CHINA.
ALICE O 2 Plenary | October 1 st, 2014 | Pierre Vande Vyvre O2 Project Status P. Buncic, T. Kollegger, Pierre Vande Vyvre 1.
1 Status report on DAQ E. Dénes – Wigner RCF/NFI ALICE Bp meeting – March
Use of GPUs in ALICE (and elsewhere) Thorsten Kollegger TDOC-PG | CERN |
The ALICE DAQ: Current Status and Future Challenges P. VANDE VYVRE CERN-EP/AID.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
ALICE Computing Model The ALICE raw data flow P. VANDE VYVRE – CERN/PH Computing Model WS – 09 Dec CERN.
Network Architecture for the LHCb DAQ Upgrade Guoming Liu CERN, Switzerland Upgrade DAQ Miniworkshop May 27, 2013.
The new CMS DAQ system for LHC operation after 2014 (DAQ2) CHEP2013: Computing in High Energy Physics Oct 2013 Amsterdam Andre Holzner, University.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Clara Gaspar, March 2005 LHCb Online & the Conditions DB.
Data Acquisition Backbone Core J. Adamczewski-Musch, N. Kurz, S. Linev GSI, Experiment Electronics, Data processing group.
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015.
Latest ideas in DAQ development for LHC B. Gorini - CERN 1.
VLVnT09A. Belias1 The on-shore DAQ system for a deep-sea neutrino telescope A.Belias NOA-NESTOR.
The KLOE computing environment Nuclear Science Symposium Portland, Oregon, USA 20 October 2003 M. Moulson – INFN/Frascati for the KLOE Collaboration.
LHCb front-end electronics and its interface to the DAQ.
1 LHCC RRB SG 16 Sep P. Vande Vyvre CERN-PH On-line Computing M&O LHCC RRB SG 16 Sep 2004 P. Vande Vyvre CERN/PH for 4 LHC DAQ project leaders.
P. Vande Vyvre – CERN/PH CERN – January Research Theme 2: DAQ ARCHITECT – Jan 2011 P. Vande Vyvre – CERN/PH2 Current DAQ status: Large computing.
LHC experimental data: From today’s Data Challenges to the promise of tomorrow B. Panzer – CERN/IT, F. Rademakers – CERN/EP, P. Vande Vyvre - CERN/EP Academic.
CWG4 – The data model The group proposes a time frame - based data model to: – Formalize the access to data types produced by both detector FEE and data.
Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015.
The Past... DDL in ALICE DAQ The DDL project ( )  Collaboration of CERN, Wigner RCP, and Cerntech Ltd.  The major Hungarian engineering contribution.
O 2 Project Roadmap P. VANDE VYVRE 1. O2 Project: What’s Next ? 2 O2 Plenary | 11 March 2015 | P. Vande Vyvre TDR close to its final state and its submission.
XLV INTERNATIONAL WINTER MEETING ON NUCLEAR PHYSICS Tiago Pérez II Physikalisches Institut For the PANDA collaboration FPGA Compute node for the PANDA.
Filippo Costa ALICE DAQ ALICE DAQ future detector readout October 29, 2012 CERN.
ALICE Online Upgrade P. VANDE VYVRE – CERN/PH ALICE meeting in Budapest – March 2012.
ICHEC Presentation ESR2: Reconfigurable Computing and FPGAs ICE-DIP Srikanth Sridharan 9/2/2015.
A. KlugeFeb 18, 2015 CRU form factor discussion & HLT FPGA processor part II A.Kluge, Feb 18,
Niko Neufeld HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Pierre VANDE VYVRE ALICE Online upgrade October 03, 2012 Offline Meeting, CERN.
CWG13: Ideas and discussion about the online part of the prototype P. Hristov, 11/04/2014.
LHCbComputing Computing for the LHCb Upgrade. 2 LHCb Upgrade: goal and timescale m LHCb upgrade will be operational after LS2 (~2020) m Increase significantly.
Monitoring for the ALICE O 2 Project 11 February 2016.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
ALICE O 2 project B. von Haller on behalf of the O 2 project CERN.
ALICE O 2 | 2015 | Pierre Vande Vyvre O 2 Project Pierre VANDE VYVRE.
Workshop ALICE Upgrade Overview Thorsten Kollegger for the ALICE Collaboration ALICE | Workshop |
16 September 2014 Ian Bird; SPC1. General ALICE and LHCb detector upgrades during LS2  Plans for changing computing strategies more advanced CMS and.
Rutherford Appleton Laboratory September 1999Fifth Workshop on Electronics for LHC Presented by S. Quinton.
ROM. ROM functionalities. ROM boards has to provide data format conversion. – Event fragments, from the FE electronics, enter the ROM as serial data stream;
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
DAQ 1000 Tonko Ljubicic, Mike LeVine, Bob Scheetz, John Hammond, Danny Padrazo, Fred Bieser, Jeff Landgraf.
P. Vande Vyvre – CERN/PH for the ALICE collaboration CHEP – October 2010.
The Evaluation Tool for the LHCb Event Builder Network Upgrade Guoming Liu, Niko Neufeld CERN, Switzerland 18 th Real-Time Conference June 13, 2012.
Predrag Buncic CERN Plans for Run2 and the ALICE upgrade in Run3 ALICE Tier-1/Tier-2 Workshop February 2015.
DAQ thoughts about upgrade 11/07/2012
Use of FPGA for dataflow Filippo Costa ALICE O2 CERN
LHCb and InfiniBand on FPGA
Challenges in ALICE and LHCb in LHC Run3
evoluzione modello per Run3 LHC
LHC experiments Requirements and Concepts ALICE
Enrico Gamberini, Giovanna Lehmann Miotto, Roland Sipos
Outline ALICE upgrade during LS2
ALICE – First paper.
Commissioning of the ALICE HLT, TPC and PHOS systems
ProtoDUNE SP DAQ assumptions, interfaces & constraints
ALICE Computing Upgrade Predrag Buncic
LHCb Trigger, Online and related Electronics
Implementation of DHLT Monitoring Tool for ALICE
Presentation transcript:

Pierre VANDE VYVRE for the O 2 project 15-Oct-2013 – CHEP – Amsterdam, Netherlands

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 2 ALICE apparatus upgrade New computing requirements O 2 system  Detector read-out  Data volume reduction  Big data O 2 project  Computing Working Groups Next steps Outline

ALICE O2 Project - CHEP 2013 P. Vande Vyvre /19 (LHC 2nd Long Shutdown) Inner Tracking System (ITS)  New, high-resolution, low-material ITS Time Project Chamber (TPC)  Upgrade of TPC with replacement of MWPCs with GEMs  New pipelined continuous readout electronics New and common computing system for online and offline computing New 5-plane silicon telescope in front of the Muon Spectrometer ALICE LS2 Upgrade

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 4 Rate increase: from 500 Hz to 50 kHz  Physics topics require measurements characterized by very small signal-over-background ratio → large statistics  Large background → traditional triggering or filtering techniques very inefficient for most physics channels.  Strategy: read out all particle interactions 50 kHz (anticipated Pb-Pb interaction rate) TPC intrinsic rate << 50 kHz  In average 5 events overlapping in the detector  Continuous read-out Requirements: Event Rate

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 5 Massive data volume reduction needed Only option is by online processing Requirements: Data Volume Detector Event Size kHz Pb-Pb (GByte/s) After Zero Suppression (MByte) TPC TRD ITS0.840 Others0.525 Total

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 6 O 2 Hardware System 2 x 10 or 40 Gb/s FLP 10 Gb/s FLP ITS TRD Muon FTP L0 L1 FLP EMC FLP TPC FLP TOF FLP PHO Trigger Detectors ~ 2500 links in total ~ 250 FLPs First Level Processors EPN Data Storage EPN Storage Network Storage Network Farm Network Farm Network 10 Gb/s ~ 1250 EPNs Event Processing nodes

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 7 GBT: custom radiation-hard optical link Detector Interface trigger link

ALICE O2 Status P. Vande Vyvre 8 DDL3: 10 Gb/s (LoI target) using a commercial standard (Ethernet or serial PCIe) Commercial products at 40 or 56 Gb/s available now  Dual-port 40 GbE Network Interface Card (NIC) (Chelsio) (40 GbE made of four lanes of multi-mode fiber at 10 Gb/s)  Dual-port 56 GbIB (QDR) (Mellanox)  Multiplex 4 x DDL3 over 1 input port of a commercial NIC o Breakout splitter cable (4 x 10 Gbe ↔ 1 x 40 GbE) o Commercial network switch (staging, DCS data demultiplex, etc)  Both options tested in the lab with equipment on loan giving the expected performance of 4 x 10 Gb/s Detector Data Link 3 (DDL3) FLP 10 x 10 Gb/s Det. Read-Out Data Source FLP NIC O 2 Farm Network O 2 Farm Network EPN DCS Server DCS Network DCS Network Commercial Network Switch 40 Gb/s Commercial Network Switch 40 Gb/s

ALICE O2 Status P. Vande Vyvre 9 DDL1 at 2 Gb/s used by all ALICE detectors for Run 1 (radiation tolerant) DDL2 at 4 and 5 Gb/s (according to needs) ready for Run 2 Prototype for one of the DDL3 option considered for Run 3 implemented (Eth. + UDP/IP) Expected performance evolution verified DDL Performance Evolution Plot F. Costa

ALICE O2 Status P. Vande Vyvre 10 1 key element for the O2 system will be the I/O bandwidth of the PCs PCIe Gen2 performance measured for the LoI PCIe Gen3 measured with a FPGA development board (Xilinx Virex-7 Connectivity Kit VC709)  Large data blocks: wire speed 8 GB/s, theoretical max 7.2, measured 5.8  FLP I/O capacity needed will at least require: 3 slots PCIe Gen 3 x8 or 2 slot x16 PC Input/Output Bandwidth Plot H. Hengel M. Husejko

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 11 TPC data volume reduction by online event reconstruction Discarding original raw data In production from the 2011 Pb-Pb run TPC Data Volume Reduction Data Format Data Reduction Factor Event Size (MByte) Raw Data1700 FEEZero Suppression3520 HLT Clustering & Compression5-7~3 Remove clusters not associated to relevant tracks 21.5 Data format optimization2-3<1 HLT Pb-Pb 2011

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 12 LHC luminosity variation during fill and efficiency taken into account for average output to computing center Total Data Volume Detector Input to Online System (GByte/s) Peak Output to Local Data Storage (GByte/s) Avg. Output to Computing Center (GByte/s) TPC TRD ITS Others Total

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 13 Shift from 1 to many platforms  Intel Xeon X64  Intel Xeon Phi (many cores)  GPUs  Low cost processors  FPGAs Benchmarking in progress to assess their relative merits Heterogeneous Platforms

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 14 Dataflow Model (1) QA Db FLPs EPNs Detectors electronics Data Sample n+1 Data Sample n Data Sample n-1 Buffering TPC - e.g. TPC clusters Condition & Calibration Databases - Global reco - Tracks/Tracklets - Drift time - Vertex Raw Data Input Local Processing Data aggregation Frame dispatch Event building Tagging Global Processing Event Building QA/DQM TRD Raw data sample Sub Time Frame Full Time Frame Event Full Time Frame Raw data Data Reduction 1 … Trigger and clock Calibration 0 (on local data, ie. partial detector) Time slicing Raw and Processed data Calibration 1 (on full detectors) Raw data sample Events Full Time Frame Information from CCDB Time slicing could actually occur before, e.g. on the front-end or receiving part Data Reduction 0 Tagging Read-Out

ALICE O2 Project - CHEP Dataflow Model (2) QA Db Data storage Grid or EPNs EPNs Condition & Calibration Databases - Global reco - Tracks/Tracklets - Drift time - Vertex Local Storage Calibration 2 (on all events so far) Data aggregation Frame dispatch Event building Tagging Global Processing Event Building Storage QA/DQM Sub Time Frame Full Time Frame Events SimulationAnalysis Events Event Full Time Frame Raw data Data Reduction 1 Permanent Storage Calibration 1 (on full detectors) Events Full Time Frame Reconstruction improvements Information from CCDB Offline processing Full Time Frame P. Vande Vyvre

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 16 HEP is not alone in the computing universe ! 1 ZB/year in 2017 (Cisco) 35 ZB in 2020 (IBM) 1 ZB = 1’000 EB = 1’000’000 PB Number of users (Kissmetrics) Internet : an inflating universe…

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 17 “Hyper giants”: the 150 companies that control 50% of all traffic on the web (Arbor Networks) “Hyper giants”: the 150 companies that control 50% of all traffic on the web (Arbor Networks) Google : 100 billion searches/month, 38’500 searches/second Google : 100 billion searches/month, 38’500 searches/second YouTube: 6 billion hours of video are watched each month YouTube: 6 billion hours of video are watched each month Facebook 350 millions photos uploaded/day Facebook 350 millions photos uploaded/day HEP should definitely try to navigate in the wake of the Big Data hyper giants HEP should definitely try to navigate in the wake of the Big Data hyper giants …with a few very large galaxies !

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 18 Very large data sets  High Energy Physics data are inherently and embarrassingly parallel… but  At the luminosity targeted for the upgrade there will be some pile-up → Continuous dataflow → New framework must handle it Issues to become a Big Data shop  Lots of legacy software not designed for this paradigm  Fraction the work into small independent manageable tasks  Merge results Big Data approach

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 19 O 2 Framework For example Data Acquisition Cluster Finding Tracking ESD, AOD Analysis Simulation

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 20 O 2 Project Institution Boards Computing Board Online Institution Board Computing Working Groups Projects O 2 Steering Board Project Leaders DAQ CWG1 Architecture CWG2 Procedure & Tools HLT CWG3 DataFlow CWG4 Data Model Offline CWG5 Platforms CWG6 Calibration CWG13 Sw Framework CWGnn CWG7 Reconstruc. CWG8 Simulation CWGnn CWGnn CWGnn CWGnn CWG9 QA, DQM, Vi CWG10 Control CWG11 Sw Lifecycle CWG12 Hardware - 50 people active in 1-3 CWGs - Service tasks

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 21 Sep 2012ALICE Upgrade LoI Jan 2013Report of the DAQ-HLT-Offline software panel on “ALICE Computer software framework for LS2 upgrade” Mar 2013 O 2 Computing Working Groups Sep 2014 O 2 Technical Design Report Overall Schedule O 2 Technical Design Report O2 Computing Working Groups

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 22 Computing Working Groups CWG1 Architecture CWG2 Tools CWG1 Architecture CWG2 Tools CWG3 Dataflow CWG1 Architecture CWG2 Tools CWG3 Dataflow CWG4 Data Model CWG5 Computing Platforms CWG6 Calibration CWG1 Architecture CWG2 Tools CWG3 Dataflow CWG4 Data Model CWG5 Computing Platforms CWG6 Calibration CWG7 Reconstruction CWG1 Architecture CWG2 Tools CWG3 Dataflow CWG4 Data Model CWG5 Computing Platforms CWG6 Calibration CWG8 Physics Simulation CWG7 Reconstruction CWG1 Architecture CWG2 Tools CWG3 Dataflow CWG4 Data Model CWG5 Computing Platforms CWG1 Architecture CWG2 Tools CWG3 Dataflow CWG4 Data Model CWG5 Computing Platforms CWG6 Calibration CWG8 Physics Simulation CWG7 Reconstruction CWG9 QA, DQM CWG1 Architecture CWG2 Tools CWG3 Dataflow CWG4 Data Model CWG5 Computing Platforms CWG6 Calibration CWG9 QA, DQM CWG10 Control, Configuration CWG8 Physics Simulation CWG7 Reconstruction CWG1 Architecture CWG2 Tools CWG3 Dataflow CWG4 Data Model CWG5 Computing Platforms CWG6 Calibration CWG9 QA, DQM CWG10 Control, Configuration CWG11 Software Lifecycle CWG8 Physics Simulation CWG7 Reconstruction CWG1 Architecture CWG2 Tools CWG3 Dataflow CWG4 Data Model CWG5 Computing Platforms CWG6 Calibration CWG9 QA, DQM CWG12 Computing Hardware CWG10 Control, Configuration CWG11 Software Lifecycle CWG8 Physics Simulation CWG7 Reconstruction CWG1 Architecture CWG2 Tools CWG3 Dataflow CWG4 Data Model

ALICE O2 Project - CHEP 2013 P. Vande Vyvre 23 Intensive period of R&D :  Collect the requirements: ITS and TPC TDRs  System modeling  Prototyping and benchmarking Technology and time are working with us  New options  Massive usage of commercial equipment very appealing Technical Design Report  Sep ’14: submission to the LHCC Next steps

Thanks !