Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Clara Gaspar on behalf of the LHCb Collaboration, “Physics at the LHC and Beyond”, Quy Nhon, Vietnam, August 2014 Challenges and lessons learnt LHCb Operations.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
ALFA: The new ALICE-FAIR software framework
ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen1 Offline Mass Data Processing using Online Computing Resources at HERA-B José Hernández DESY-Zeuthen.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
Quality Control B. von Haller 8th June 2015 CERN.
Tom Dietel University of Cape Town for the ALICE Collaboration Computing for ALICE at the LHC.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
1 Kittikul Kovitanggoon*, Burin Asavapibhop, Narumon Suwonjandee, Gurpreet Singh Chulalongkorn University, Thailand July 23, 2015 Workshop on e-Science.
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Predrag Buncic CERN ALICE Status Report LHCC Referee Meeting September 22, 2015.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015.
Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 1 Computing at the HL-LHC Predrag Buncic on behalf of the Trigger/DAQ/Offline/Computing Preparatory.
1 The PHENIX Experiment in the RHIC Run 7 Martin L. Purschke, Brookhaven National Laboratory for the PHENIX Collaboration RHIC from space Long Island,
HSF Workshop, SLAC | January 19, 2015 | Predrag Buncic HSF Workshop – ALICE input Predrag Buncic.
Pierre VANDE VYVRE for the O 2 project 15-Oct-2013 – CHEP – Amsterdam, Netherlands.
Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 1 Computing at the HL-LHC Predrag Buncic on behalf of the Trigger/DAQ/Offline/Computing Preparatory.
CBM Computing Model First Thoughts CBM Collaboration Meeting, Trogir, 9 October 2009 Volker Friese.
Data Acquisition, Trigger and Control
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
O 2 Project Roadmap P. VANDE VYVRE 1. O2 Project: What’s Next ? 2 O2 Plenary | 11 March 2015 | P. Vande Vyvre TDR close to its final state and its submission.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Workflows and Data Management. Workflow and DM Run3 and after: conditions m LHCb major upgrade is for Run3 (2020 horizon)! o Luminosity x 5 ( )
Predrag Buncic ALICE Status Report LHCC Referee Meeting CERN
ATLAS Distributed Computing perspectives for Run-2 Simone Campana CERN-IT/SDC on behalf of ADC.
Pierre VANDE VYVRE ALICE Online upgrade October 03, 2012 Offline Meeting, CERN.
CWG13: Ideas and discussion about the online part of the prototype P. Hristov, 11/04/2014.
Ian Bird WLCG Networking workshop CERN, 10 th February February 2014
Predrag Buncic (CERN/PH-SFT) Software Packaging: Can Virtualization help?
LHCbComputing Computing for the LHCb Upgrade. 2 LHCb Upgrade: goal and timescale m LHCb upgrade will be operational after LS2 (~2020) m Increase significantly.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
ALICE Software Evolution Predrag Buncic. GDB | September 11, 2013 | Predrag Buncic 2.
Monitoring for the ALICE O 2 Project 11 February 2016.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
ALICE O 2 project B. von Haller on behalf of the O 2 project CERN.
ALICE O 2 | 2015 | Pierre Vande Vyvre O 2 Project Pierre VANDE VYVRE.
Main parameters of Russian Tier2 for ATLAS (RuTier-2 model) Russia-CERN JWGC meeting A.Minaenko IHEP (Protvino)
Workshop ALICE Upgrade Overview Thorsten Kollegger for the ALICE Collaboration ALICE | Workshop |
16 September 2014 Ian Bird; SPC1. General ALICE and LHCb detector upgrades during LS2  Plans for changing computing strategies more advanced CMS and.
Big Data for Big Discoveries How the LHC looks for Needles by Burning Haystacks Alberto Di Meglio CERN openlab Head DOI: /zenodo.45449, CC-BY-SA,
Joint Institute for Nuclear Research Synthesis of the simulation and monitoring processes for the data storage and big data processing development in physical.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Domenico Elia1 ALICE computing: status and perspectives Domenico Elia, INFN Bari Workshop CCR INFN / LNS Catania, Workshop Commissione Calcolo.
CRISP WP18, High-speed data recording Krzysztof Wrona, European XFEL PSI, 18 March 2013.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
Computing infrastructures for the LHC: current status and challenges of the High Luminosity LHC future Worldwide LHC Computing Grid (WLCG): Distributed.
Predrag Buncic CERN Data management in Run3. Roles of Tiers in Run 3 Predrag Buncic 2 ALICEALICE ALICE Offline Week, 01/04/2016 Reconstruction Calibration.
Predrag Buncic CERN Plans for Run2 and the ALICE upgrade in Run3 ALICE Tier-1/Tier-2 Workshop February 2015.
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
LHC collisions rate: Hz New PHYSICS rate: Hz Event selection: 1 in 10,000,000,000,000 Signal/Noise: Raw Data volumes produced.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
Ian Bird WLCG Workshop San Francisco, 8th October 2016
WP18, High-speed data recording Krzysztof Wrona, European XFEL
Predrag Buncic ALICE Status Report LHCC Referee Meeting CERN
evoluzione modello per Run3 LHC
Workshop Computing Models status and perspectives
Data Challenge with the Grid in ATLAS
ALICE analysis preservation
Outline ALICE upgrade during LS2
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
ALICE Computing Model in Run3
ALICE Computing Upgrade Predrag Buncic
New strategies of the LHC experiments to meet
Computing at the HL-LHC
Presentation transcript:

Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015

Upgrade Roadmap Predrag Buncic 2 ALICEALICE Future IT Challenges for ALICE Run1 Run2 Run3 ALICE + LHCb Run4 ATLAS + CMS

The ALICE Experiment Predrag Buncic 3 ALICEALICE Event display from a real Pb-Pb collision. Each line is a particle traversing the ALICE apparatus, leaving a signal in the crossed detector Future IT Challenges for ALICE

Towards Higher Luminosities Predrag Buncic 4 ALICEALICE LHC proton-proton now: Luminosity 7 × cm –2 s –1 25 ns LHC proton-proton after 2018: Luminosity 4 × cm –2 s –1 Future IT Challenges for ALICE

Data Acquisition (DAQ) Design Concept Predrag Buncic 5 ALICEALICE Acquire data of tens of millions of channels Store them in a matrix of hundreds of memories Multiplex to a computer farm Assemble and store data from the same event MULTIPLEXER Memory Matrix Computer FarmComplete Events Future IT Challenges for ALICE

Paradigm shift: Continuous readout Predrag Buncic 6 ALICEALICE Continuous detector reading: replace events with time windows (20 ms,  events). Self sufficient small dataset (10 GB) Calibrate and reconstruction online: reduce data volume & structure the data MULTIPLEXER Memory Matrix Computer FarmComplete Events Future IT Challenges for ALICE

Run 3 Predrag Buncic 7 ALICEALICE Storage Online/Offline Facility 90 GB/s 1.1 TB/s x12 compression factor Future IT Challenges for ALICE

O2 Facility Predrag Buncic 8 ALICEALICE 60 PB 10^5 CPUs 5000 GPUs Future IT Challenges for ALICE

O2 Facility Predrag Buncic 9 ALICEALICE 60 PB 10^5 CPUs 5000 GPUs Future IT Challenges for ALICE

O2 Facility Predrag Buncic 10 ALICEALICE 60 PB 10^5 CPUs 5000 GPUs Future IT Challenges for ALICE

O2 Facility Predrag Buncic 11 ALICEALICE 60 PB 10^5 CPUs 5000 GPUs Future IT Challenges for ALICE

O2 Facility Predrag Buncic 12 ALICEALICE 60 PB 10^5 CPUs 5000 GPUs Future IT Challenges for ALICE

O2 Facility Predrag Buncic 13 ALICEALICE 60 PB 10^5 CPUs 5000 GPUs Future IT Challenges for ALICE

O2 Facility Predrag Buncic 14 ALICEALICE 60 PB 10^5 CPUs 5000 GPUs Future IT Challenges for ALICE

O2 Facility Predrag Buncic 15 ALICEALICE 60 PB 10^5 CPUs 5000 GPUs Future IT Challenges for ALICE

Roles of Tiers in Run 3 Predrag Buncic 16 ALICEALICE Reconstruction Calibration Analysis Reconstruction Calibration Archiving Analysis Simulation T0/T1 CTF -> ESD -> AOD AF AOD -> HISTO, TREE O 2 RAW -> CTF -> ESD -> AOD 1 T2/HPC MC -> CTF -> ESD -> AOD 1..n 1..3 CTF AOD Future IT Challenges for ALICE

O2 Software Framework Predrag Buncic 17 ALICEALICE Input File(s) T 5 Output File T 1 T 3 T 2 T 4 T 6 t0 time t1 ALFA - Common development between GSI/FAIR experiments and ALICE Distributed, multi process application Each process can be multi-threaded and possibly adapted to use hardware accelerators Message Queues for data exchange Future IT Challenges for ALICE

Detector readout at 1.1 TB/s and early data reduction in FLPs FPGA based data transfer, cluster finder, data compression FPGA programming, tools and training FLP to EPN data transfer at 500 GB/s Network modeling, simulation and cost optimization Lossy data compression on EPNs requiring precise quasi-online calibration and time frame based reconstruction Low latency database (key value store) for calibration Using ~ cores and 5000 GPUs for synchronous processing Management and s/w deployment/scheduling on heterogeneous clusters Use of lightweight virtualization (containers) CernVM like docker containers? Provisioning and maintaining 60 PB disk buffer to store compressed time frames and assure subsequent calibration/reconstruction iterations EOS or cluster file system? Summary of future IT challenges for ALICE Predrag Buncic 18 ALICEALICE Future IT Challenges for ALICE

Extending O2 style computing to Grid Optimizing scheduling and throughput of distributed application Modeling of the system in order to optimize cost and performance Using dedicated facilities or re-purposed T1/T2s for analysis HPC style big data processing, 5PB/day Software framework developments Performance on all levels Use of hardware accelerators Efficient streaming I/O Configuring and deploying distributed multi-process applications Simulation Developing fast simulation tools Use of opportunistic and HPC resources O2 Facility design and implementation 2.4 MW facility on limited budget Summary of future IT challenges for ALICE Predrag Buncic 19 ALICEALICE Future IT Challenges for ALICE