Predrag Buncic CERN ALICE Status Report LHCC Referee Meeting September 22, 2015.

Slides:



Advertisements
Similar presentations
Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
Advertisements

Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
Batch Production and Monte Carlo + CDB work status Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
UNCLASSIFIED: LA-UR Data Infrastructure for Massive Scientific Visualization and Analysis James Ahrens & Christopher Mitchell Los Alamos National.
Cambodia-India Entrepreneurship Development Centre - : :.... :-:-
ALICE Operations short summary and directions in 2012 Grid Deployment Board March 21, 2011.
ALICE Operations short summary LHCC Referees meeting June 12, 2012.
ALICE Operations short summary and directions in 2012 WLCG workshop May 19-20, 2012.
1 Status of the ALICE CERN Analysis Facility Marco MEONI – CERN/ALICE Jan Fiete GROSSE-OETRINGHAUS - CERN /ALICE CHEP Prague.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
Status of CMS Matthew Nguyen Recontres LCG-France December 1 st, 2014 *Mostly based on information from CMS Offline & Computing Week November 3-7.
The ALICE Analysis Framework A.Gheata for ALICE Offline Collaboration 11/3/2008 ACAT'081A.Gheata – ALICE Analysis Framework.
Claudio Grandi INFN Bologna CMS Operations Update Ian Fisk, Claudio Grandi 1.
Offline Coordinators  CMSSW_7_1_0 release: 17 June 2014  Usage:  Generation and Simulation samples for run 2 startup  Limited digitization and reconstruction.
Your university or experiment logo here Caitriana Nicholson University of Glasgow Dynamic Data Replication in LCG 2008.
Status of 2015 pledges 2016 requests RRB Report Concezio Bozzi INFN Ferrara LHCb NCB, November 3 rd 2014.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
1. Maria Girone, CERN  Q WLCG Resource Utilization  Commissioning the HLT for data reprocessing and MC production  Preparing for Run II  Data.
Continuous Integration and Code Review: how IT can help Alex Lossent – IT/PES – Version Control Systems 29-Sep st Forum1.
ATLAS in LHCC report from ATLAS –ATLAS Distributed Computing has been working at large scale Thanks to great efforts from shifters.
Grid Computing at Yahoo! Sameer Paranjpye Mahadev Konar Yahoo!
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
ALICE – networking LHCONE workshop 10/02/ Quick plans: Run 2 data taking Both for Pb+Pb and p+p – Reach 1 nb -1 integrated luminosity for rare.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
Jean-Roch Vlimant, CERN Physics Performance and Dataset Project Physics Data & MC Validation Group McM : The Evolution of PREP. The CMS tool for Monte-Carlo.
2012 RESOURCES UTILIZATION REPORT AND COMPUTING RESOURCES REQUIREMENTS September 24, 2012.
Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015.
LHCbComputing LHCC status report. Operations June 2014 to September m Running jobs by activity o Montecarlo simulation continues as main activity.
GridKa Summer 2010 T. Kress, G.Quast, A. Scheurer Migration of data from old to new dCache instance finished on Nov. 23 rd almost 500'000 files (600.
Predrag Buncic ALICE Status Report LHCC Referee Meeting CERN
ATLAS Distributed Computing perspectives for Run-2 Simone Campana CERN-IT/SDC on behalf of ADC.
Predrag Buncic CERN Future of the Offline. Data Preparation Group.
Predrag Buncic CERN ALICE Status Report LHCC Referee Meeting 01/12/2015.
Predrag Buncic (CERN/PH-SFT) Software Packaging: Can Virtualization help?
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
LHCbComputing Computing for the LHCb Upgrade. 2 LHCb Upgrade: goal and timescale m LHCb upgrade will be operational after LS2 (~2020) m Increase significantly.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
ALICE Grid operations +some specific for T2s US-ALICE Grid operations review 7 March 2014 Latchezar Betev 1.
ALICE Run 2 Readiness WLCG Collaboration Workshop Okinawa Apr 11, 2015 Maarten Litmaath CERN v1.2 1.
16 September 2014 Ian Bird; SPC1. General ALICE and LHCb detector upgrades during LS2  Plans for changing computing strategies more advanced CMS and.
Next Generation of Apache Hadoop MapReduce Owen
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
1 June 11/Ian Fisk CMS Model and the Network Ian Fisk.
Domenico Elia1 ALICE computing: status and perspectives Domenico Elia, INFN Bari Workshop CCR INFN / LNS Catania, Workshop Commissione Calcolo.
Some topics for discussion 31/03/2016 P. Hristov 1.
Advancing CernVM-FS and its Development Infrastructure José Molina Colmenero CERN EP-SFT.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
MAUS Status A. Dobbs CM43 29 th October Contents MAUS Overview Infrastructure Geometry and CDB Detector Updates CKOV EMR KL TOF Tracker Global Tracking.
Predrag Buncic CERN Data management in Run3. Roles of Tiers in Run 3 Predrag Buncic 2 ALICEALICE ALICE Offline Week, 01/04/2016 Reconstruction Calibration.
Predrag Buncic CERN Plans for Run2 and the ALICE upgrade in Run3 ALICE Tier-1/Tier-2 Workshop February 2015.
LHCb LHCb GRID SOLUTION TM Recent and planned changes to the LHCb computing model Marco Cattaneo, Philippe Charpentier, Peter Clarke, Stefan Roiser.
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
Predrag Buncic CERN ALICE Status Report LHCC Referee Meeting 24/05/2015.
Tom Van Steenkiste Supervisor: Predrag Buncic
T1/T2 workshop – 7th edition - Strasbourg 3 May 2017 Latchezar Betev
Predrag Buncic ALICE Status Report LHCC Referee Meeting CERN
evoluzione modello per Run3 LHC
Workshop Computing Models status and perspectives
ALICE analysis preservation
LHCb Software & Computing Status
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Simulation use cases for T2 in ALICE
ALICE Computing Model in Run3
ALICE Computing Upgrade Predrag Buncic
QA tools – introduction and summary of activities
Module 01 ETICS Overview ETICS Online Tutorials
Presentation transcript:

Predrag Buncic CERN ALICE Status Report LHCC Referee Meeting September 22, 2015

Software consolidation – Improved barrel tracking at high p T – Development and testing of code for new detectors – Validation of G4 Re-processing of RAW data from RUN1 – p-p and p-A data recalibrated – All processing with the same software – General-purpose and special MC completed ALICE during LS1 Detector upgrades – TPC, TRD readout electronics consolidation – TRD full azimuthal coverage – +1 PHOS calorimeter module – New DCAL calorimeter 2

ALICE Grid keeps growing 60 in Europe 11 in Asia 2 in Africa 1 operational 1 future 2 in South America 1 operational 1 future 8 in North America 7 operational 1 future ORNL – US UNAM – Mexico RRC-KI (T1)- Russia Wuhan – China Bandung, Cibinong – Indonesia WUT - Poland

Grid utilization: end of RUN1 - today Reaching new highs 96K parallel jobs Consistent and very good efficiency in all computing centres

ALICE Resources usage in 2015 (CPU)

ALICE Resources usage in 2015 (DISK)

Data popularity Regular storage cleanup less than 5% of unpopular data on disk

Run2 progress Steady data taking – Fast calibration and reconstruction – Quasi-online data processing Without backlog Accumulation of raw data in Run 2

Resource requirements CPU (kHEPSPEC06) Tier0CAFTier1sTier2s Disk (PB) Tier0 1) CAFTier1s 2) Tier2s , ) Excluding the 3.0 PB of disk buffer in front of the taping system 2) Excluding the 2.35 PB of disk buffer in front of the taping system

Beyond Run2 – the O 2 project Technical Design Report for the Upgrade of the Online-Offline Computing System -Submitted to LHCC -Framework and code development is well under way

Rebalancing disk/CPU ratio on T2s 11 Expecting Grid resources (CPU, storage) to grow at 20% per year rate Large number of disk will be used by Run 1 and Run 2 data Since T2s will be used almost exclusively for simulation jobs (no input) and resulting AODs will be exported to T1s/AFs, we expect to significantly lower needs for storage on T2s CPU (cores) Disk (PB)

Combined long term disk space requirement (T0+T1s+T2s+O2) 12 CRSG Scrutinized request PB

First O2 milestones

Dynamical Deployment System (DSS) 18/09/1514 Connecting the FairMQ devices/tasks requires knowledge of connection parameters that need to be propagated startup time - the time which took DDS to distribute user tasks, propagate properties and enter into RUN state DDS Topology editor An interactive web tool that allows: creation, modification and visualization of a DDS topology Aims to provide the DDS user with an intuitive platform for creating the desired topology.

Parameter management based on Riak key-value store 18/09/1515 high availability scalability fault tolerance configurable Two storage back-ends were tested: Bitcask best latency nodes out-of-memory LevelDb similar performance compressed storage Message-Queue based concept make it possible to use directly the native Java client of Riak

16 Agile lifecycle: from code to published builds quickly, with automated tests Philosophy: automatize as much as possible, trivialize human intervention Objective for O²: refreshing aging infrastructure and paradigm Setup the infrastructure for running builds and tests: done Basic tools to build+publish current software and O²: done Visualization interface for the tests: Nov 2015 Automatic test and flagging of Pull Requests for O²: Nov 2015 Basic infrastructure is ready way ahead of time: as of Sep 9, old system retired and already using it in production for current software Develop Build and test Deploy and publish Git-basedIntegration buildsCVMFS, RPM, Ubuntu source code tarballs test results Software Lifecycle

17 Quick development of the new infrastructure possible thanks to use of common OpenSource tools supported CERN IT Source code on GitHub and CERN GitLab Builds are scheduled by Jenkins Full infrastructure running on CERN OpenStack virtual machines We can build on multiple platforms seamlessly by using Docker containers Containers are deployed by Apache Mesos Software distribution by CVMFS Visualization of performance test trends via Kibana Software Lifecycle

Summary The main ALICE goals in LS1 in terms of computing and software were achieved – Software quality improved and the entire Run1 dataset was re- processed with same software version The reprocessing was helped by an effective use of computing resources – Often going beyond the pledges for CPU Computing resource requests for Run 2 remain unchanged – Must keep up in order to match start of Run3 O2 will absorb the impact on resources at the start of Run3 Work on O2 has started following the plan presented in TDR – First milestones on the horizon

Backup

O2 Milestones