O2 Project – Phase 2 Predrag Buncic

Slides:



Advertisements
Similar presentations
Enabling Access to Sound Archives through Integration, Enrichment and Retrieval WP1. Project Management.
Advertisements

DataGrid is a project funded by the European Union 22 September 2003 – n° 1 EDG WP4 Fabric Management: Fabric Monitoring and Fault Tolerance
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
8.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
ALFA: The new ALICE-FAIR software framework
Status and roadmap of the AlFa Framework Mohammad Al-Turany GSI-IT/CERN-PH-AIP.
Chapter 9: Moving to Design
Trigger and online software Simon George & Reiner Hauser T/DAQ Phase 1 IDR.
Quality Control B. von Haller 8th June 2015 CERN.
Chapter 9 Elements of Systems Design
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
FLP data flow – FLP prototype Filippo Costa ALICE O 2 9 th June 2015 WUHAN, CHINA.
REVIEW OF NA61 SOFTWRE UPGRADE PROPOSAL. Mandate The NA61 experiment is contemplating to rewrite its fortran software in modern technology and are requesting.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
ALICE O 2 Plenary | October 1 st, 2014 | Pierre Vande Vyvre O2 Project Status P. Buncic, T. Kollegger, Pierre Vande Vyvre 1.
1 Alice DAQ Configuration DB
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Predrag Buncic CERN ALICE Status Report LHCC Referee Meeting September 22, 2015.
LHC Computing Review Recommendations John Harvey CERN/EP March 28 th, th LHCb Software Week.
1 Planning for Reuse (based on some ideas currently being discussed in LHCb ) m Obstacles to reuse m Process for reuse m Project organisation for reuse.
February 17, 2015 Software Framework Development P. Hristov for CWG13.
9 Systems Analysis and Design in a Changing World, Fourth Edition.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
ATLAS Data Challenges US ATLAS Physics & Computing ANL October 30th 2001 Gilbert Poulard CERN EP-ATC.
9 Systems Analysis and Design in a Changing World, Fourth Edition.
11 CLUSTERING AND AVAILABILITY Chapter 11. Chapter 11: CLUSTERING AND AVAILABILITY2 OVERVIEW  Describe the clustering capabilities of Microsoft Windows.
Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015.
1 The ILC Control Work Packages. ILC Control System Work Packages GDE Oct Who We Are Collaboration loosely formed at Snowmass which included SLAC,
O 2 Project Roadmap P. VANDE VYVRE 1. O2 Project: What’s Next ? 2 O2 Plenary | 11 March 2015 | P. Vande Vyvre TDR close to its final state and its submission.
Andrea Valassi (CERN IT-DB)CHEP 2004 Poster Session (Thursday, 30 September 2004) 1 HARP DATA AND SOFTWARE MIGRATION FROM TO ORACLE Authors: A.Valassi,
Predrag Buncic ALICE Status Report LHCC Referee Meeting CERN
Predrag Buncic CERN Future of the Offline. Data Preparation Group.
The MEG Offline Project General Architecture Offline Organization Responsibilities Milestones PSI 2/7/2004Corrado Gatto INFN.
CWG13: Ideas and discussion about the online part of the prototype P. Hristov, 11/04/2014.
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
ALICE O 2 | 2015 | Pierre Vande Vyvre O 2 Project Pierre VANDE VYVRE.
Follow-up to SFT Review (2009/2010) Priorities and Organization for 2011 and 2012.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Current activities and short term plans 24/04/2015 P. Hristov.
Online Software November 10, 2009 Infrastructure Overview Luciano Orsini, Roland Moser Invited Talk at SuperB ETD-Online Status Review.
1 Presented by: Val Pennell, Test Tool Manager Date: March 9, 2004 Software Testing Tools – Load Testing.
9 Systems Analysis and Design in a Changing World, Fifth Edition.
Barthélémy von Haller CERN PH/AID For the ALICE Collaboration The ALICE data quality monitoring system.
Summary of IAPP scientific activities into 4 years P. Giannetti INFN of Pisa.
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
QC-specific database(s) vs aggregated data database(s) Outline
WP18, High-speed data recording Krzysztof Wrona, European XFEL
BA Continuum India Pvt Ltd
CMS High Level Trigger Configuration Management
Chapter 6: Database Project Management
LHC experiments Requirements and Concepts ALICE
ALICE analysis preservation
The ILC Control Work Packages
Jacek Otwinowski (for the DPG QA tools and WP7 groups)
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
EIN 6133 Enterprise Engineering
JCOP Review Closeout March 13, 2003 Chip Watson, Robin Lauckner,
ALICE Computing Upgrade Predrag Buncic
PROCESS - H2020 Project Work Package WP6 JRA3
Copyright © 2011 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 2 Database System Concepts and Architecture.
SQL Server BI on Windows Azure Virtual Machines
US ATLAS Physics & Computing
Module 01 ETICS Overview ETICS Online Tutorials
Wide Area Workload Management Work Package DATAGRID project
ATLAS DC2 & Continuous production
Development of LHCb Computing Model F Harris
Presentation transcript:

O2 Project – Phase 2 Predrag Buncic

Timeline 2020 2019 2018 2017 2015 2016 2014 Commissioning CWG structure Work on TDR started Test Time Frame Reconstruction and simulation for ITS and TPC Simulation challenge Commissioning 2019 10 % data challenge O2 TDR accepted by the LHCC. Work started in CWG structure First milestones achieved and delayed

Time to draw the line… R&D phase of the project The goal was to write TDR and test possible solutions. Learn how to work together in CWG structure. Shared responsibility (by construction). Decisions by consensus in plenary meeting. Implementation phase of the project The goal is to deliver well defined system components (hardware and software) following the strict schedule. Work will be organized in Work Packages. Institutional responsibility for the WP. Decisions and milestones follow-up by the O2 Technical Board.

O2 Project Management Monthly O2 plenary meetings Decisions by consensus Weekly O2 Steering Board meeting Following project progress Bi-monthly O2 plenary meetings Information exchange and technical presentations. Bi-monthly O2 Technical Board meeting TB meeting schedule interleaved with plenaries. Attended by the WP leaders and SB members. Progress and milestone follow-up, ensuring coherence of WPs, preparation of decisions for confirmation by O2 Steering Board. O2 Steering Board Weekly project follow-up, long term planning meetings twice per year.

Work Packages http://goo.gl/QZUE1d

WP1 – Data Model Deliverables Participating institutes: FIAS, CERN Description of ALICE data model in terms of human and machine readable format that is maintainable over time. Includes all data structures that are exchanged between hardware and software devices or stored to a file. Development of AOD data model that is specifically adapted for efficient use with ALFA/FairMQ transport. Development of software needed to support 1) and 2). Participating institutes: FIAS, CERN WP Leader: Mikolaj Krzewicki

WP2 – Data Flow and System Simulation Deliverables Develop a tool for detailed simulation of network topologies for O2 facility allowing for quick evaluation of various scenarios. Develop a tool for coarse simulation of O2 system components allowing for quick estimate of computing resource needs. Extend 2) to estimate overall long term computing needs during Run 3 and 4 considering O2 and Grid resources. Participating institutes: CERN, Split WP Leader: Iosif Legrand

WP3 – Common tools and infrastructure Deliverables Develop and support common software tools used by all WPs such as: software build and distribution tool; continuous integration and release testing tools. Implement and document software lifecycle procedures and protocols. Provide general user support and in particular 1) and 2). Participating institutes: CERN WP Leader: Dario Berzano

WP3 – Common tools and infrastructure Deliverables Develop software components that are unique to ALICE and do not belong to a common ALFA layer. Integration with CCDB (Parameter Manager). Software provided by the other WPs and the detector groups integrated into common software framework. Common framework for unit and functional tests. Monitor framework performance and proactively work on software optimization. Participating institutes: CERN, GSI WP Leader: Giulio Eulisse

WP4 – O2 Software Framework Deliverables Develop software components that are unique to ALICE and do not belong to a common ALFA layer. Integration with CCDB (Parameter Manager). Software provided by the other WPs and the detector groups integrated into common software framework. Common framework for unit and functional tests. Monitor framework performance and proactively work on software optimization. Participating institutes: CERN, GSI WP Leader: Giulio Eulisse

WP5 – Data distribution and load balancing Deliverables Sub-time frame building on FLP nodes. Time frame building on EPN nodes. Load balancing of FLP->EPN traffic. Interface to Common Trigger Processor (CTP). Participating institutes: FIAS, CERN WP Leader: Gvozden Neskovic

WP6 – Detector readout Deliverables GBT read-out with the G-RORC (firmware and software) as a temporary solution before the RCU is available. DDL2  read-out with the C-RORC (firmware and software) for the detector who will use the DDL2 after LS2. GBT read-out with the CRU (firmware and software). DCS interface with the CRU (firmware and software). Participating institutes: CERN, Budapest WP Leader: Filippo Costa

WP7 – Quality Control Deliverables QC framework including basic QC client, API, repository, checkers and benchmark. Data Sampling and merging infrastructure allowing correlation and trending supporting alarms and notifications. Assist software development by detectors teams including infrastructure and procedures to validate detectors software Participating institutes: CERN, CERN, Warsaw, Krakow, Jammu, INFN, COMSATS WP Leader: Barthelemy Von Haller

WP8 – Control, Configuration and Monitoring Deliverables Control Resource allocation and scheduling. A set of control agents that represent and control system components such as (FLP and EPN nodes, detectors and partitions). Bookkeeping and GUI. Monitoring. Deploy end-to-end monitoring system that collects and presents monitoring information. Configuration Deploy end-to-end configuration system that stores and distributes process parameters Participating institutes: CERN, Bari, IFI Frankfurt, KMUTT Bangkok, NPI Prague WP Leader: Vasco Chibante

WP9 – Event Display Deliverables Adapt and improve the current ALICE event display for use within ALICE O2 software framework while maintaining the end extending the existing functionality Participating institutes: WUT

WP10 – CCDB Deliverables Adapt the current OCDB to use cases and constraints of O2 building on a proven architecture: database for metadata queries; blob/object store for data; time and run based access keys (for backward compatibility). API for CCDB access from the O2 framework Extend/Adapt FairRoot Parameter Manager for this purpose. Import existing OCDB data into new format and provide a service for database access and performance evaluation. Participating institutes: CERN, GSI WP Leader: Costin Grigoras

WP10 – CCDB Deliverables Common data transport layer based on FairMQ with 0MQ, nanomsg interfaces and low level bindings to the network fabric using libfabric abstraction. Performance evaluation of FairMQ using various network technologies and topologies. Distributed Deployment System (DDS) for deployment of process graphs (topologies). Common parameter manager as CCDB interface. Performance evaluations of different message passing serialization protocols. Participating institutes: GSI, CERN WP Leader: Mohammad Al-Turany

WP12 – Simulation Deliverables Up-to-date description of geometry in O2 framework and materials for all detectors. Detector response (digitization) software for all detectors. Simulation output compatible with timeframe format. Support for simulation QA, navigation and comparison between reconstruction and simulation. Simulation performance monitoring and improvement. Integration with new transports and geometry models Geant V, VecGeom. Support for fast and parameterized and fast simulation. Unified interfaces to modern MC generators. Participating institutes: CERN (+detector projects) WP Leader: Sandro Wenzel

WP13 – Reconstruction and Calibration Deliverables Develop reconstruction and calibration components of the O2 software framework. Coordinate development and integration of detector specific reconstruction and calibration algorithms. Assure that overall data compression factor following the synchronous reconstruction meets TDR requirements. Monitor computing and physics performance of the reconstruction and calibration code and keeping it compatible with the available computing capacity. Participating institutes: CERN (+detector projects) WP Leader: Ruben Shahoyan l

WP14 – Analysis framework and facilities Deliverables Refactoring of the current analysis framework and OADB to comply with O2 architecture covering use cases of filtering, skimming, augmentation, user specific formats, bookkeeping. Converters from the old to new AOD format. Analysis Facility infrastructure (software and hardware) capable of processing multi-PB datasets within 12h and providing: high performance cluster file system (1PB+); low latency and high throughput network; software distribution compatible with O2 needs; micro-services compatible scheduler. Support for organized analysis running on the AF infrastructure. Participating institutes: CERN, GSI WP Leader: Peter Hristovl

WP15 – Data Management Deliverables Identify or develop a solution for data management that spans the O2 facility and Grid scale: 60 PB in O2; 500 PB across the participating Grid tiers. Identify or provide a uniform data access method that adheres to AAA procedures compatible with Grid requirements Provide a horizontally scalable storage service without a single point of failure that meets the requirements of O2 facility in terms of data volume throughput and I/O ops/s. Provide a solution to access or migrate legacy data and metadata to a new system. Participating institutes: CERN WP Leader: Latchezar Betev

WP16 – Computing Room CR1 (FLP) Deliverables Installation of the infrastructure (power, fibres). Installation and configuration of the FLPs and the CRUs. Installation and configuration of the computing services. Installation and configuration of the control and data networks. Installation and configuration of the system and application software. Operation and support of the CR1 farm. Participating institutes: CERN, FIAS WP Leader: Ulrich Fuchs

WP17 – Computing Room CR0 (EPN) Deliverables Installation/validation of the infrastructure. Installation and configuration of the EPN nodes. Installation and configuration of the computing services. Installation and configuration of the control and data networks. Installation and configuration of the system and application software. Operation and support of the CR0 farm. Participating institutes: FIAS, CERN WP Leader: Johannes Lehrbach

WP Summary table WP Title Leading Institute WP Leader 1 Data Model FIAS Mikolaj Krzewicki 2 Data Flow and System Simulation CERN Iosif Legrand 3 Common tools and infrastructure Dario Berzano 4 O2 Software Framework Giulio Eulisse 5 Data distribution and load balancing Gvozden Neskovic 6 Detector readout Filippo Costa 7 Quality Control Barthelemy Von Haller 8 Control, Configuration and Monitoring Vasco Chibante 9 Event Display WUT t.b.c. 10 CCDB Costing Grigoras 11 ALFA GSI Mohammad Al-Turany 12 Simulation Sandro Wenzel 13 Reconstruction and Calibration Ruben Shahoyan 14 Analysis framework and facilities Peter Hristov 15 Data Management Latchezar Betev 16 Computing Room CR1 (FLP) Ulrich Fuchs 17 Computing Room CR0 (EPN) Johannes Lehrbach

Still working together! By re-aligning the responsibilities for the WPs with projects where predominant competence exist we do not want to abandon the O2 spirit. On the contrary, we must continue working together closer that before for this project to have any chance to succeed!