CERN - IT Department CH-1211 Genève 23 Switzerland www.cern.ch/i t Persistency Framework CORAL, POOL, COOL status and plans Andrea Valassi (IT-PSS) On.

Slides:



Advertisements
Similar presentations
Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
Advertisements

CERN - IT Department CH-1211 Genève 23 Switzerland t CORAL Server A middle tier for accessing relational database servers from CORAL applications.
CERN - IT Department CH-1211 Genève 23 Switzerland t LCG Persistency Framework CORAL, POOL, COOL – Status and Outlook A. Valassi, R. Basset,
CORAL and COOL news for ATLAS (update since March 2012 ATLAS sw workshop)March 2012 Andrea Valassi (IT-SDC) ATLAS Database Meeting 24 th October 2013.
CERN IT Department CH-1211 Genève 23 Switzerland t ES Discussion of COOL - CORAL - POOL priorities for ATLAS Andrea Valassi (IT-ES) For the.
1 Databases in ALICE L.Betev LCG Database Deployment and Persistency Workshop Geneva, October 17, 2005.
D. Düllmann - IT/DB LCG - POOL Project1 POOL Release Plan for 2003 Dirk Düllmann LCG Application Area Meeting, 5 th March 2003.
CERN - IT Department CH-1211 Genève 23 Switzerland t Relational Databases for the LHC Computing Grid The LCG Distributed Database Deployment.
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL March 25, 2003 CHEP 2003 Data Analysis Environment and Visualization.
Components and Architecture CS 543 – Data Warehousing.
CERN - IT Department CH-1211 Genève 23 Switzerland t Partitioning in COOL Andrea Valassi (CERN IT-DM) R. Basset (CERN IT-DM) Distributed.
Andrea Valassi (CERN IT-SDC) DPHEP Full Costs of Curation Workshop CERN, 13 th January 2014 The Objectivity migration (and some more recent experience.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
CERN - IT Department CH-1211 Genève 23 Switzerland t The High Performance Archiver for the LHC Experiments Manuel Gonzalez Berges CERN, Geneva.
CERN IT Department CH-1211 Genève 23 Switzerland t SDC Stabilizing SQL execution plans in COOL using Oracle hints Andrea Valassi (IT-SDC)
Conditions DB in LHCb LCG Conditions DB Workshop 8-9 December 2003 P. Mato / CERN.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES PhEDEx Monitoring Nicolò Magini CERN IT-ES-VOS For the PhEDEx.
LHC: ATLAS Experiment meeting “Conditions” data challenge Elizabeth Gallas - Oxford - August 29, 2009 XLDB3.
Software Solutions for Variable ATLAS Detector Description J. Boudreau, V. Tsulaia University of Pittsburgh R. Hawkings, A. Valassi CERN A. Schaffer LAL,
Workshop Summary (my impressions at least) Dirk Duellmann, CERN IT LCG Database Deployment & Persistency Workshop.
CHEP 2006, Mumbai13-Feb-2006 LCG Conditions Database Project COOL Development and Deployment: Status and Plans Andrea Valassi On behalf of the COOL.
Databases E. Leonardi, P. Valente. Conditions DB Conditions=Dynamic parameters non-event time-varying Conditions database (CondDB) General definition:
CERN IT Department CH-1211 Genève 23 Switzerland t ES Future plans for CORAL and COOL Andrea Valassi (IT-ES) For the Persistency Framework.
ALICE, ATLAS, CMS & LHCb joint workshop on
ATLAS Detector Description Database Vakho Tsulaia University of Pittsburgh 3D workshop, CERN 14-Dec-2004.
CERN Physics Database Services and Plans Maria Girone, CERN-IT
Database Monitoring Requirements Salvatore Di Guida (CERN) On behalf of the CMS DB group.
The Persistency Patterns of Time Evolving Conditions for ATLAS and LCG António Amorim CFNUL- FCUL - Universidade de Lisboa A. António, Dinis.
3rd November Richard Hawkings Luminosity, detector status and trigger - conditions database and meta-data issues  How we might apply the conditions.
CERN - IT Department CH-1211 Genève 23 Switzerland t COOL Conditions Database for the LHC Experiments Development and Deployment Status Andrea.
Peter Chochula ALICE Offline Week, October 04,2005 External access to the ALICE DCS archives.
CERN - IT Department CH-1211 Genève 23 Switzerland t CORAL COmmon Relational Abstraction Layer Radovan Chytracek, Ioannis Papadopoulos (CERN.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Database authentication in CORAL and COOL Database authentication in CORAL and COOL Giacomo Govi Giacomo Govi CERN IT/PSS CERN IT/PSS On behalf of the.
A. Valassi – Python bindings for C++ in PyCool ROOT Workshop – 16 th Sep Python bindings for C++ via PyRoot User experience from PyCool in COOL.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
The ATLAS TAGs Database - Experiences and further developments Elisabeth Vinek, CERN & University of Vienna on behalf of the TAGs developers group.
CERN IT Department CH-1211 Genève 23 Switzerland t COOL performance optimization using Oracle hints Andrea Valassi and Romain Basset (IT-DM)
Andrea Valassi (CERN IT-DB)CHEP 2004 Poster Session (Thursday, 30 September 2004) 1 HARP DATA AND SOFTWARE MIGRATION FROM TO ORACLE Authors: A.Valassi,
Overview of C/C++ DB APIs Dirk Düllmann, IT-ADC Database Workshop for LHC developers 27 January, 2005.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Overview of DMLite Ricardo Rocha ( on behalf of the LCGDM team.
11th November Richard Hawkings Richard Hawkings (CERN) ATLAS reconstruction jobs & conditions DB access  Conditions database basic concepts  Types.
CERN IT Department CH-1211 Genève 23 Switzerland t COOL Performance Tests ATLAS Conditions Database example Romain Basset, IT-DM October.
CORAL CORAL a software system for vendor-neutral access to relational databases Ioannis Papadopoulos, Radoval Chytracek, Dirk Düllmann, Giacomo Govi, Yulia.
CERN - IT Department CH-1211 Genève 23 Switzerland t Operating systems and Information Services OIS Proposed Drupal Service Definition IT-OIS.
The DCS Databases Peter Chochula. 31/05/2005Peter Chochula 2 Outline PVSS basics (boring topic but useful if one wants to understand the DCS data flow)
Summary of persistence discussions with LHCb and LCG/IT POOL team David Malon Argonne National Laboratory Joint ATLAS, LHCb, LCG/IT meeting.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Status of tests in the LCG 3D database testbed Eva Dafonte Pérez LCG Database Deployment and Persistency Workshop.
David Adams ATLAS ATLAS Distributed Analysis (ADA) David Adams BNL December 5, 2003 ATLAS software workshop CERN.
Conditions Database Status and Plans for 2005 Andrea Valassi (CERN IT-ADC) LCG Applications Area Review 31 March 2005.
M.Frank, CERN/LHCb Persistency Workshop, Dec, 2004 Distributed Databases in LHCb  Main databases in LHCb Online / Offline and their clients  The cross.
Maria del Carmen Barandela Pazos CERN CHEP 2-7 Sep 2007 Victoria LHCb Online Interface to the Conditions Database.
ATLAS The ConditionDB is accessed by the offline reconstruction framework (ATHENA). COOLCOnditions Objects for LHC The interface is provided by COOL (COnditions.
2007 Workshop on INFN ComputingRimini, 7 May 2007 Conditions Database Service implications at Tier1 and Tier2 sites Andrea Valassi (CERN IT-PSS)
CERN IT Department CH-1211 Genève 23 Switzerland t Load testing & benchmarks on Oracle RAC Romain Basset – IT PSS DP.
CERN IT Department CH-1211 Genève 23 Switzerland t DPM status and plans David Smith CERN, IT-DM-SGT Pre-GDB, Grid Storage Services 11 November.
CERN - IT Department CH-1211 Genève 23 Switzerland t COOL – Status and Plans Andrea Valassi (CERN IT-PSS-DP) LCG Application Area Review,
INFN Tier1/Tier2 Cloud WorkshopCNAF, 22 November 2006 Conditions Database Services How to implement the local replicas at Tier1 and Tier2 sites Andrea.
Database Replication and Monitoring
(on behalf of the POOL team)
Andrea Valassi (IT-ES)
Database Readiness Workshop Intro & Goals
POOL persistency framework for LHC
Conditions Data access using FroNTier Squid cache Server
LHCb Conditions Database TEG Workshop 7 November 2011 Marco Clemencic
Presentation transcript:

CERN - IT Department CH-1211 Genève 23 Switzerland t Persistency Framework CORAL, POOL, COOL status and plans Andrea Valassi (IT-PSS) On behalf of the Persistency Framework team LHCC, 19th November 2007 Thanks to the experiments for their input for this talk! Physics Services Support

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 2 Outline Introduction Main achievements in 2007 PF usage in the experiments Plans for 2008 Conclusions

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 3 Persistency Framework Components CORAL –Abstraction of access to relational databases –Support for Oracle, MySQL, SQLite, FroNtier POOL –Technologically-neutral hybrid data storage Streaming of objects (e.g. to ROOT or RDBMS) Object metadata catalogs (e.g. in relational databases) COOL –Conditions data management Conditions object metadata (interval of validity, version) Conditions object data payload (user-defined attributes)

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 4 Component Interaction

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 5 Persistency Framework and AA Interaction with other projects –SPI (external libs, configuration, nightly builds) –SEAL (base libraries - will soon be replaced) –ROOT (object streaming in POOL; PyCool) –GRID middleware (LFC for CORAL authentication) –3D (relational data deployment and distribution) New in 2007 –New configuration and build system Move from SCRAM to CMT Nightly builds and QMTEST tests based on CMT –Support for new platforms SLC4 on 64-bit Linux (lxplus) MacOSX/Intel (no Oracle yet)

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 6 Main achievements in 2007 (1) CORAL –LFC-based authentication and authorization –Python bindings (PyCoral) and database copy tool PyCoral and LFC work in cooperation with RRCAT, India –Connection pooling –Improved thread safety (Atlas online requirement) –Support for stored procedures POOL –Major reimplementation of collections Using CORAL database connectivity and authentication

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 7 Main achievements in 2007 (2) COOL –Major API and schema changes in COOL 2.0 Portable data types (e.g. cool::Int64) for 64-bit platforms Tag locking functionality Channel metadata management Dynamic replication tool Schema evolution tool and tests Improved DB authentication and replica lookup –Performance optimizations in COOL 2.1 and 2.2 Multi-channel bulk insertion and several query use cases –COOL is now deployed at T0 and several T1 sites (with Streams distribution – see 3D presentation)

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 8 LHCb feedback and PF usage POOL –For event data (SIMU, DIGI, DST, tags…) ROOT backend and XML catalog But RAW data will be in flat files (no POOL or ROOT) –LHCb requests to remove the SEAL dependency LHCb does not request any new functionality in POOL COOL (and CORAL via COOL) –For conditions data (online and offline) Oracle at the pit, T0 and T1 (with Streams replication) –Evaluating LFC-based authentication in CORAL SQLite files for MC production

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 9 ATLAS feedback and PF usage (1) POOL –For event data (RDO, EST, AOD, tags…) ROOT backend Separation of transient/persistent definitions simplifies schema evolution and is used to improve performance New POOL collections were mainly developed by ATLAS CORAL (directly) –For the detector description (geometry database) Oracle master, SQLite (was MySQL) for data distribution –For the online configuration and trigger databases MySQL server and proxies Motivation for most of the CORAL developments in 2007

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 10 ATLAS feedback and PF usage (2) COOL (and CORAL via COOL) –For conditions data (extensive online/offline use) Oracle at online, T0 and T1 (with Streams replication) COOL replication to SQLite (MC) and MySQL (HLT) –Largest data volume is from DCS (>300 GB/year) ATLAS tools exist to transfer data from PVSS to COOL –ATLAS requests are regularly discussed at the weekly COOL meetings Functionalities (e.g. for channel and tag management) Performance (e.g. retrieval of tagged data) –Worries about limited manpower and time spent on debugging non-core platforms (Windows) and non-COOL issues (SEAL, CORAL…)

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 11 CMS feedback and PF usage POOL (and CORAL via POOL-ORA) –POOL-ORA is the basis of all conditions data modeling and storage in CMS Using the Oracle, SQLite and FroNtier backends Switch to streaming to BLOB columns in POOL-ORA to optimize performance (switch is transparent via Reflex) Work in progress with POOL team on schema evolution –Worries about continuity of development/support due to expected changes in the development team CORAL (directly) –To read conditions data from the Oracle online db (then written via POOL-ORA into the offline db)

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 12 Plans for next year CORAL –Move SEAL functionalities into CORAL Will then need to be picked up by POOL and COOL –CORAL proxy server development POOL –Schema evolution in POOL-ORA COOL –Further performance optimizations –Enhancements for channel and tag management

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 13 Move SEAL functionalities into CORAL Motivation –SEAL support and maintenance are not staffed –Several problems in multi-threading environments e.g. from 2 nd CORAL thread that closes stale connections Multi-threading outside the original SEAL design scope Main components to replace –Component model and dynamic loading of plugins

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 14 CORAL proxy server Motivation –Secure authentication and authorisation Authenticate using Grid certificates on the proxy in spite of missing database vendor support for X.509 certificates –Scalability for many connections Serve several (mostly idle) CORAL connections to the proxy by fewer active connections to the database Interests both users and service managers –IT physics database and security teams Better load management, hide DB ports behind firewall –ATLAS and CMS Possible addition of data caching functionality A MySQL-based proxy is currently used in ATLAS online

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 15 CORAL proxy server - a possible scenario

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 16 POOL-ORA schema evolution Deal with class definition changes –e.g. add or remove attributes of a class –e.g. change the type of an attribute of a class –e.g. move data member from/into a base class Deal with changes in the storage layout –e.g. move C-array data into inline columns Work is already in progress –Tools for users with schema modification privileges –Mainly for CMS (no persistent/transient separation)

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 17 COOL enhancements and optimizations Feature enhancements –Partial tag locks, easier use of channel names… –Most requests from ATLAS, some from LHCb too Performance optimizations –Main pending issue: retrieval of tagged IOVs –Many use cases mean many queries to optimize Similar problems (query time increases as table size increase) with similar solutions (query rewrite, indexes…) Consolidating code to factor out commonalities –Local scalability tests to try to prevent problems from showing up later at T0 and T1 sites

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 18 Manpower Small contributions from the experiments –ATLAS: POOL (0.5) and COOL (0.2) COOL contributions also from user community and DBAs –CMS: CORAL (0.1) – not counting Frontier team –LHCb: POOL (0.1) and COOL (0.2) Main contribution from CERN IT-PSS –POOL/CORAL (3.5) and COOL (0.8) Three key CORAL developers from IT-PSS will leave in 2008 –Transfer expertise to new hires (one, possibly two) –Reallocate tasks to smaller PF team

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 19 Comments from the previous review Both remarks are still valid and relevant –Reduction and turnover in the CORAL/POOL team “Possible manpower crisis” foreseen in the 2006 report –SEAL replacement is in the PF work plan for 2008

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 20 Conclusions ATLAS, CMS and LHCb are relying on the PF for their event and/or conditions data Development plans for 2008 –Many items on the CORAL and COOL work plan –POOL is mostly in maintenance mode Overall manpower reduction –Low but relatively stable contributions for COOL –Team reduction and turnover for CORAL/POOL

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 21 Reserve slides

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 22 SV Single-channel browse IOVs Example: get all IOVs in t=[20,30] for channel 5 –SV (single version): there is only one version at any time t –SC (single channel): just select ChannelId=5 Problem until COOL included –Retrieval time is longer for IOVs at the end of the IOV table – ( Since<=20 AND 20<Until ) OR ( 20<Since AND Since<=30 ) For more details: “task #3675”“task #3675” IOV valid at t=20 (inefficient lookup – two columns) Fixed in COOL –Optimize lookup of first IOV As in fix for SV SC single-IOV find –Two separate SQL queries 1.MAX(Since) WHERE Since<20 2.Since = maxSince (from query1) New strategy in COOL –Merge two queries in a single SQL query (use subqueries) –Needed for SV MC case

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 23 SV Multi-channel browse IOVs Example: get all IOVs in t=[20,30] in channels 1-99 –SV (single version): there is only one version at any time t –MC (multi channel): ( 1<=ChannelId AND ChannelId<=99 ) Special case: all channels (no selection on ChannelId ) Problem until COOL included –Retrieval time is longer for IOVs at the end of the IOV table – ( Since<=20 AND 20<Until ) OR ( 20<Since AND Since<=30 ) –Same problem as for single-channel case in COOL Fixed in COOL –Optimize lookup of first IOV for each channel as in fix for SC case With Max(Since) subquery –Loop over selected channels via a join on the IOV and channel tables Execution plan (table order in join) depends on first value used (“bind variable peeking”): fix it using hints, /*+ LEADING(c i) USE_NL(c i) */ For more details: “task #4402”“task #4402” Was a showstopper for Atlas distributed readback tests at T1 sites in Q2 2007

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 24 Deployment in LHCb Computing model –Reconstruction at T0/T1 –Only MC prod at T2 COOL stores only conditions data for event reconstruction –Oracle at PIT, T0, T1 with replication via Streams –Geometry and conditions for MC sent to T2 as SQLite file Online db master at PIT –Replicated forward to T0 and T1 via Streams –Data from PVSS processes Offline db master at T0 –Replicated back to PIT and forward to T1 via Streams –Data computed in offline calibration/alignment jobs (Marco Clemencic, COOL meeting 3 July 2006) COOL

CERN - IT Department CH-1211 Genève 23 Switzerland t LHCC, 19th November 2007 A. Valassi – Persistency Framework – 25 Deployment in Atlas Largest COOL data set comes from DCS –Via the PVSS2COOL data transfer (1.5 GB/day) From the online RAC in the T0 computer centre –For offline reconstruction and detector experts Many options open for T2 replication –Many use cases (simulation, calibration, analysis) –Static/dynamic replication to sqlite/mysql, Frontier (Florbela Viegas, CHEP 2007) Persistency Framework – 25