Persistency Project (POOL) Status and Work Plan Proposal

Slides:



Advertisements
Similar presentations
Object Persistency & Data Handling Session C - Summary Object Persistency & Data Handling Session C - Summary Dirk Duellmann.
Advertisements

Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
The POOL Persistency Framework POOL Summary and Plans.
1 Databases in ALICE L.Betev LCG Database Deployment and Persistency Workshop Geneva, October 17, 2005.
RLS Production Services Maria Girone PPARC-LCG, CERN LCG-POOL and IT-DB Physics Services 10 th GridPP Meeting, CERN, 3 rd June What is the RLS -
D. Düllmann - IT/DB LCG - POOL Project1 POOL Release Plan for 2003 Dirk Düllmann LCG Application Area Meeting, 5 th March 2003.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
LCG Persistency Workshop, June 5th 2002
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
RLS Tier-1 Deployment James Casey, PPARC-LCG Fellow, CERN 10 th GridPP Meeting, CERN, 3 rd June 2004.
EGEE is a project funded by the European Union under contract IST Testing processes Leanne Guy Testing activity manager JRA1 All hands meeting,
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL U.S. ATLAS Physics and Computing Advisory Panel Review Argonne National Laboratory Oct 30, 2001.
Workshop Summary (my impressions at least) Dirk Duellmann, CERN IT LCG Database Deployment & Persistency Workshop.
CHEP 2003 March 22-28, 2003 POOL Data Storage, Cache and Conversion Mechanism Motivation Data access Generic model Experience & Conclusions D.Düllmann,
CERN Physics Database Services and Plans Maria Girone, CERN-IT
3D Workshop Outline & Goals Dirk Düllmann, CERN IT More details at
POOL Status and Plans Dirk Düllmann IT-DB & LCG-POOL Application Area Meeting 10 th March 2004.
Feedback from the POOL Project User Feedback from the POOL Project Dirk Düllmann, LCG-POOL LCG Application Area Internal Review October 2003.
SEAL Project Core Libraries and Services 18 December 2002 P. Mato / CERN Shared Environment for Applications at LHC.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
The POOL Persistency Framework POOL Project Review Introduction & Overview Dirk Düllmann, IT-DB & LCG-POOL LCG Application Area Internal Review October.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
DGC Paris WP2 Summary of Discussions and Plans Peter Z. Kunszt And the WP2 team.
LCG Applications Area Meeting - 5 Feb 2003 ATLAS Strategy for the Data Dictionary & Persistency David R. Quarrie Lawrence Berkeley National Laboratory.
CHEP 2004, Core Software Integration of POOL into three Experiment Software Frameworks Giacomo Govi CERN IT-DB & LCG-POOL K. Karr, D. Malon, A. Vaniachine.
D. Duellmann - IT/DB LCG - POOL Project1 The LCG Pool Project and ROOT I/O Dirk Duellmann What is Pool? Component Breakdown Status and Plans.
Some Ideas for a Revised Requirement List Dirk Duellmann.
K. Harrison CERN, 3rd March 2004 GANGA CONTRIBUTIONS TO ADA RELEASE IN MAY - Outline of Ganga project - Python support for AJDL - LCG analysis service.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
Data Placement Intro Dirk Duellmann WLCG TEG Workshop Amsterdam 24. Jan 2012.
David Foster LCG Project 12-March-02 Fabric Automation The Challenge of LHC Scale Fabrics LHC Computing Grid Workshop David Foster 12 th March 2002.
D. Duellmann - IT/DB LCG - POOL Project1 The LCG Dictionary and POOL Dirk Duellmann.
Summary of persistence discussions with LHCb and LCG/IT POOL team David Malon Argonne National Laboratory Joint ATLAS, LHCb, LCG/IT meeting.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
G.Govi CERN/IT-DB 1GridPP7 June30 - July 2, 2003 Data Storage with the POOL persistency framework Motivation Strategy Storage model Storage operation Summary.
ROOT Based CMS Framework Bill Tanenbaum US-CMS/Fermilab 14/October/2002.
Site Services and Policies Summary Dirk Düllmann, CERN IT More details at
EJB Enterprise Java Beans JAVA Enterprise Edition
POOL & ARDA / EGEE POOL Plans for 2004 ARDA / EGEE integration Dirk Düllmann, IT-DB & LCG-POOL LCG workshop, 24 March 2004.
D. Duellmann - IT/DB LCG - POOL Project1 Internal Pool Release V0.2 Dirk Duellmann.
D. Duellmann, IT-DB POOL Status1 POOL Persistency Framework - Status after a first year of development Dirk Düllmann, IT-DB.
D. Düllmann - IT/DB LCG - POOL Project1 POOL Release V1.0 Dirk Düllmann LCG Application Area Meeting, 14 th May 2003.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
A quick summary and some ideas for the 2005 work plan Dirk Düllmann, CERN IT More details at
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 6 April 2005.
POOL Based CMS Framework Bill Tanenbaum US-CMS/Fermilab 04/June/2003.
LCG Application Area Internal Review Persistency framework summary and input for phase-2 planning Dirk Duellmann, CERN IT
Jean-Philippe Baud, IT-GD, CERN November 2007
LCG Applications Area Milestones
EGEE Middleware Activities Overview
(on behalf of the POOL team)
U.S. ATLAS Grid Production Experience
LCG 3D Distributed Deployment of Databases
More Interfaces Workplan
3D Application Tests Application test proposals
Database Readiness Workshop Intro & Goals
POOL: Component Overview and use of the File Catalog
Maximum Availability Architecture Enterprise Technology Centre.
Readiness of ATLAS Computing - A personal view
LCGAA nightlies infrastructure
POOL File Catalog: Design & Status
POOL persistency framework for LHC
LCG Distributed Deployment of Databases A Project Proposal
Dirk Düllmann CERN Openlab storage workshop 17th March 2003
Workshop Summary Dirk Duellmann.
LCG middleware and LHC experiments ARDA project
First Internal Pool Release 0.1
POOL/RLS Experience Current CMS Data Challenges shows clear problems wrt to the use of RLS Partially due to the normal “learning curve” on all sides in.
POOL Status & Release Plan for V0.4
SEAL Project Core Libraries and Services
Presentation transcript:

Persistency Project (POOL) Status and Work Plan Proposal Dirk Düllmann dirk.duellmann@cern.ch PEB Meeting, July 16th 2002

Persistency Project Timeline Started officially 19 April Initially staffed with 1.6 FTE Zhen Xie/CMS (100%) - MySQL scalability and reliability test DD (60%) – Completed discussion about requirement list and deployment plans with experiments Persistency Workshop 5-6 June at CERN More requirement and implementation discussions Draft work package breakdown and release sequence proposed Additional experiment resource commitments received Since beginning of July Real design discussions in work package started active participation now ~5FTE Daily meetings with core component WPs Meet twice a week for Catalog WP Collection and Meta Data discussion just starting now Project h/w and s/w infrastructure being defined and becoming available

Experiment Deployment Plans Summary of first round of discussions with the experiments Numbers referring to minimal requirements for a first release (rather than to constraints imposed by the design/implementation) Timescale for first release – September (CMS) Others are interested but more relaxed about date Volume – 10-50TB (CMS) Files – several 100k (ALICE) Distribution – O(10) sites (CMS) Number of population jobs – 10k (CMS) Use of REFs per Event (CMS/ATLAS/LHCb) LHCb: O(100) refs per event; CMS (100-1000); ALICE: not relying on Refs so far Recovery from jobs failures (CMS, ATLAS) CMS: Existing setup allows to just re-issue same job Append to existing files LHCb: each file is written by exactly one job ALICE: single event spans several files

Experiment Focus of Interest (prioritised by # of votes) RootIO integration with grid aware catalog (ALL) Transparent Navigation (ATLAS/CMS/LHCb) ALICE: maybe, but only in some places EDG (ALL), Alien (ALICE), Magda (ATLAS) MySQL as RDBMS implementation until first release (ALL) Consistency between streaming data and meta-data (CMS/ATLAS) At application defined checkpoints during a job Early separation of persistent and transient dictionary (ATLAS/LHCb) Initial release supports persistency for non-TObjects (ATLAS/LHCb) without changes to user class definitions Support for shallow (catalog-only) data copies (CMS) formerly known as cloned Federations Support for deep (extracted part-of-object hierarchy) copies (CMS)

Practical Project Organisation Stay close to initial RTAG component model Work Packages are (nearly) completely aligned Constraint: at least two significant contributions (>30% FTE) per work package to insure independent work have core component developers co-located at CERN Some interface/design discussion will need to take place inside co-located work packages outside CERN by email / phone meetings small workshops at CERN

Proposed Work Package Split Earlier Drafts have been shown on the Persistency Workshop and also later by Torre Catalog and Grid Integration Zhen Xie (50%), Maria Girone (50%), Mathias Steinecke Storage Manager & Refs Markus Frank (50%), Giacomo Govi (50%), Fons Rademakers Dictionary & Conversion Services Craig Tull, Stefan Roiser (50%), Victor Perevoztchikov (30%) Collections and Meta Data David Malon(30%), Chris Laine (50%), Julius Hrivnac (20%), Sasha Vaniachine (20%), RD Schaffer(30%) Common services, integration and testing Giacomo Govi (50%), Maria Girone (50%), Zhen Xie (50%) some commitment made by experiments still under discussion

File Catalog – First Results Scalability and performance for native MySQL catalog close to completion no big surprises compared to Persistency Workshop reports nearly flat scaling up to 200 clients from 10 nodes 0.2 - 0.3 ms per catalog update (depending on client load) impact of batching several updates in one transactions Some instabilities observed database had to be recreate after schema corruption server lockup after file system full condition Need to understand both in more detail before any larger scale production deployment

Catalog Scaling Results (M. Girone/Z. Xie)

EDG/Globus Integration Close contact with EDG WP2 about Giggle (EDG/Globus Replica Location Service) Need to understand software integration with EDG provided middleware software dependencies, release cycle, supported platforms POOL will assume standard EDG testbed installation for client and server machines Plan to repeat catalog scalability and reliability tests with Giggle as backend currently some issues since EDG software is not yet released for RedHat 7.2

Basic S/W Infrastructure Setup Persistency is the first LCG project will need to operate before some of the basic LCG services are fully agreed Need to participate during setup basic s/w infrastructure Repository directory structure Need to build on a few basic services Error Reporting and Exception hierarchy UUID generation Component loading and configuration Configuration control, build and automated test system For now start as pragmatic (and simple) as possible Extract implementation for all of the above from existing projects Expect to change as soon as wider scale agreement is achieved in LCG will require additional work for re-factoring eg after V0.0.1 release

Work Package Progress Core Components – StorageMgr Draft header files and UML diagrams are circulated Gaudi Introspection has been made independent from GAUDI Received example code implementing persistency for foreign (non-TObject) classes Expect to conclude first round of interface discussions this week start with implementation of component prototypes Collections & Meta Data Design discussions between people involved have started but are still in an early state Aim for first interface proposals during the next two weeks

POOL Hardware Resources Discussed hardware requirements with other groups in IT Proposal distributed by Bernd Panzer Now-December 2 disk server machines to be setup as catalog servers one with native MySQL catalog / meta data server one installed as EDG testbed node acting as replica location server 10 client nodes for concurrency test installed as testbed nodes once EDG for RH7.2 is available September-December stability and distribution test (decoupled from development setup) 2 additional disk servers for decoupled stability and distribution tests 3 disk servers for ROOT data November-December more resources for experiment framework integration tests from experiment data challenge allocations as required by experiments

Proposed Release Strategy AIM complete first internal release cycle as soon as possible Start by integration of core components StorageMgr, FileCatalog, Reflection, CacheMgr & Refs result transparent navigational store operational Other (more external) components will start design and implementation already but will only be integrated in later releases in particular Collections and MetaData are essential for final product but less tightly coupled Only the first external release will have all components integrated usable for expert users (eg no automatic installation)

Proposed Release Sequence R 0.0.1 - Basic Navigation all core components for navigation exist and interoperate StorageMgr, Refs & CacheMgr, Dictionary (r/o), FileCatalog some remaining simplifications Assume TObject on read/write – simplified conversion R 0.0.2 – Collections first collection implementation integrated support implicit and explicit collections on either RDBMS or RootIO persistency for foreign classes working persistency for non-TObject classes without need for user code instrumentation EDG/Globus FileCatalog integrated R 0.0.3 – Meta Data & Query (external) meta data annotation and query added for event, event collection and file based meta data

Summary Real development work has (just) started more than half of the people who signed up for the project are actively working towards a very first release Consultancy and (in some areas) modifications from ROOT team expect some inefficiencies because of vacations and project being a test case for new LCG s/w components and procedures September release date and list of requested features seem incompatible Need to limit the influx of new requirements and feature request Propose to aim rather for relevant release content than a too aggressive release date Propose to reschedule V0.0.3 Release for November Assuming that committed resources become available

Root Storage Manager Alternatives User objects are kept in a Root tree Root trees are typed (as eg RDBMS tables) trees need to created by SM objects need to be placed in the right tree are optimised for sequential access should allow foreign (non-TObject) persistency User objects are kept as keyed Objects could allow more efficient random access but some scalability issues all keys are read if a directory is opened? could probably be fixed Do keyed foreign objects work already? if not could they be made work? CMS is actively prototyping second approach for their COBRA Root prototype POOL project will concentrate on first approach and re-synchronise after first development release

Storage Manager Several half day brainstorming sessions with Markus Frank last week cache manager interface and implementation issues role of context integration with dictionary and conversion token structure for Root/RDBMS tables basic model file-container-object container maybe a Root tree, a Root directory, a RDBMS table in more detail flow of control inside SM during read, write, commit operations first UML diagrams have been circulated