A Flexible Distributed Event-level Metadata System for ATLAS David Malon*, Jack Cranshaw, Kristo Karr (Argonne), Julius Hrivnac, Arthur Schaffer (LAL Orsay)

Slides:



Advertisements
Similar presentations
Metadata Progress GridPP18 20 March 2007 Mike Kenyon.
Advertisements

Metadata (data about data) GridPP-15, Paul Millar.
Database Planning, Design, and Administration
©Silberschatz, Korth and Sudarshan12.1Database System Concepts Chapter 12: Indexing and Hashing Basic Concepts Ordered Indices B+-Tree Index Files B-Tree.
Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
Advanced Database Systems September 2013 Dr. Fatemeh Ahmadi-Abkenari 1.
Database Replication techniques: a Three Parameter Classification Authors : Database Replication techniques: a Three Parameter Classification Authors :
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL March 25, 2003 CHEP 2003 Data Analysis Environment and Visualization.
1 Distributed Databases Chapter Two Types of Applications that Access Distributed Databases The application accesses data at the level of SQL statements.
29 July 2008Elizabeth Gallas1 An introduction to “TAG”s for ATLAS analysis Elizabeth Gallas Oxford Oxford ATLAS Physics Meeting Tuesday 29 July 2008.
1 Distributed Databases Chapter What is a Distributed Database? Database whose relations reside on different sites Database some of whose relations.
AMI S.A. Datasets… Solveig Albrand. AMI S.A. A set is… A number of things grouped together according to a system of classification, or conceived as forming.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
CERN - IT Department CH-1211 Genève 23 Switzerland t Partitioning in COOL Andrea Valassi (CERN IT-DM) R. Basset (CERN IT-DM) Distributed.
By N.Gopinath AP/CSE. Why a Data Warehouse Application – Business Perspectives  There are several reasons why organizations consider Data Warehousing.
Argonne National Laboratory ATLAS Core Database Software U.S. ATLAS Collaboration Meeting New York 22 July 1999 David Malon
Database System Development Lifecycle © Pearson Education Limited 1995, 2005.
Event Metadata Records as a Testbed for Scalable Data Mining David Malon, Peter van Gemmeren (Argonne National Laboratory) At a data rate of 200 hertz,
ITEC224 Database Programming
ATLAS DQ2 Deletion Service D.A. Oleynik, A.S. Petrosyan, V. Garonne, S. Campana (on behalf of the ATLAS Collaboration)
David Adams ATLAS ATLAS Distributed Analysis David Adams BNL March 18, 2004 ATLAS Software Workshop Grid session.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
LHC: ATLAS Experiment meeting “Conditions” data challenge Elizabeth Gallas - Oxford - August 29, 2009 XLDB3.
30 Jan 2009Elizabeth Gallas1 Introduction to TAGs Elizabeth Gallas Oxford ATLAS-UK Distributed Computing Tutorial January 2009.
File System Management File system management encompasses the provision of a way to store your data in a computer, as well as a way for you to find and.
Databases E. Leonardi, P. Valente. Conditions DB Conditions=Dynamic parameters non-event time-varying Conditions database (CondDB) General definition:
Datasets on the GRID David Adams PPDG All Hands Meeting Catalogs and Datasets session June 11, 2003 BNL.
ATLAS Detector Description Database Vakho Tsulaia University of Pittsburgh 3D workshop, CERN 14-Dec-2004.
An RTAG View of Event Collections, and Early Implementations David Malon ATLAS Database Group LHC Persistence Workshop 5 June 2002.
ATLAS Data Challenges US ATLAS Physics & Computing ANL October 30th 2001 Gilbert Poulard CERN EP-ATC.
Caitriana Nicholson, CHEP 2006, Mumbai Caitriana Nicholson University of Glasgow Grid Data Management: Simulations of LCG 2008.
The PHysics Analysis SERver Project (PHASER) CHEP 2000 Padova, Italy February 7-11, 2000 M. Bowen, G. Landsberg, and R. Partridge* Brown University.
08-Nov Database TEG workshop, Nov 2011 ATLAS Oracle database applications and plans for use of the Oracle 11g enhancements Gancho Dimitrov.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
AliEn AliEn at OSC The ALICE distributed computing environment by Bjørn S. Nilsen The Ohio State University.
Physical Database Design Purpose- translate the logical description of data into the technical specifications for storing and retrieving data Goal - create.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL November 17, 2003 SC2003 Phoenix.
David Adams ATLAS ATLAS distributed data management David Adams BNL February 22, 2005 Database working group ATLAS software workshop.
The ATLAS TAGs Database - Experiences and further developments Elisabeth Vinek, CERN & University of Vienna on behalf of the TAGs developers group.
ATLAS Database Access Library Local Area LCG3D Meeting Fermilab, Batavia, USA October 21, 2004 Alexandre Vaniachine (ANL)
The ATLAS Computing Model and USATLAS Tier-2/Tier-3 Meeting Shawn McKee University of Michigan Joint Techs, FNAL July 16 th, 2007.
Magda Distributed Data Manager Prototype Torre Wenaus BNL September 2001.
David Adams ATLAS Datasets for the Grid and for ATLAS David Adams BNL September 24, 2003 ATLAS Software Workshop Database Session CERN.
TAGS in the Analysis Model Jack Cranshaw, Argonne National Lab September 10, 2009.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Site Services and Policies Summary Dirk Düllmann, CERN IT More details at
The ATLAS Computing & Analysis Model Roger Jones Lancaster University ATLAS UK 06 IPPP, 20/9/2006.
Enabling Grids for E-sciencE CMS/ARDA activity within the CMS distributed system Julia Andreeva, CERN On behalf of ARDA group CHEP06.
ELSSISuite Services QIZHI ZHANG Argonne National Laboratory on behalf of the TAG developers group ATLAS Software and Computing Week, 4~8 April, 2011.
David Adams ATLAS ATLAS Distributed Analysis (ADA) David Adams BNL December 5, 2003 ATLAS software workshop CERN.
Finding Data in ATLAS. May 22, 2009Jack Cranshaw (ANL)2 Starting Point Questions What is the latest reprocessing of cosmics? Are there are any AOD produced.
David Adams ATLAS ATLAS Distributed Analysis and proposal for ATLAS-LHCb system David Adams BNL March 22, 2004 ATLAS-LHCb-GANGA Meeting.
Next-Generation Navigational Infrastructure and the ATLAS Event Store Abstract: The ATLAS event store employs a persistence framework with extensive navigational.
Database Access Patterns in ATLAS Computing Model G. Gieraltowski, J. Cranshaw, K. Karr, D. Malon, A. Vaniachine ANL P, Nevski, Yu. Smirnov, T. Wenaus.
ATLAS Distributed Computing ATLAS session WLCG pre-CHEP Workshop New York May 19-20, 2012 Alexei Klimentov Stephane Jezequel Ikuo Ueda For ATLAS Distributed.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
Magda Distributed Data Manager Torre Wenaus BNL October 2001.
Metadata and Supporting Tools on Day One David Malon Argonne National Laboratory Argonne ATLAS Analysis Jamboree Chicago, Illinois 22 May 2009.
ATLAS Distributed Computing Tutorial Tags: What, Why, When, Where and How? Mike Kenyon University of Glasgow.
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL May 19, 2003 BNL Technology Meeting.
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
Module 11: File Structure
(on behalf of the POOL team)
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Data Lifecycle Review and Outlook
Grid Data Integration In the CMS Experiment
ATLAS DC2 & Continuous production
Presentation transcript:

A Flexible Distributed Event-level Metadata System for ATLAS David Malon*, Jack Cranshaw, Kristo Karr (Argonne), Julius Hrivnac, Arthur Schaffer (LAL Orsay) CHEP’06, Mumbai February 2005

13-17 February 2006 Malon et al CHEP'06 2 Acknowledgment  Thanks to Caitriana Nicholson (Glasgow) for presenting on our behalf  Some of us would rather be in Mumbai!

13-17 February 2006 Malon et al CHEP'06 3 Event-level metadata in the ATLAS computing model  ATLAS Computing Model proposes an event-level metadata system--a “tag” database--for rapid and efficient event selection  Budget allows for approximately 1 kilobyte of “payload” metadata per event, so storage requirements are at the scale of a small number of terabytes  Should be widely replicable in principle--all Tier 1s and most Tier 2s should be able to accommodate it if its unique resource demands are not onerous

13-17 February 2006 Malon et al CHEP'06 4 Underlying technology  Persistence technology for tags are currently based upon POOL collections  Collections store references to objects, along with a corresponding attribute list upon which one might base object-level selection  Implemented in ROOT and in relational database backends

13-17 February 2006 Malon et al CHEP'06 5 Data flow  Event tags are written into ROOT files when Analysis Object Data (AOD) are produced at Tier 0  Strictly speaking, tags are produced when (relatively small) AOD files are merged into larger files  File-based tags are bulk loaded into relational database at CERN  File-based tags may not be discarded--they may serve as indices for simple attribute-based selection and direct addressing of specific events in the corresponding data files  Tags are sent from Tier 0 to Tier 1s, and thence to Tier 2s  May send only file-based tags to Tier 2s: depends on Tier 2 capabilities

13-17 February 2006 Malon et al CHEP'06 6 Rome Workshop Event Tag Production AOD Files (1000 evts each) Root Collections (1 Per AOD File) Dataset n Dataset 2 Dataset 1 Table 1 Table 2 Table n Event Tag Database (Tier 0) Replica Databases (Tier 1) Relational Database Site 1 Site 2 Site n

13-17 February 2006 Malon et al CHEP'06 7 Machinery and middleware  Queries return lists of references to events, grouped by id (GUID) of the containing files  ATLAS infrastructure stores references both to AOD and to upstream processing stages (e.g., ESD) and can return any or all of these  Utility also returns the list of distinct file GUIDs for use by resource brokers and job schedulers  (Some) ATLAS distributed analysis prototypes are already capable of splitting the event list on these file GUID boundaries and spawning multiple jobs accordingly, to allow parallel processing of the sample

13-17 February 2006 Malon et al CHEP'06 8 ATLAS physics experience with tags  Tags were put into the hands of ATLAS physicists in advance of the June 2005 ATLAS Rome Physics Workshop  Physicists defined tag “schema” and provided content  Event store group ensured that tags contained, not only pointers to events in the latest processing stage, but to upstream data as well  Rome data production was globally distributed; only datasets that were moved back to CERN had tags inserted into collaboration-wide tag database  Just under 3 million events in master tag database  Feedback was positive: triggered initiation of a collaboration-wide tag content review in Fall 2005  Report and recommendations due late February 2006

13-17 February 2006 Malon et al CHEP'06 9 Performance tests and experience  Rome tag production with genuine tag content (from simulated data) provided a testbed for many things, including implementation alternatives, scalability, and performance tests  Used for tests of indexing strategies, technology comparisons, …; details on ATLAS Twiki and elsewhere  Performance was “adequate” for a few million events  Conclusions: some grounds for optimism, some grounds for concern about scalability of a master tag database  Clearly not ready yet for 10**9 events  Room for divide-and-conquer strategies (horizontal partitioning, e.g., by trigger stream, vertical partitioning (e.g., by separating trigger from physics metadata), as well as indexing and back-end optimizations

13-17 February 2006 Malon et al CHEP'06 10 Event collections and streaming  One natural use case: instead of extracting into a set of files the events that satisfy your selection criteria (a “skim” in the parlance of some experiments), what about simply building a list of references to those events?  Should you be disappointed if you are a loser in the collaboration-wide negotiations--your “skim” is not one of the standard ones--and all you have instead is a list of event references?  Note that you can always use your event collection to extract the events you want into your own files on your own resources--we have utilities for this.  What are the consequences of using reference lists to avoid the storage waste of building overlapping streams?  e.g., you get references to events that satisfy your selection criteria but “belong” to someone else’s stream

13-17 February 2006 Malon et al CHEP'06 11 Event collections versus streams tests  Used the Rome tag database to test the performance implications of iterating through N events scattered uniformly through M files versus iterating through them after they had been gathered into a single file (or file sequence)  Results: Cost is approximately the cost of opening M-1 additional files-- navigational overhead was too small to measure  Rule of thumb for ATLAS users: ~1 second per additional file in CERN Castor; much smaller on AFS, though space constraints are more stringent; don’t have quotable SRM/dCache figures yet  ATLAS has this month commissioned a streaming work group to decide about stream definitions and the processing stages at which streaming will be done:  Prototyping based upon tag database work will be integral to the strategy evaluation and comparison process

13-17 February 2006 Malon et al CHEP'06 12 Collection vs. Direct File Access Analysis Result User Single Large File User Collection Analysis Result Database Multiple Files File Catalog

13-17 February 2006 Malon et al CHEP'06 13 Distributed data management (DDM) integration  ATLAS has, since the Rome physics workshop, substantially altered its distributed data management model to be more “dataset”-oriented  Poses challenges to the event-level metadata system, and to the production system as well  Tag payloads today are references to events, which, following the LCG POOL model, embed file ids. This was “easy” for the distributed data management system in the past: take the output list of unique file GUIDs, and find the files or the sites that host them.

13-17 February 2006 Malon et al CHEP'06 14 Distributed data management integration issues  “Dataset” questions  What is the corresponding dataset? Is it known at the time the tag is written? Does this imply that a job that creates an event file needs to know the output dataset affiliation in advance (true in general for production jobs, but in general?)?  How are datasets identified? Is versioning relevant? Is any (initial?) dataset assignment of an event immutable?  Result of a query to an event collection is another event collection, which can be published as a “dataset” in the DDM sense  How is the resulting dataset marshalled from the containing (file-based) datasets?  We now have answers to many of these questions, and integration work is progressing in advance of the next round of commissioning test.  Glasgow group (including our substitute presenter, Caitriana--thanks!) is addressing event-level metadata/DDM integration

13-17 February 2006 Malon et al CHEP'06 15 Replication  Used Octopus-based replication tools for heterogeneous replication (Oracle to MySQL) CERN-->Brookhaven  Plan to use Oracle streams for Tier 0 to Tier 1 replication in LHC Service Challenge 4 tests later this year

13-17 February 2006 Malon et al CHEP'06 16 Stored procedures  Have done some preliminary work with Java stored procedures in Oracle for queries that are procedurally simple, but complicated (or lengthy) to express in SQL  Capabilities look promising; no performance numbers yet  We may use this approach for decoding trigger information (don’t know yet--Rome physics simulation included no trigger simulation, and hence no trigger signature representation)

13-17 February 2006 Malon et al CHEP'06 17 Ongoing work  Need POOL collections improvements if POOL is to provide a basis for a genuine ATLAS event-level metadata system  Much work is underway: bulk loading improvements, design for horizontal partitioning (multiple table implementation), and for tag extensibility  Production system and distributed data management integration already mentioned  We haven’t yet investigated less-than-naïve indexing strategies, bit- sliced indexing (though we have some experience with it from previous projects), or any kind of server-side tuning  Computing System Commissioning tests in 2006 will tell us much about the future shape of event-level metadata systems in ATLAS