STAR C OMPUTING The STAR Databases: Objectivity and Post-Objectivity Torre Wenaus BNL ATLAS Computing Week CERN August 31, 1999.

Slides:



Advertisements
Similar presentations
Object Persistency & Data Handling Session C - Summary Object Persistency & Data Handling Session C - Summary Dirk Duellmann.
Advertisements

Chapter 10: Designing Databases
March 24-28, 2003Computing for High-Energy Physics Configuration Database for BaBar On-line Rainer Bartoldus, Gregory Dubois-Felsmann, Yury Kolomensky,
STAR C OMPUTING STAR Overview and OO Experience Torre Wenaus STAR Computing and Software Leader Brookhaven National Laboratory, USA CHEP 2000, Padova February.
1 Databases in ALICE L.Betev LCG Database Deployment and Persistency Workshop Geneva, October 17, 2005.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
O. Stézowski IPN Lyon AGATA Week September 2003 Legnaro Data Analysis – Team #3 ROOT as a framework for AGATA.
Objectivity Data Migration Marcin Nowak, CERN Database Group, CHEP 2003 March , La Jolla, California.
8 Systems Analysis and Design in a Changing World, Fifth Edition.
DATA PRESERVATION IN ALICE FEDERICO CARMINATI. MOTIVATION ALICE is a 150 M CHF investment by a large scientific community The ALICE data is unique and.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
SSIS Over DTS Sagayaraj Putti (139460). 5 September What is DTS?  Data Transformation Services (DTS)  DTS is a set of objects and utilities that.
2/10/2000 CHEP2000 Padova Italy The BaBar Online Databases George Zioulas SLAC For the BaBar Computing Group.
Argonne National Laboratory ATLAS Core Database Software U.S. ATLAS Collaboration Meeting New York 22 July 1999 David Malon
INFO425: Systems Design INFORMATION X Finalizing Scope (functions/level of automation)  Finalizing scope in terms of functions and level of.
STAR C OMPUTING ROOT in STAR Torre Wenaus STAR Computing and Software Leader Brookhaven National Laboratory, USA ROOT 2000 Workshop, CERN February 3, 2000.
Event Metadata Records as a Testbed for Scalable Data Mining David Malon, Peter van Gemmeren (Argonne National Laboratory) At a data rate of 200 hertz,
Chapter 3: Operating-System Structures System Components Operating System Services System Calls System Programs System Structure Virtual Machines System.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
IMPLEMENTATION OF SOFTWARE INPUT OUTPUT CONTROLLERS FOR THE STAR EXPERIMENT J. M. Burns, M. Cherney*, J. Fujita* Creighton University, Department of Physics,
Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration Migrating PHENIX databases from object to relational model.
Object-Oriented Software Engineering Practical Software Development using UML and Java Chapter 1: Software and Software Engineering.
Offline Coordinators  CMSSW_7_1_0 release: 17 June 2014  Usage:  Generation and Simulation samples for run 2 startup  Limited digitization and reconstruction.
Chapter 4 Realtime Widely Distributed Instrumention System.
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 3: Operating-System Structures System Components Operating System Services.
David N. Brown Lawrence Berkeley National Lab Representing the BaBar Collaboration The BaBar Mini  BaBar  BaBar’s Data Formats  Design of the Mini 
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Summary of 1 TB Milestone RD Schaffer Outline : u Goals and assumptions of the exercise u The hardware constraints u The basic model u What have we understood.
Object-Oriented Software Engineering Practical Software Development using UML and Java Chapter 1: Software and Software Engineering.
HENP Computing at BNL Torre Wenaus STAR Software and Computing Leader BNL RHIC & AGS Users Meeting Asilomar, CA October 21, 1999.
STAR C OMPUTING STAR Computing Infrastructure Torre Wenaus BNL STAR Collaboration Meeting BNL Jan 31, 1999.
1 Geospatial and Business Intelligence Jean-Sébastien Turcotte Executive VP San Francisco - April 2007 Streamlining web mapping applications.
5 May 98 1 Jürgen Knobloch Computing Planning for ATLAS ATLAS Software Week 5 May 1998 Jürgen Knobloch Slides also on:
NOVA Networked Object-based EnVironment for Analysis P. Nevski, A. Vaniachine, T. Wenaus NOVA is a project to develop distributed object oriented physics.
Grand Challenge and PHENIX Report post-MDC2 studies of GC software –feasibility for day-1 expectations of data model –simple robustness tests –Comparisons.
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
Introduction to Database Systems1. 2 Basic Definitions Mini-world Some part of the real world about which data is stored in a database. Data Known facts.
1 GCA Application in STAR GCA Collaboration Grand Challenge Architecture and its Interface to STAR Sasha Vaniachine presenting for the Grand Challenge.
STAR Event data storage and management in STAR V. Perevoztchikov Brookhaven National Laboratory,USA.
STAR C OMPUTING STAR Analysis Operations and Issues Torre Wenaus BNL STAR PWG Videoconference BNL August 13, 1999.
STAR Collaboration, July 2004 Grid Collector Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani Lawrence Berkeley National.
CASE (Computer-Aided Software Engineering) Tools Software that is used to support software process activities. Provides software process support by:- –
PHENIX and the data grid >400 collaborators 3 continents + Israel +Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
NOVA A Networked Object-Based EnVironment for Analysis “Framework Components for Distributed Computing” Pavel Nevski, Sasha Vanyashin, Torre Wenaus US.
STAR C OMPUTING Plans for Production Use of Grand Challenge Software in STAR Torre Wenaus BNL Grand Challenge Meeting LBNL 10/23/98.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
Managing Enterprise GIS Geodatabases
Infrastructure for Data Warehouses. Basics Of Data Access Data Store Machine Memory Buffer Memory Cache Data Store Buffer Bus Structure.
STAR C OMPUTING The STAR Databases: From Objectivity to ROOT+MySQL Torre Wenaus BNL ATLAS Computing Week CERN August 31, 1999.
Some Ideas for a Revised Requirement List Dirk Duellmann.
Welcome to the PRECIS training workshop
The ATLAS TAGs Database - Experiences and further developments Elisabeth Vinek, CERN & University of Vienna on behalf of the TAGs developers group.
Andrea Valassi (CERN IT-DB)CHEP 2004 Poster Session (Thursday, 30 September 2004) 1 HARP DATA AND SOFTWARE MIGRATION FROM TO ORACLE Authors: A.Valassi,
The MEG Offline Project General Architecture Offline Organization Responsibilities Milestones PSI 2/7/2004Corrado Gatto INFN.
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
System Components Operating System Services System Calls.
Systems Analysis and Design in a Changing World, Fifth Edition
CMS High Level Trigger Configuration Management
Systems Analysis – ITEC 3155 Evaluating Alternatives for Requirements, Environment, and Implementation.
Chapter 6 Database Design
Nuclear Physics Data Management Needs Bruce G. Gibbard
OO-Design in PHENIX PHENIX, a BIG Collaboration A Liberal Data Model
Chapter 2: Operating-System Structures
Simulation and Physics
Chapter 2: Operating-System Structures
Presentation transcript:

STAR C OMPUTING The STAR Databases: Objectivity and Post-Objectivity Torre Wenaus BNL ATLAS Computing Week CERN August 31, 1999

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Content I will address mainly the event store side of STAR databases I will discuss  Our starting point in terms of environment and requirements  Objectivity l Initial decisions l Work and experience l Lessons learned  Subsequent decisions and altered directions in light of experience  Where we are now in STAR databases and event store  Requirements and decision factors revisited today  Conclusions

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 STAR at RHIC RHIC: Relativistic Heavy Ion Collider at Brookhaven National Laboratory  Colliding Au - Au nuclei at 200GeV/nucleon  Principal objective: Discovery and characterization of the Quark Gluon Plasma  Additional spin physics program in polarized p - p  Engineering run 6-8/99; first year physics run 1/00 STAR experiment  One of two large ‘HEP-scale’ experiments at RHIC, >400 collaborators each (PHENIX is the other)  Heart of experiment is a Time Projection Chamber (TPC) drift chamber (operational) together with Si tracker (year 2) and electromagnetic calorimeter (staged over years 1-3)  Hadrons, jets, electrons and photons over large solid angle

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 The STAR Computing Task Data recording rate of 20MB/sec; ~12MB raw data per event (~1Hz)  ~4000+ tracks/event recorded in tracking detectors (factor of 2 uncertainty in physics generators)  High statistics per event permit event by event measurement and correlation of QGP signals such as strangeness enhancement, J/psi attenuation, high Pt parton energy loss modifications in jets, global thermodynamic variables (eg. Pt slope correlated with temperature)  17M Au-Au events (equivalent) recorded in nominal year Relatively few but highly complex events requiring large processing power  Wide range of physics studies: ~100 concurrent analyses in ~7 physics working groups

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Computing Requirements Nominal year processing and data volume requirements: Raw data volume: 200TB Reconstruction: 2800 Si95 total CPU, 30TB DST data  10x event size reduction from raw to reco  1.5 reconstruction passes/event assumed Analysis: 4000 Si95 total analysis CPU, 15TB micro-DST data  Si95-sec/event per MB of DST depending on analysis l Wide range, from CPU-limited to I/O limited  ~100 active analyses, 5 passes per analysis  micro-DST volumes from.1 to several TB Simulation: 3300 Si95 total including reconstruction, 24TB Total nominal year data volume: 270TB Total nominal year CPU: 10,000 Si95

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Computing Facilities Dedicated RHIC computing center at BNL, the RHIC Computing Facility  Data archiving and processing for reconstruction and analysi l Simulation done offsite  10,000 (reco) + 7,500 (analysis) Si95 CPU l Primarily Linux; some Sun for I/O intensive analysis  ~50TB disk, 270TB robotic tape, 200MB/s, managed by HPSS  Current scale (STAR allocation, ~40% of total): l ~2500 Si95 CPU l 3TB disk Support for (a subset of) physics analysis computing at home institutions

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Offline Software Environment Current software base a mix of Fortran (55%) and C++ (45%)  from ~80%/20% (~95%/5% in non-infrastructure code) in 9/98  New development, and all post-reco analysis, in C++ Framework built over ROOT adopted 11/98  Origins in the ‘Makers’ of ATLFAST  Supports legacy Fortran codes, table (IDL) based data structures developed in previous StAF framework without change  Deployed in offline production and analysis in our ‘Mock Data Challenge 2’, 2-3/99 Post-reconstruction analysis: C++/OO data model ‘StEvent’  StEvent interface is ‘generic C++’; analysis codes are unconstrained by ROOT and need not (but may) use it Next step: migrate the OO data model upstream to reco

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Database Development Effort Infrastructure (including databases) development team: ~7 people  Two less than planned minimum due to budget constraints Database/event store development ~ 2.5 FTE-years in last 1.5 years  Various fractions of (primarily) l Valery Fine, BNL l Yuri Fisyak, BNL l Victor Perevoztchikov, BNL l Jeff Porter, BNL l Sasha Vanyashin, Royal Stockholm Institute of Technology l Torre Wenaus, BNL Strong reliance on leveraging community, commercial, open software

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Initial RHIC DB Technology Choices A RHIC-wide Event Store Task Force in Fall ‘97 addressed data management alternatives  Requirements formulated by the four experiments  Objectivity and ROOT were the ‘contenders’ put forward  STAR and PHENIX selected Objectivity as the basis for data management l Concluded that only Objectivity met the requirements of their event stores  ROOT selected by the smaller experiments and seen by all as analysis tool with great potential  Issue for the two larger experiments: l Where to draw a dividing line between Objectivity and ROOT in the data model and data processing

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Requirements -- And Fall ‘97 View of Fulfillment

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Objectivity in STAR, with a BaBar Kick-Start Objectivity selected for all STAR Computing database needs  Event store  Conditions and calibrations, configurations Decision to use Objectivity would have been inconceivable were it not for BaBar leading the way with an intensive development effort to deploy Objectivity on the same timescale as RHIC and at a similar scale  Benefit both from their code and their brush-clearing experience Full BaBar software distribution was imported and used as an ‘external library’ providing the foundation for STAR event store and conditions database  BaBar’s SRT and site customization features used to adapt their code without direct modification to the STAR environment

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Event Store Implementation BaBar event store software adapted to implement STAR event store down to the level of event components  Event collections; Event collection dictionary with organization in Unix directory tree style  System, group, user level authorization controls  SRT based management, build tools Event component objects implemented in-house  Direct mapping between objects and existing IDL-defined data structures (C structs)  Flat structure under event header; flexible partitioning of components to different streams based on access characteristics

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Alleviating Risk Early (and easy) decisions taken to limit technology risk:  Full separation of persistent and transient event models l BaBar transient-persistent mapping scheme not used  Raw data not stored in Objectivity l Accommodation of file-based components in event collections  StAF-standard XDF I/O retained as production standard and backup  ROOT I/O pursued l to provide a future backup, replacing XDF, that provides a true object store §remove restriction to table-based persistent data structures l as a competitive partial replacement of Objectivity §addressing data storage, but not data management

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Event Component Implementation IDL-based persistent data model ensures XDF compatibility (fully technology-neutral; C structs with relations via indices)  But restricts how we use Objectivity: don’t even pretend it’s C++ No impact on the transient event because of its decoupling from persistent  We accept the cost of the translation; it’s worth it  At the DST level, data model is stable so code burden isn’t an issue; neither is performance  Decouples design optimization and development timescales of persistent and transient representations Of course, in doing this decoupling, we are throwing away a good part of the advantage argued for Objectivity from the beginning  No ‘impedance mismatch’; an ODBMS can map directly onto the OO data models we want to build and work with Regarded it as a necessity wrought of uncertainty  can broaden our use of Objectivity later, in an environment of less uncertainty and more experience

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Objectivity Deployment Objectivity event store deployed in Mock Data Challenge 1, Sep-Oct ‘98  MDC: A RHIC-wide exercise in stress-testing RCF and experiment hardware and software in full production environment from (simulated) raw data through physics analysis In STAR’s MDC1  217k Au-Au events (1.7TB) simulated on offsite Cray T3Es with network transport to RCF HPSS  168k events reconstructed (600GB of DSTs generated)  50GB Objectivity DST database built (disk space limited) and used at a small scale with STAR analysis codes l Integrated with the offline framework with translators to/from the transient table-based DST l Used at a small scale with STAR analysis codes l Used by the Grand Challenge in testing their architecture

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 MDC1 Experience Development and deployment quite smooth, but Objectivity development not a fast, efficient process  Modify/rebuild/execute development cycle is heavy and slow, particularly if schema is changed  It’s an elegant tool but it requires highly complex software and software environment to develop real applications in the real world No problems storing and working with 50GB of DSTs, in a very restricted setting  Very small number of users l Access control and schema management tools not exercised in a realistic environment  Serial usage only  One platform (Sun/Solaris) only, and not our principal one l Linux port slow to appear and not keeping up with compilers l Big issue lurking: BaBar software not yet ported to Linux

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Objectivity Burdens The list of burdens imposed by Objectivity grew as our experience and lessons from BaBar mounted  Management, development burden imposed by ensuring consistent schema in a single experiment-wide federation  Schema evolution unusable if forward compatibility is desired (ability to run old executables on new data)  Do-it-yourself access control, particularly with AMS  Risk of major impact from platform lock-in due to porting delays; both Linux and Sun  Scalability concerns (fall ‘98) -- lock manager performance issues in parallel usage?

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 A Hybrid Event Store for STAR? No MDC1 show-stoppers, but many open issues and increasing concern over viability at our manpower levels Series of review/decision workshops held in Oct/Nov ‘98  How do we store our events?  How and where do we use ROOT?  How do we present our events to analysis codes, and what form should those codes take? Decisions relating to data management:  Deploy ROOT, together with Objectivity, as DST store in MDC2 (Feb-Mar ‘99) l Decide after MDC2 experience on ROOT vs. Objectivity  Use ROOT as analysis toolkit and underlying basis for framework l Use ROOT for post-DST storage (micro-DSTs) §Reap advantages of close coupling to analysis toolkit

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 ROOT I/O Adopted as a Storage Solution ROOT-based DST storage fully deployed and exercised in MDC2 production  Revealed problems in robustness against crashes, subsequently fixed  Meets requirements in multiple data streams for different event components Robust I/O top priority for ROOT team (FNAL ROOT Workshop, March) CDF’s decision to use ROOT I/O in Run II also a factor Conclusion: A functional and viable solution for an object store  Easy decision to adopt it over Objectivity ROOT has also been implemented as a persistent back end of our OO/C++ data model StEvent, giving us a direct object store without ROOT appearing in the data model interface.

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Requirements: STAR 8/99 View (My Version)

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 RHIC Data Management: Factors For Evaluation My perception of changes in the STAR view from ‘97 to now are shown Objy Root+MySQLFactor  Cost  Performance and capability as data access solution  Quality of technical support  Ease of use, quality of doc   Ease of integration with analysis   Ease of maintenance, risk   Commonality among experiments   Extent, leverage of outside usage   Affordable/manageable outside RCF   Quality of data distribution mechanisms   Integrity of replica copies   Availability of browser tools   Flexibility in controlling permanent storage location   Level of relevant standards compliance, eg. ODMG   Java access   Partitioning DB and resources among groups

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Hybrid Event Store, Yes; With Objectivity, No Adoption of ROOT I/O for the event store leaves Objectivity with one role left to cover: the true ‘database’ functions of the event store  Navigation among event collections, runs/events, event components  Data locality (now translates basically to file lookup)  Management of dynamic, asynchronous updating of the event store from one end of the processing chain to the other l From initiation of an event collection (run) in online through addition of components in reconstruction, analysis and their iterations But with the concerns and weight of Objectivity it is overkill for this role. So we went shopping…  looking to leverage the world around us, as always  and eyeing particularly the rising wave of Internet-driven tools and open software and came up with MySQL in May.

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 MySQL as the STAR Database  Relational DB, open software, very fast, widely used on the web  Not a full featured heavyweight like Oracle (hence simple and fast) l No transactions, journalling  Development pace is very fast with a wide range of tools to use (good interfaces to Perl,C/C++, Java; easy and powerful web interfacing)  Wonderful experience so far. Like a quick protyping tool that is also production-capable for the right applications  In production use since June for file cataloguing and run logging l ~tables of ~20k rows cataloguing ~9TB of data  Under test as the basis for compact event tags (and to test scalability) l No problem so far with 2M row tables of 100bytes/row l Query response in a few seconds  Multiple servers, databases can be used as needed to address scalability, access and locking characteristics  MySQL based conditions (calibrations, geometry, parameters, conditions) implemented  Developing generic tools for STAR and outside use.

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 MySQL Based Tools Just flash a few screen shots

STAR C OMPUTING Torre Wenaus, BNL ATLAS meeting 8/99 Conclusions The circumstances of STAR  Startup this year  Slow start in addressing event store implementation, C++ migration  Large base of legacy software  Extremely limited manpower and computing resources drive us to very practical and pragmatic data management choices  Beg, steal and borrow from the community  Deploy community and industry standard technologies  Isolate implementation choices behind standard interfaces, to revisit and re-optimize in the future which leverage existing STAR strengths  Component and standards-based software greatly eases integration of new technologies l preserving compatibility with existing tools for selective and fall-back use l while efficiently migrating legacy software and legacy physicists After some course corrections, we have a capable data management architecture for startup that scales to STAR’s data volumes … but Objectivity is no longer in the picture.