STAR C OMPUTING ROOT in STAR Torre Wenaus STAR Computing and Software Leader Brookhaven National Laboratory, USA ROOT 2000 Workshop, CERN February 3, 2000.

Slides:



Advertisements
Similar presentations
STAR C OMPUTING STAR Overview and OO Experience Torre Wenaus STAR Computing and Software Leader Brookhaven National Laboratory, USA CHEP 2000, Padova February.
Advertisements

1 Databases in ALICE L.Betev LCG Database Deployment and Persistency Workshop Geneva, October 17, 2005.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
O. Stézowski IPN Lyon AGATA Week September 2003 Legnaro Data Analysis – Team #3 ROOT as a framework for AGATA.
Grid Collector: Enabling File-Transparent Object Access For Analysis Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani.
EventStore Managing Event Versioning and Data Partitioning using Legacy Data Formats Chris Jones Valentin Kuznetsov Dan Riley Greg Sharp CLEO Collaboration.
CERN IT Department CH-1211 Genève 23 Switzerland t Integrating Lemon Monitoring and Alarming System with the new CERN Agile Infrastructure.
STAR C OMPUTING The STAR Databases: Objectivity and Post-Objectivity Torre Wenaus BNL ATLAS Computing Week CERN August 31, 1999.
2/10/2000 CHEP2000 Padova Italy The BaBar Online Databases George Zioulas SLAC For the BaBar Computing Group.
LCIO A persistency framework for LC detector simulation studies Frank Gaede, DESY, IT 4 th ECFA/DESY LC Workshop Amsterdam April 1 st -4 th 2003.
STAR C OMPUTING Maker and I/O Model in STAR Victor Perevoztchikov.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration Migrating PHENIX databases from object to relational model.
Indo-US Workshop, June23-25, 2003 Building Digital Libraries for Communities using Kepler Framework M. Zubair Old Dominion University.
HPS Online Software Discussion Jeremy McCormick, SLAC Status and Plans.
MySQL and GRID Gabriele Carcassi STAR Collaboration 6 May Proposal.
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL U.S. ATLAS Physics and Computing Advisory Panel Review Argonne National Laboratory Oct 30, 2001.
David N. Brown Lawrence Berkeley National Lab Representing the BaBar Collaboration The BaBar Mini  BaBar  BaBar’s Data Formats  Design of the Mini 
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within.
Event Data History David Adams BNL Atlas Software Week December 2001.
HENP Computing at BNL Torre Wenaus STAR Software and Computing Leader BNL RHIC & AGS Users Meeting Asilomar, CA October 21, 1999.
Web application for detailed real-time database transaction monitoring for CMS condition data ICCMSE 2009 The 7th International Conference of Computational.
ATLAS, U.S. ATLAS, and Databases David Malon Argonne National Laboratory DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National Laboratory.
STAR C OMPUTING STAR Computing Infrastructure Torre Wenaus BNL STAR Collaboration Meeting BNL Jan 31, 1999.
5 May 98 1 Jürgen Knobloch Computing Planning for ATLAS ATLAS Software Week 5 May 1998 Jürgen Knobloch Slides also on:
NOVA Networked Object-based EnVironment for Analysis P. Nevski, A. Vaniachine, T. Wenaus NOVA is a project to develop distributed object oriented physics.
CERN – European Organization for Nuclear Research Administrative Support - Internet Development Services CET and the quest for optimal implementation and.
1 GCA Application in STAR GCA Collaboration Grand Challenge Architecture and its Interface to STAR Sasha Vaniachine presenting for the Grand Challenge.
Hall-D/GlueX Software Status 12 GeV Software Review III February 11[?], 2015 Mark Ito.
STAR Event data storage and management in STAR V. Perevoztchikov Brookhaven National Laboratory,USA.
STAR C OMPUTING STAR Analysis Operations and Issues Torre Wenaus BNL STAR PWG Videoconference BNL August 13, 1999.
STAR Collaboration, July 2004 Grid Collector Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani Lawrence Berkeley National.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
PHENIX and the data grid >400 collaborators 3 continents + Israel +Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
NOVA A Networked Object-Based EnVironment for Analysis “Framework Components for Distributed Computing” Pavel Nevski, Sasha Vanyashin, Torre Wenaus US.
STAR C OMPUTING Plans for Production Use of Grand Challenge Software in STAR Torre Wenaus BNL Grand Challenge Meeting LBNL 10/23/98.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
1 fileCatalog, tagDB and GCA A. Vaniachine Grand Challenge STAR fileCatalog, tagDB and Grand Challenge Architecture A. Vaniachine presenting for the Grand.
STAR C OMPUTING The STAR Databases: From Objectivity to ROOT+MySQL Torre Wenaus BNL ATLAS Computing Week CERN August 31, 1999.
STAR C OMPUTING Introduction Torre Wenaus BNL May ‘99 STAR Computing Meeting BNL May 24, 1999.
The ATLAS TAGs Database - Experiences and further developments Elisabeth Vinek, CERN & University of Vienna on behalf of the TAGs developers group.
Andrea Valassi (CERN IT-DB)CHEP 2004 Poster Session (Thursday, 30 September 2004) 1 HARP DATA AND SOFTWARE MIGRATION FROM TO ORACLE Authors: A.Valassi,
General requirements for BES III offline & EF selection software Weidong Li.
ATLAS Database Access Library Local Area LCG3D Meeting Fermilab, Batavia, USA October 21, 2004 Alexandre Vaniachine (ANL)
Magda Distributed Data Manager Prototype Torre Wenaus BNL September 2001.
November 1, 2004 ElizabethGallas -- D0 Luminosity Db 1 D0 Luminosity Database: Checklist for Production Elizabeth Gallas Fermilab Computing Division /
Pavel Nevski DDM Workshop BNL, September 27, 2006 JOB DEFINITION as a part of Production.
Summary of User Requirements for Calibration and Alignment Database Magali Gruwé CERN PH/AIP ALICE Offline Week Alignment and Calibration Workshop February.
The MEG Offline Project General Architecture Offline Organization Responsibilities Milestones PSI 2/7/2004Corrado Gatto INFN.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
TAG and iELSSI Progress Elisabeth Vinek, CERN & University of Vienna on behalf of the TAG developers group.
Database Issues Peter Chochula 7 th DCS Workshop, June 16, 2003.
Barthélémy von Haller CERN PH/AID For the ALICE Collaboration The ALICE data quality monitoring system.
Fermilab Scientific Computing Division Fermi National Accelerator Laboratory, Batavia, Illinois, USA. Off-the-Shelf Hardware and Software DAQ Performance.
A Semi-Automated Digital Preservation System based on Semantic Web Services Jane Hunter Sharmin Choudhury DSTC PTY LTD, Brisbane, Australia Slides by Ananta.
CMS High Level Trigger Configuration Management
LHC experiments Requirements and Concepts ALICE
Readiness of ATLAS Computing - A personal view
US ATLAS Physics & Computing
Data Lifecycle Review and Outlook
Nuclear Physics Data Management Needs Bruce G. Gibbard
Module 01 ETICS Overview ETICS Online Tutorials
OO-Design in PHENIX PHENIX, a BIG Collaboration A Liberal Data Model
ATLAS DC2 & Continuous production
Use of GEANT4 in CMS The OSCAR Project
DBOS DecisionBrain Optimization Server
Presentation transcript:

STAR C OMPUTING ROOT in STAR Torre Wenaus STAR Computing and Software Leader Brookhaven National Laboratory, USA ROOT 2000 Workshop, CERN February 3, 2000

STAR C OMPUTING Torre Wenaus, BNL ROOT 2000 Outline STAR and STAR Computing ROOT in STAR  Framework and analysis environment §cf. Valery Fine’s talk  Event data store §cf. Victor Perevoztchikov’s talk l Complemented by MySQL l OO Event persistency Current status

STAR C OMPUTING Torre Wenaus, BNL ROOT 2000 STAR at RHIC RHIC: Relativistic Heavy Ion Collider at Brookhaven National Laboratory  Colliding Au - Au nuclei at 100GeV/nucleon  Principal objective: Discovery and characterization of the Quark Gluon Plasma (QGP)  First year physics run April-August 2000 STAR experiment  One of two large experiments at RHIC, >400 collaborators each l PHENIX is the other  Hadrons, jets, electrons and photons over large solid angle l Principal detector: 4m TPC drift chamber (operational) l Si tracker (year 2) l Pb/scint EM calorimeter (staged over years 1-3)  ~4000+ tracks/event recorded in tracking detectors  High statistics per event permit event by event measurement and correlation of QGP signals

Summer ’99 Engineering run Beam gas event

STAR C OMPUTING Torre Wenaus, BNL ROOT 2000 Computing at STAR Data recording rate of 20MB/sec; 15-20MB raw data per event (~1Hz)  17M Au-Au events (equivalent) recorded in nominal year  Relatively few but highly complex events Requirements:  200TB raw data/year; 270TB total for all processing stages  10,000 Si95 CPU/year  Wide range of physics studies: ~100 concurrent analyses in ~7 physics working groups Principal facility: RHIC Computing Facility (RCF) at Brookhaven  All reconstruction, some analysis, no simulation  20 kSi95 CPU, 50TB disk, 270TB robotic (HPSS) in ’01 Secondary STAR facility: NERSC (LBNL)  Scale similar to STAR component of RCF

STAR C OMPUTING Torre Wenaus, BNL ROOT 2000 STAR Software Environment C++:Fortran ~ 65%:35% in Offline  from ~20%/80% in 9/98 l ~5%/95% in application codes In Fortran:  Most simulation and reconstruction In C++:  All post-reconstruction physics analysis code  New/recent simulation, reconstruction codes  Virtually all infrastructure code ~75 packages Online system in C++  Java GUIs Migratory Fortran => C++ software environment central to STAR offline design

STAR C OMPUTING Torre Wenaus, BNL ROOT 2000 STAR Offline Framework STAR’s legacy Fortran developed in a migration-friendly Framework – StAF – enforcing IDL-based data structures and component interfaces  But proceeding with StAF to a fully OO Framework / analysis environment required major in-house development and support  Instead … leverage a very capable tool from the community 11/98 adopted new Framework built over ROOT  Origins in the ROOT-based ATLFAST  Modular components ‘Makers’ instantiated in a processing chain progressively build (and ‘own’) event components  Supports legacy Fortran and IDL based data structures developed in previous StAF framework without change (automated wrapping)  Deployed in offline production and analysis since RHIC’s second ‘Mock Data Challenge’, 2-3/99 cf. Valery’s talk for details

STAR C OMPUTING Torre Wenaus, BNL ROOT 2000 Event Store Design/Implementation Requirements Flexible partitioning of event components to different streams based on access characteristics Support for both IDL-defined data structures and an OO event model, compatible with Fortran => C++ migration Robust schema evolution (new codes reading old data and vice versa) Easy management and navigation; dynamic addition of event components  Named collections of events, production- and user- defined  Navigation from from a run/event/component request to the data No requirement for on-demand access  Desired event components are specified at start of job, and optimized retrieval of these components is managed for the whole job (Grand Challenge)  On-demand access not compatible with job-level optimization of event component retrieval

STAR C OMPUTING Torre Wenaus, BNL ROOT 2000 Objectivity or ROOT as Event Data Store? Original (1997 RHIC event store task force) STAR choice: Objectivity Prototype Objectivity event store and conditions DB’s deployed Fall ’98  50GB DST event store tested in first Mock Data Challenge  Basically worked OK, but many concerns surrounding Objectivity Decided to develop and deploy ROOT as DST event store in Mock Data Challenge 2 (Feb-Mar ‘99) and make a choice  Performed well; met requirements in multiple data streams for different event components Conclusion: A functional and viable solution for an event store  Easy decision to adopt ROOT over Objectivity  Particularly with CDF’s decision to use ROOT I/O in Run II

STAR C OMPUTING Torre Wenaus, BNL ROOT 2000 Features of ROOT-based Event Store Data organized as named components resident in different files constituting a ‘file family’  Successive processing stages add new components No separation between transient and persistent data structures  Transient object interfaces do not express the persistency mechanism, but the object implementations are directly persistent  Used for our OO/C++ event model StEvent, giving us a direct object store without ROOT appearing in the event model interface Automatic schema evolution implemented  Extension of existing manual ROOT schema evolution c.f. Victor’s talk for details

STAR C OMPUTING Torre Wenaus, BNL ROOT 2000 STAR Event Model StEvent Introduction of C++/OO into STAR: in physics analysis  Until a year ago, essentially no post-reconstruction analysis code intended for real data analysis  A community interested and willing to move to C++/OO  Allowed us to break completely away from Fortran at the DST StEvent C++/OO event model developed  Targeted initially at DST and post-reconstruction analysis  To be extended later – upstream to reconstruction and downstream to micro DSTs – now in progress Requirements reflected in StEvent design:  Event model seen by application codes is “generic C++” – does not express implementation and persistency choices l Developed initially (deliberately) as a purely transient model – no dependencies on ROOT or persistency mechanisms l Later implemented in ROOT to provide direct persistency

STAR C OMPUTING Torre Wenaus, BNL ROOT 2000 ROOT+MySQL Based Data Management Adoption of ROOT I/O for the event store left Objectivity with one role remaining to cover: the true ‘database’ functions of the event store  Navigation among event collections, runs/events, event components  Data locality (now translates basically to file lookup)  Management of dynamic, asynchronous updating of the event store from one end of the processing chain to the other l From initiation of an event collection (run) in online through addition of components in reconstruction, analysis and their iterations  and the conditions database But with the concerns and weight of Objectivity it is overkill for this. So we went shopping…  with particular attention to the rising wave of Internet-driven tools and open software and came up with MySQL

STAR C OMPUTING Torre Wenaus, BNL ROOT 2000 Requirements: STAR 1/00 View (My Version)

STAR C OMPUTING Torre Wenaus, BNL ROOT 2000 MySQL as the STAR Database Relational DB, open software, very fast, widely used on the web Not a full featured heavyweight like Oracle  Relatively simple and fast  No transactions, no unwinding based on journalling Development pace is very fast with a wide range of tools to use  Good interfaces to Perl, C/C++, Java  Easy and powerful web interfacing Great experience so far  Like a quick protyping tool that is also production capable – for the right applications l Metadata and compact data Multiple servers, databases can be used as needed to address scalability, access and locking characteristics

STAR C OMPUTING Torre Wenaus, BNL ROOT 2000 MySQL based applications in STAR File catalogs for simulated and real data  Catalogues ~22k files, ~10TB of data Being integrated with Grand Challenge Architecture (GCA) Production run log used in datataking Event tag database  Good results with preliminary tests of 10M row table, 100bytes/row l 140sec for full SQL query, no indexing (70 kHz) Conditions (constants, geometry, calibrations, configurations) database Production database  Job configuration catalog, job logging, QA, I/O file management Distributed (LAN or WAN) processing monitoring system  Monitors STAR analysis facilities at BNL; planned extension to NERSC Distributed analysis job editing/management system Web-based browsers for all of the above

STAR Databases and Navigation Between Them

STAR C OMPUTING Torre Wenaus, BNL ROOT 2000 Current Status Offline software infrastructure, data management, applications and production operations are operational in production and ‘ready’ to receive year 1 physics data  ‘Ready’ in quotes because there is much essential work still under way l Reconstruction, analysis software, data mining and analysis operations infrastructure including Grand Challenge deployment  We’re in production at year 1 throughput levels this week, in another ‘mini Mock Data Challenge’ exercise Our final Mock Data Challenge prior to physics data is in March  Stress testing analysis software and infrastructure, uDSTs  Target for an operational Grand Challenge ROOT is a vital and robust component through much of STAR software  And is a major reason we’re as ready for data as we are Thank you to the core ROOT team and ROOT developers everywhere!