Jerome Lauret – ROOT Workshop 2005, CERN, CH ROOT4STAR: a ROOT based framework for user analysis and data mining J é r ô me Lauret -

Slides:



Advertisements
Similar presentations
Adding scalability to legacy PHP web applications Overview Mario A. Valdez-Ramirez.
Advertisements

1 Databases in ALICE L.Betev LCG Database Deployment and Persistency Workshop Geneva, October 17, 2005.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
O. Stézowski IPN Lyon AGATA Week September 2003 Legnaro Data Analysis – Team #3 ROOT as a framework for AGATA.
CLEO’s User Centric Data Access System Christopher D. Jones Cornell University.
ROOT An object oriented HEP analysis framework.. Computing in Physics Physics = experimental science =>Experiments (e.g. at CERN) Planning phase Physics.
2/10/2000 CHEP2000 Padova Italy The BaBar Online Databases George Zioulas SLAC For the BaBar Computing Group.
Click to edit Master title style 21 st Nov, 2003 Valeri Fine (BNL) ROOT and ROOT4STAR framework V. Fine or Software for “C++ speaking” people.
Zubanov Alexei, 2006 Aug 07 – Sep 28 QtROOT package extension with Coin3D lib.
STAR C OMPUTING Maker and I/O Model in STAR Victor Perevoztchikov.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
The GlueX Collaboration Meeting October 4-6, 2012 Jefferson Lab Curtis Meyer.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
100 Million events, what does this mean ?? STAR Grid Program overview.
Java Root IO Part of the FreeHEP Java Library Tony Johnson Mark Dönszelmann
Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May STAR Computing Doug Olson for: J é r ô me Lauret.
STAR Analysis Meeting, BNL, Dec 2004 Alexandre A. P. Suaide University of Sao Paulo Slide 1 BEMC software and calibration L3 display 200 GeV February.
STAR C OMPUTING STAR Computing Infrastructure Torre Wenaus BNL STAR Collaboration Meeting BNL Jan 31, 1999.
Root based event display Dmitry Romanov October 19, 2010.
5 May 98 1 Jürgen Knobloch Computing Planning for ATLAS ATLAS Software Week 5 May 1998 Jürgen Knobloch Slides also on:
NOVA Networked Object-based EnVironment for Analysis P. Nevski, A. Vaniachine, T. Wenaus NOVA is a project to develop distributed object oriented physics.
Tracker data quality monitoring based on event display M.S. Mennea – G. Zito University & INFN Bari - Italy.
Grand Challenge and PHENIX Report post-MDC2 studies of GC software –feasibility for day-1 expectations of data model –simple robustness tests –Comparisons.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
CMS pixel data quality monitoring Petra Merkel, Purdue University For the CMS Pixel DQM Group Vertex 2008, Sweden.
STAR Collaboration Meeting, BNL, Feb 2005 Alexandre A. P. Suaide University of Sao Paulo Slide 1 BEMC software update L3 display 200 GeV February.
Virtual Monte Carlo and new geometry description in STAR Maxim Potekhin STAR Collaboration Meeting, BNL July 17, 2004 July 17, 2004.
Analysis trains – Status & experience from operation Mihaela Gheata.
STAR Event data storage and management in STAR V. Perevoztchikov Brookhaven National Laboratory,USA.
STAR Collaboration, July 2004 Grid Collector Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani Lawrence Berkeley National.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
STAR C OMPUTING Plans for Production Use of Grand Challenge Software in STAR Torre Wenaus BNL Grand Challenge Meeting LBNL 10/23/98.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
V.Fine ALICE-STAR Joint meeting April, 9, 2000 The STAR offline framework* V. Fine *) See also:
Computing R&D and Milestones LHCb Plenary June 18th, 1998 These slides are on WWW at:
Data Communications and Networks Chapter 9 – Distributed Systems ICT-BVF8.1- Data Communications and Network Trainer: Dr. Abbes Sebihi.
V.Fine for STAR collaboration The STAR offline framework and 3D visualization V. Fine, Y. Fisyak, V. Perevoztchikov, T.Wenaus.
Online Monitoring System at KLOE Alessandra Doria INFN - Napoli for the KLOE collaboration CHEP 2000 Padova, 7-11 February 2000 NAPOLI.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Super Scaling PROOF to very large clusters Maarten Ballintijn, Kris Gulbrandsen, Gunther Roland / MIT Rene Brun, Fons Rademakers / CERN Philippe Canal.
Status & development of the software for CALICE-DAQ Tao Wu On behalf of UK Collaboration.
The ATLAS DAQ System Online Configurations Database Service Challenge J. Almeida, M. Dobson, A. Kazarov, G. Lehmann-Miotto, J.E. Sloper, I. Soloviev and.
October Test Beam DAQ. Framework sketch Only DAQs subprograms works during spills Each subprogram produces an output each spill Each dependant subprogram.
S.Linev: Go4 - J.Adamczewski, H.G.Essel, S.Linev ROOT 2005 New development in Go4.
The MEG Offline Project General Architecture Offline Organization Responsibilities Milestones PSI 2/7/2004Corrado Gatto INFN.
Alignment in real-time in current detector and upgrade 6th LHCb Computing Workshop 18 November 2015 Beat Jost / Cern.
V.Fine for STAR collaboration The STAR offline framework V. Fine, Y. Fisyak, V. Perevoztchikov, T.Wenaus.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
ATLAS Distributed Analysis Dietrich Liko IT/GD. Overview  Some problems trying to analyze Rome data on the grid Basics Metadata Data  Activities AMI.
The ATLAS Computing & Analysis Model Roger Jones Lancaster University ATLAS UK 06 IPPP, 20/9/2006.
STAR Persistent Pointers in the STAR Micro-DST V. Perevoztchikov Brookhaven National Laboratory,USA.
Production Mode Data-Replication Framework in STAR using the HRM Grid CHEP ’04 Congress Centre Interlaken, Switzerland 27 th September – 1 st October Eric.
StEvent I/O Model And Writing a Maker Or How to Add a New Detector Akio Ogawa BNL 2003 Nov Dubna.
Proposal for the MEG Offline System Assisi 9/21/2004Corrado Gatto General Architecture Computing Model Organization & Responsibilities Milestones.
Pavel Nevski STAR simulations GSTAR framework OO geometry/event model NOVA components.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Update on CHEP from the Computing Speaker Committee G. Carlino (INFN Napoli) on behalf of the CSC ICB, October
1 SLAC simulation workshop, May 2003 Ties Behnke Mokka and LCDG4 Ties Behnke, DESY and SLAC MOKKA: european (france) developed GEANT4 based simulation.
MAUS Status A. Dobbs CM43 29 th October Contents MAUS Overview Infrastructure Geometry and CDB Detector Updates CKOV EMR KL TOF Tracker Global Tracking.
Barthélémy von Haller CERN PH/AID For the ALICE Collaboration The ALICE data quality monitoring system.
(on behalf of the POOL team)
CMS High Level Trigger Configuration Management
A C++ generic model for the GLAST Geometric Description
Nuclear Physics Data Management Needs Bruce G. Gibbard
Support for ”interactive batch”
OO-Design in PHENIX PHENIX, a BIG Collaboration A Liberal Data Model
Offline framework for conditions data
Presentation transcript:

Jerome Lauret – ROOT Workshop 2005, CERN, CH ROOT4STAR: a ROOT based framework for user analysis and data mining J é r ô me Lauret -

Jerome Lauret – ROOT Workshop 2005, CERN, CH Introduction This should really have been a talk with a title such as “How ROOT helps an experiment with its data taking and analysis – Real life experience from PBytes experiments” ROOT is HEAVILY used in RHIC/STAR  In fact, all RHIC experiments use it  It is used to build frameworks for standard or distributed computing  It is versatile enough to allow developments (QtRoot, …)  It is stable enough that we use it from online to offline … But before more ROOT commercials …

Jerome Lauret – ROOT Workshop 2005, CERN, CH RHIC facility... The Solenoidal Tracker At RHIC is an experiment located at the Brookhaven National Laboratory (BNL), USA

Jerome Lauret – ROOT Workshop 2005, CERN, CH STAR experiment... The Solenoidal Tracker At RHIC  A collaboration of 616 people wide, spanning over 12 countries for a total of 52 institutions  A Pbytes scale experiment overall (raw+reconstructed) with several Million of files The Physics  A multi-purpose detector system for Heavy Ion and p+p program

Jerome Lauret – ROOT Workshop 2005, CERN, CH First use of ROOT – Online Right from the very moment of a collision, STAR computing uses ROOT to ● Provide Online QA ● Monitoring ● Event display Standard approach here again … ● GUI are easy to build … ● Interfacing event-Pools and ROOT histogram is not complicated … Trigger RO Evt Builder Evt Builder Trigger

Jerome Lauret – ROOT Workshop 2005, CERN, CH Event display Such images are produced minutes after collision. During the run, collision “movies” are available from the Web for public display … More exotic We had the events displayed on the façade in Munich “live”

Jerome Lauret – ROOT Workshop 2005, CERN, CH The Offline STAR framework

Jerome Lauret – ROOT Workshop 2005, CERN, CH Framework in STAR - root4star STAR had a few phases of framework  Early adoption of ROOT ~ – V. Fine / V. Perevotchikov  ROOT4STAR used for ALL STAR published analysis to date ROOT Provides a lot …  Histogram and fitting, Graphics (2D, 3D), IO, NTuples/Trees  Collection of classes, schema evolution, stability  UI Browsers, Panels, Tree Viewers Buildt-in OO model, C++ As many features as one needs for a start … So the early adoption …

Jerome Lauret – ROOT Workshop 2005, CERN, CH root4star or ROOT? ROOT4STAR = ROOT+STAR specific  All ROOT features  Interfaces with G3 (strong linking)  STAR additional base Classes (TTable,...)  Qt based GUIs and Event display ... A single framework for  Simulation  Data mining  User analysis Physics ready micro-DST DO NOT need the framework (root+few libs suffices) Even works on a laptop running Windows ;-)

Jerome Lauret – ROOT Workshop 2005, CERN, CH The two ways to see a framework …

Jerome Lauret – ROOT Workshop 2005, CERN, CH In real life (hopefully the user end) The STAR code is  Codes deriving from a base class StMaker Arranged into libraries, blocks are components Makers are loaded via dynamic libraries User can start from template example  A hierarchical collection of STAR Makers handling data-sets  A single instance of a “chain”, a “steering” component All dependencies in one another sorted NO NEED TO KNOW from users A few special makers  IOMaker, handles all IO - Persistent event model StEvent  Messenger manages all messages (do not prevent cout / printf)  DBMaker manages transparently all DB related access (event timestamp based)

Jerome Lauret – ROOT Workshop 2005, CERN, CH root4star A TDataSet class  from which data sets and makers inherit  allows the construction of hierarchical organizations of components and data  centralizes almost all common tasks Data set navigation, IO, database, inter-component communication

Jerome Lauret – ROOT Workshop 2005, CERN, CH root4star STAR framework designed to support  chained components  can themselves be composite sub-chains  components (“makers”) managing “datasets” they have created and are responsible for  “makers” can communicate and Get/Add global datasets Usual OO approach Base class have common methods Init() InitRun() Finish() FinishRun() Make() … Presented in past ROOT Workshops and CHEP conference …

Jerome Lauret – ROOT Workshop 2005, CERN, CH OO analogy … ROOT TObject, … ROOT4STAR TDataSets, StMaker…

Jerome Lauret – ROOT Workshop 2005, CERN, CH Under the hood … class StMaker : public TDataSet{ public: StMaker(const char *name="",const char *dummy=0); virtual ~StMaker(); virtual Int_t IsChain() const {return 0;} /// User defined functions virtual void Clear(Option_t *option=""); virtual Int_t InitRun(int runumber); virtual Int_t Init(); virtual void StartMaker(); virtual Int_t Make(); virtual Int_t Finish(); virtual Int_t FinishRun(int oldrunumber); // Get methods virtual TDataSet *GetData(const char *name, const char* dir=".data") const; virtual TDataSet *GetDataSet (const char* logInput) const {return FindDataSet(logInput);} virtual Int_t GetEventNumber() const ; virtual Int_t GetRunNumber() const ; virtual TDatime GetDateTime() const; virtual Int_t GetDate() const ; virtual Int_t GetTime() const ; … For example, methods to handle timestamps (simulation or reco)

Jerome Lauret – ROOT Workshop 2005, CERN, CH Why ROOT4STAR instead of ROOT? Good question … In principle, we are (were) only tight to G3 Evaluated VMC approach  Got rid of legacy codes using (yes  ) c-blocks  Introduced TGeo based geometry Currently made using g2root from G3 Geo comparison evaluated – seems to do  Shaping a common IO model for data / simulation VMC means we retire our FZ format New model was already there … We will have news by CHEP 06

Jerome Lauret – ROOT Workshop 2005, CERN, CH Geometry user front end … We like XML ;-)  Worked with AGDD and GraXML  Note that our current geometry do not make use of db parameters to “re-scale” (perfect geometry)  Basic [initial] idea: [?]GDD to TGeo So far  Have basic description in place  Missing “many” (still), polyhedral, negative dimension volume (expanding) Next  Putting a regression test in place  Porting and testing the full geometry

Jerome Lauret – ROOT Workshop 2005, CERN, CH The Offline STAR data & model

Jerome Lauret – ROOT Workshop 2005, CERN, CH Analysis... DAQ DST/event MuDST production Analysis Nano or Pico DST, ROOT trees.. ALL is classic-root based … Possible Further data reduction …

Jerome Lauret – ROOT Workshop 2005, CERN, CH Data Sets sizes Raw Data Size  <> ~ 2-3 MB/event - All on Mass Storage (HPSS as MSS)  Needed only for calibration, production – Not centrally or otherwise stored Real Data size  Data Summary Tape+QA histos+Tags+run information and summary: <> ~ 2-3 MB/event  Micro-DST: KB/event Total Year4 (example) DAQ DST/event MuDST production Analysis

Jerome Lauret – ROOT Workshop 2005, CERN, CH Data Sets sizes - Tier0 Projections Note – Reconstructed ~ raw size 2.2 reconstructed passes … Huge storage demand (even greater processing)

Jerome Lauret – ROOT Workshop 2005, CERN, CH Distributed Disks = SE attached to specific CE – ROOTD allows accessing those … Where does this data go ?? D D DD DataCarousel Client Script adds records Pftp on local disk Complex FileCatalog Management Processing nodes = DD nodes STARTED WITH A VERY HOMEMADE VERY “STATIC” MODEL ROOTD A purely economic model … DD cost 1/5 to 1/10 th of CD

Jerome Lauret – ROOT Workshop 2005, CERN, CH Distributed Disks – Xrootd era D D DD Pftp on local disk from Xrootd In the process of replacing “it” with XROOTD, XROOT+SRM XROOTD provides load balancing, scalability, a way to avoid LFN/PFN translation ● Deployed on 380 nodes (biggest Xrootd usage?) ● Needed to wait for the 64 node limitations removal (reported in February, available ~ in April/May) ● Warning: moving from ROOTD to Xrootd is not as trivial as it seems in ROOT 4.xx.xx … ● Different security model, shaky initial implementation [Should be fixed - Gerri] ● ROOTD does only PFN, Xrootd cannot do both PFN and LFN [in progress??] ● Several patches sent to the Xrootd team

Jerome Lauret – ROOT Workshop 2005, CERN, CH Xrootd stability Xrootd provides load balancing, scalability. A way to avoid LFN/PFN translation..  Initial version was not stable enough for us (end of 2004) …  Needed to wait for the 64 nodes limitation removal (reported in February, available ~ in April/May)  Warning: moving from ROOTD to Xrootd is not as trivial as it seems in ROOT 4.xx.xx … Different security model, erroneous initial implementation [Should be fixed - Gerri] ROOTD does only PFN, Xrootd cannot do both PFN and LFN [in progress??]  Several patches sent to the Xrootd team Deployed on 380 nodes (biggest Xrootd usage?)  Stable enough now for sure …  We are ready to go along ROOTD -> XROOTD ALL THE WAY !!!!

Jerome Lauret – ROOT Workshop 2005, CERN, CH GridCollector Usage in STAR – c.f. Kesheng/John Wu  Based on TAGS produced at reco time  Rest on now well tested and robust SRM (DRM+HRM) deployed in STAR anyhow Immediate Access and managed SE Files moved transparently by delegation to SRM service  Easier to maintain, prospects are enormous “Smart” IO-related improvements and home- made formats no faster than using GridCollector (a priori) Physicists could get back to physics And STAR technical personnel better off supporting GC It is a WORKING prototype of  Grid interactive analysis framework  VERY POWERFULL Event “server” based (no longer files) root4star -b -q doEvents.C'(25,"select MuDst where Production=P04ie \ and trgSetupName=production62GeV and magScale=ReversedFullField \ and chargedMultiplicity>3300 and NV0>200", "gc,dbon")'

Jerome Lauret – ROOT Workshop 2005, CERN, CH Summary Is ROOT great ??  I think it fair to say YES  Fair to mention that I had at maximum 4 “major” issues in 4 years (resolved within 24 hours, special thanks to Philippe, Rene, …) Outstanding support and outstanding team  Fait to say that it provided all features needed for data processing, framework, analysis, … Could it be better ? Always …  We need more display capabilities Qt like work important (to us and other efforts like QScan, …) Geometry work (Andrei, …) should be a priority …  We need MORE distributed computing aware capabilities / Grid Managing SE (Xrootd, perhaps GridCollector …) Managing CE, … Ideally, event based a-la-GridCollector is necessary (in the plans?)