Analysis infrastructure/framework A collection of questions, observations, suggestions concerning analysis infrastructure and framework Compiled by Marco.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Fox Scientific, Inc. ONLINE ORDERING 101. Welcome to our website On our main page you can find current promotions, the vendors we offer, technical references.
Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
1 User Analysis Workgroup Update  All four experiments gave input by mid December  ALICE by document and links  Very independent.
Distributed Databases John Ortiz. Lecture 24Distributed Databases2  Distributed Database (DDB) is a collection of interrelated databases interconnected.
The LEGO Train Framework
– Unfortunately, this problems is not yet fully under control – No enough information from monitoring that would allow us to correlate poor performing.
Trains status&tests M. Gheata. Train types run centrally FILTERING – Default trains for p-p and Pb-Pb, data and MC (4) Special configuration need to be.
ALICE Operations short summary and directions in 2012 Grid Deployment Board March 21, 2011.
ALICE Operations short summary LHCC Referees meeting June 12, 2012.
ALICE Operations short summary and directions in 2012 WLCG workshop May 19-20, 2012.
Staging to CAF + User groups + fairshare Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE Offline week,
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Computers in the real world Objectives Understand what is meant by memory Difference between RAM and ROM Look at how memory affects the performance of.
The ALICE Analysis Framework A.Gheata for ALICE Offline Collaboration 11/3/2008 ACAT'081A.Gheata – ALICE Analysis Framework.
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 3: Operating-System Structures System Components Operating System Services.
Costin Grigoras ALICE Offline. In the period of steady LHC operation, The Grid usage is constant and high and, as foreseen, is used for massive RAW and.
Workshop on Computing for Neutrino Experiments - Summary April 24, 2009 Lee Lueking, Heidi Schellman NOvA Collaboration Meeting.
240-Current Research Easily Extensible Systems, Octave, Input Formats, SOA.
Infrastructure for QA and automatic trending F. Bellini, M. Germain ALICE Offline Week, 19 th November 2014.
Optimizing CMS Data Formats for Analysis Peerut Boonchokchuay August 11 th,
PWG3 Analysis: status, experience, requests Andrea Dainese on behalf of PWG3 ALICE Offline Week, CERN, Andrea Dainese 1.
Andrei Gheata, Mihaela Gheata, Andreas Morsch ALICE offline week, 5-9 July 2010.
Analysis trains – Status & experience from operation Mihaela Gheata.
5/2/  Online  Offline 5/2/20072  Online  Raw data : within the DAQ monitoring framework  Reconstructed data : with the HLT monitoring framework.
AliRoot survey P.Hristov 11/06/2013. Offline framework  AliRoot in development since 1998  Directly based on ROOT  Used since the detector TDR’s for.
ALICE Operations short summary ALICE Offline week June 15, 2012.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
INFORMATION TECHNOLOGY
The ATLAS Computing Model and USATLAS Tier-2/Tier-3 Meeting Shawn McKee University of Michigan Joint Techs, FNAL July 16 th, 2007.
A. Gheata, ALICE offline week March 09 Status of the analysis framework.
1 Offline Week, October 28 th 2009 PWG3-Muon: Analysis Status From ESD to AOD:  inclusion of MC branch in the AOD  standard AOD creation for PDC09 files.
Summary of User Requirements for Calibration and Alignment Database Magali Gruwé CERN PH/AIP ALICE Offline Week Alignment and Calibration Workshop February.
Data processing Offline review Feb 2, Productions, tools and results Three basic types of processing RAW MC Trains/AODs I will go through these.
M. Gheata ALICE offline week, October Current train wagons GroupAOD producersWork on ESD input Work on AOD input PWG PWG31 (vertexing)2 (+
PWG3 Analysis: status, experience, requests Andrea Dainese on behalf of PWG3 ALICE Offline Week, CERN, Andrea Dainese 1.
PWG3 analysis (barrel)
Analysis experience at GSIAF Marian Ivanov. HEP data analysis ● Typical HEP data analysis (physic analysis, calibration, alignment) and any statistical.
David Adams ATLAS ATLAS Distributed Analysis (ADA) David Adams BNL December 5, 2003 ATLAS software workshop CERN.
Analysis train M.Gheata ALICE offline week, 17 March '09.
M. Gheata ALICE offline week, 24 June  A new analysis train macro was designed for production  /ANALYSIS/macros/AnalysisTrainNew.C /ANALYSIS/macros/AnalysisTrainNew.C.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE Operations: Evolution of the Role of.
David Adams ATLAS ADA: ATLAS Distributed Analysis David Adams BNL December 15, 2003 PPDG Collaboration Meeting LBL.
Analysis framework plans A.Gheata Offline week 13 July 2011.
Some topics for discussion 31/03/2016 P. Hristov 1.
CCR e INFN-GRID Workshop, Palau, Andrea Dainese 1 L’analisi per l’esperimento ALICE Andrea Dainese INFN Padova Una persona attiva come utente.
AliRoot survey: Calibration P.Hristov 11/06/2013.
System Components Operating System Services System Calls.
29/04/2008ALICE-FAIR Computing Meeting1 Resulting Figures of Performance Tests on I/O Intensive ALICE Analysis Jobs.
ANALYSIS TRAIN ON THE GRID Mihaela Gheata. AOD production train ◦ AOD production will be organized in a ‘train’ of tasks ◦ To maximize efficiency of full.
CALIBRATION: PREPARATION FOR RUN2 ALICE Offline Week, 25 June 2014 C. Zampolli.
1 14th June 2012 CPass0/CPass1 status and development.
The ALICE Analysis -- News from the battlefield Federico Carminati for the ALICE Computing Project CHEP 2010 – Taiwan.
Jan Fiete Grosse-Oetringhaus
Data Formats and Impact on Federated Access
SuperB and its computing requirements
Analysis trains – Status & experience from operation
The New World of User Engagement – User Perspective
ALICE Monitoring
Developments of the PWG3 muon analysis code
Production status – christmas processing
Status of the CERN Analysis Facility
ALICE analysis preservation
MICE Collaboration Meeting Saturday 22nd October 2005 Malcolm Ellis
New calibration strategy – follow-up from the production review
Analysis Trains - Reloaded
Data Preparation Group Summary of the Activities
Analysis framework - status
ALICE Computing Model in Run3
Presentation transcript:

Analysis infrastructure/framework A collection of questions, observations, suggestions concerning analysis infrastructure and framework Compiled by Marco van Leeuwen Collected input from experienced analysers Christian KB, Silvia M, Andrea D, Marco v L, Constantin L, Pietro A, Laurent A, and friends Possible action items highlighted in red

Running on the GRID Individual experiences differ significantly –Depends on dataset, format, (person/priviliges) Can run over all (ESD) LHC10h data in one day! p+p MC very fragmented lots of small jobs, cumbersome (‘recent’ option of merging via JDL in plugin helps a lot !) Success rate of jobs low (50-70%) –Also depends on use-case –Resubmission mechanism convenient, but often need >3 resubmissions to reach 90% –Not clear whether there is a single main cause for this Large memory usage is an issue (sometimes >2-3 GB) –drawback of running trains with many tasks? I/O errors –Recent work on R__unzip, xrootd may improve this? Site-issues contribute, but normally transient? (User experiences) Low success rate should be a concern – keep monitoring and improving Ask users for experience/input if needed

Stability of central services/grid connection Update of alien made aliensh more responsive However, still regular hiccups –Always quickly fixed (mail on atf list) –There may be work ongoing to improve this (e.g. reconfiguring/share of pcapiservers) ? It is practically impossible to keep a grid data connection (running merge or a test job on a local machine) open for long (>4 hrs) –Not so clear why; site-issues/R_unzip etc contribute –Related to recent observation of expiring certificates after N bytes? (fix ongoing) –Does this also affect jobs running on the grid? No show-stoppers, but definite room for improvement; need to keep an eye on these issues: –Keep users informed and ask for feedback on the mailing list (e.g.: we changed... let us know if you see improvements/problems) –Suggestion: agree on scheduled maintenance times ?

AOD vs ESD AODs actively used in cases where they have distinct advantages: –Analysis needs small subset of data (e.g. MUON AODs) –Common intermediate analysis objects (jets, Heavy Flavour) Other (‘minbias’) analyses often use ESD: –Analysis code development sometimes started on ESD –More info available; more cross-checks possible –For some analyses, essential info is (not yet) in AOD Hybrid situation is ‘natural’ –Different analyses have different needs –Putting ‘everything’ in the AOD would be counter-productive ! In mean time: keep encouraging people to use AODs Requires regular re-production of AODs and more information to the users about new sets and features What is the forum to discuss AOD content and (re)production schedule?  Andreas’ Analysis and QA meeting?  re-establish Analysis discussion during Offline Weekly? Fixed time slot? Many analyses need only a small subset of ESD info AODs: selected info to reduce resource usage

Analysis trains Single ‘central’ analysis train (or 1 per PWG) probably not practical/manageable –Too many carts –Different analyses run over different inputs Currently many analysis trains exist; mostly organised around ‘common need’ –This is probably the natural situation –Train operators concentrate knowledge/experience Would like to have a forum to exchange this experience –Train operators could/should have special priviliges Being realised via ‘official’ PWG trains? Central trains: –PWG1/QA train –Common AOD production Suggestion: set up monthly meeting of PWG analysis software coordinators and train operators Forum to exchange experiences and report common issues Need to refresh (and publicize?) list of PWG analysis software coordinators?

How to keep the information flowing? Only way to ‘reach everyone’ is via mail (or website if people know where to look) –If you have/know something of general use, send it to the list; need to encourage everyone ! –Especially important for announcements (from offline or others) Alice-analysis-taskforce list works well –Only down-side: it is a firehose, with many topics/mails ALICE A web of interconnected groups/meetings + central mailing lists, website(s) Suggestion: separate operational (day-to-day issues) from ‘high-level’ questions/discussion Set up new mailing list for analysis software discussion

Next slide: bonus track

AOD and ESD consistency, tender and repass Reality: number of improvements in software + calibration since 2010 data were produced. To make those available for analysis: –New reconstruction pass? Most straightforward option, but (too?) expensive in terms of CPU and/or storage –Alternative: use tender (only fixes a subset of issues?) Option I: produce AODs with tender Option II: use ESD+tender in analysis –Or: use existing ESD+tender to produce new ESD Need input for further discussion: CPU+storage impact for different options Amount of human effort for each option NB: Hopefully, the current situation is specific to ‘first-year’ data. No guarantees, though (expect sw+calibs to improve further over the years) Is there a rule/paradigm that info ESD and AOD should be consistent? Can we afford to stick to it, or need rethink?