April HEPCG Workshop 2006 GSI

Slides:



Advertisements
Similar presentations
1 ALICE Grid Status David Evans The University of Birmingham GridPP 14 th Collaboration Meeting Birmingham 6-7 Sept 2005.
Advertisements

Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
June, 20013rd ROOT Workshop1 PROOF and ROOT Grid Features Fons Rademakers.
PROOF and AnT in PHOBOS Kristjan Gulbrandsen March 25, 2004 Collaboration Meeting.
The ALICE Framework at GSI Kilian Schwarz ALICE Meeting August 1, 2005.
1 PROOF & GRID Update Fons Rademakers. 2 Parallel ROOT Facility The PROOF system allows: parallel execution of scripts parallel analysis of trees in a.
The German HEP Community Grid for the German HEP Community Grid 27-March-2007, ISGC2007, Taipei Agenda: D-Grid in context HEP Community.
1 Status of the ALICE CERN Analysis Facility Marco MEONI – CERN/ALICE Jan Fiete GROSSE-OETRINGHAUS - CERN /ALICE CHEP Prague.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
PROOF - Parallel ROOT Facility Kilian Schwarz Robert Manteufel Carsten Preuß GSI Bring the KB to the PB not the PB to the KB.
CHEP – Mumbai, February 2006 The LCG Service Challenges Focus on SC3 Re-run; Outlook for 2006 Jamie Shiers, LCG Service Manager.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
DataGrid Applications Federico Carminati WP6 WorkShop December 11, 2000.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
PROOF Cluster Management in ALICE Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE CAF / PROOF Workshop,
ROOT for Data Analysis1 Intel discussion meeting CERN 5 Oct 2003 Ren é Brun CERN Distributed Data Analysis.
The ALICE Computing F.Carminati May 4, 2006 Madrid, Spain.
Status of PDC’07 and user analysis issues (from admin point of view) L. Betev August 28, 2007.
ARDA Prototypes Andrew Maier CERN. ARDA WorkshopAndrew Maier, CERN2 Overview ARDA in a nutshell –Experiments –Middleware Experiment prototypes (basic.
FP6−2004−Infrastructures−6-SSA E-infrastructure shared between Europe and Latin America High Energy Physics Applications in EELA.
DataGrid is a project funded by the European Commission under contract IST rd EU Review – 19-20/02/2004 WP8 - Demonstration ALICE – Evolving.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Super Scaling PROOF to very large clusters Maarten Ballintijn, Kris Gulbrandsen, Gunther Roland / MIT Rene Brun, Fons Rademakers / CERN Philippe Canal.
September, 2002CSC PROOF - Parallel ROOT Facility Fons Rademakers Bring the KB to the PB not the PB to the KB.
AliEn2 and GSI batch farm/disks/tape Current status Kilian Schwarz.
Enabling Grids for E-sciencE CMS/ARDA activity within the CMS distributed system Julia Andreeva, CERN On behalf of ARDA group CHEP06.
ALICE Computing Status1 ALICE Computing Status Are we ready? What about our choices? Workshop on LHC Computing 26 October Ren é Brun CERN Several slides.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
Sept. 2000CERN School of Computing1 PROOF and ROOT Grid Features Fons Rademakers.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
ALICE Physics Data Challenge ’05 and LCG Service Challenge 3 Latchezar Betev / ALICE Geneva, 6 April 2005 LCG Storage Management Workshop.
The German HEP-Grid initiative for the German HEP Community Grid 13-Feb-2006, CHEP06, Mumbai Agenda: D-Grid in context.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Monthly video-conference, 18/12/2003 P.Hristov1 Preparation for physics data challenge'04 P.Hristov Alice monthly off-line video-conference December 18,
Availability of ALICE Grid resources in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
Kilian Schwarz ALICE Computing Meeting GSI, October 7, 2009
Massimo Sgaravatto INFN Padova
ALICE Computing Data Challenge VI
Workload Management Workpackage
(Prague, March 2009) Andrey Y Shevel
ARDA-ALICE activity in 2005 and tasks in 2006
ALICE and LCG Stefano Bagnasco I.N.F.N. Torino
Blueprint of Persistent Infrastructure as a Service
Design rationale and status of the org.glite.overlay component
The “Understanding Performance!” team in CERN IT
Status of the CERN Analysis Facility
PROOF – Parallel ROOT Facility
BOSS: the CMS interface for job summission, monitoring and bookkeeping
INFN-GRID Workshop Bari, October, 26, 2004
Alice DC Status P. Cerello March 19th, 2004.
GSIAF & Anar Manafov, Victor Penso, Carsten Preuss, and Kilian Schwarz, GSI Darmstadt, ALICE Offline week, v. 0.8.
ALICE Physics Data Challenge 3
ALICE – Evolving towards the use of EDG/LCG - the Data Challenge 2004
Grid2Win: Porting of gLite middleware to Windows XP platform
BOSS: the CMS interface for job summission, monitoring and bookkeeping
AliRoot status and PDC’04
MC data production, reconstruction and analysis - lessons from PDC’04
Quattor Usage at Nikhef
Simulation use cases for T2 in ALICE
RGLite A gLite plug-in for ROOT
Partner: LMU (Atlas), GSI (Alice)
LCG middleware and LHC experiments ARDA project
Distributed Data Bases & GRID Interactive Access
ExaO: Software Defined Data Distribution for Exascale Sciences
Kristjan Gulbrandsen March 25, 2004 Collaboration Meeting
Support for ”interactive batch”
Use of Geant4 in experiment interactive frameworks AliRoot
Wide Area Workload Management Work Package DATAGRID project
Short to middle term GRID deployment plan for LHCb
Presentation transcript:

K.Schwarz@gsi.de April 27 2006 HEPCG Workshop 2006 GSI The German HEP-Grid initiative HEP CG WP3: Distributed Data Analysis Tools for ALICE K.Schwarz@gsi.de April 27 2006 HEPCG Workshop 2006 GSI

HEP CG WP3: Distributed interactive data analysis Coordination P.Malzacher , GSI (LMU, GSI, unfunded: LRZ, MPI M, RZ Garching, Uni Karlsruhe, MPI Heidelberg) Optimize application specific job scheduling Analyse and test of software environment required Job management and Bookkeeping of distributed analysis Distribution of analysis, sum-up of results Interactive Analysis: Creation of dedicated analysis clusters Dynamic partitioning of Grid analysis clusters

Motivation: ALICE Computing The main aim of this study is to investigate whether we can use “ALICE–like” interactive analysis in the Grid environment without requiring AliEn everywhere. As a good basis we took gLite middleware, since it is going to be general middleware for LHC experiments in near future (in terms of EGEE project). In the same time we try to proof that the same can be done with other types of Grid middleware like Globus4, for instance.

SC4 and PDC06 April/May: Generate 1 M Pb-Pb and 100 M p-p events on 7500 KSI2k of cpu, 20% at CERN, 29% at T1s (11% GridKa), 51% T2. Network/reconstruction stress test July: repeat network/reconstruction stress test September: scheduled analysis test, user analysis

ALICE: A simple environment ROOT AliRoot STEER Virtual MC G3 G4 FLUKA HIJING MEVSIM PYTHIA6 PDF EVGEN HBTP HBTAN ISAJET EMCAL ZDC ITS PHOS TRD TOF RICH ESD AliAnalysis AliReconstruction PMD CRT FMD MUON TPC START STRUCT AliSimulation AliEn AliEn AliRoot & Co WLCG OSG NGDF

ALICE Analysis concepts Analysis Models Prompt analysis at T0 using PROOF (+file catalogue) infrastructure; batch Analysis using Grid infrastructure; interactive Analysis using PROOF (+Grid) infrastructure. PROOF/ROOT Single-/Multi- tier static and dynamic PROOF cluster; Grid API class TGrid (virtual) ---> TAliEn (implementation). User Interface ALICE users access any Grid Infrastructure via AliEn or ROOT/PROOF UI. AliEn Native and “Grid on a Grid” (LCG/EGEE, ARC, OSG); integrate as much as possible common components LFC, FTS, WMS, MonALISA ...

Tasks GAP Analysis Application specific job scheduling Interactive Analysis ... We decided to concentrate first on “Interactive Analysis”

Start with Gap Analysis Analysis based on PROOF Investigating different versions of PROOF clusters Connect ROOT and gLite: TGlite Developing a ROOT interface for gLite in Poster Session (K. Schwarz) class TGrid : public TObject { public: … virtual TGridResult *Query ( … static TGrid *Connect ( const char *grid, const char *uid = 0, const char *pw = 0 ClassDef(TGrid,0) };

Parallel Analysis of Event Data #proof.conf slave node1 slave node2 slave node3 slave node4 Local PC Remote PROOF Cluster proof proof = master server root stdout/obj proof proof = slave server ana.C proof proof proof *.root TFile node1 ana.C TNetFile *.root $ root root [0] tree.Process(“ana.C”) $ root root [0] tree.Process(“ana.C”) root [1] gROOT->Proof(“remote”) root [2] dset->Process(“ana.C”) $ root root [0] tree.Process(“ana.C”) root [1] gROOT->Proof(“remote”) $ root node2 TFile *.root node3 TFile *.root node4

Various possibilities to set up a PROOF Cluster Static: proofd started via xinetd on dedicated PROOF Cluster (e.g ALICE PROOF Analysis facility @ CERN) Dynamic: proofd integrated in local batch farm ( GSI) Dynamic 2: proofd being sent as Grid Jobs to sites were the data to be analysed can be found (query of FC) realized: PROOF/AliEn soon to be: PROOF/gLite necessary for this: ROOT/Grid communication via TGridXXXX interface provided by ROOT

Gap Analysis To achieve our aim we need to get access to a gLite testbed or install it ourselves; investigate available gLite API and tools; implement TGridXXXX ROOT interfaces for gLite; try to reproduce “ALICE-like” analysis using our new basis; learn and understand the tools; investigate possible gaps in TGrid and in gLite.

Project timeline For testing purposes a complete gLite 1.5 testbed including central services has been installed on virtual Xen hosts at GSI A first alpha release of TGliteXXXX implementation is expected in the end of March 2006.

Set of ROOT interfaces to the Grid TGrid Abstract base class defining interface to common Grid services. TGridResult Abstract base class defining interface to a Grid result. Objects of this class are created by TGrid methods. TGridJob Pure Abstract class. Abstract base class defining interface to a Grid job. TGridJDL Pure Abstract class. To generate JDL files for job submission to the Grid. TGridJobStatus Pure Abstract class. Abstract base class containing the status of a Grid job. TGridCollection Class which manages collection files on the Grid. TAlienXXXX TGLiteXXXX “realized”

Grid Access using ROOT facility Abstract Base Classes UI to Grid Service + Files Grid Plug-in Classes Alien API Service TGrid::Connect(“alien://...) TGrid TAlien gLite C++ API TGrid TGLite TGrid::Connect(“glite://...) Globus4 API TGrid::Connect(“globus4...) TGrid TGlobus4 Future development ? // Connect TGrid grid = TGrid::Connect(“alien://”); // Connect(“glite…” // Query TGridResult *res = grid.Query(“/home/test_user/analysis/”, ”*.root“); // List of files TList *listf = res->GetFileInfoList(); // Create chain TChain chain(“Events", “session"); Chain.AddFileInfoList(listf); // Start PROOF TProof proof(“remote”); // Process your query Chain.Process(“selector.C”); TGrid example with AliEn/gLite

Application specific job scheduling Still in an early stage Need to get more experience with applications first More interactive applications in the LHC environment need to be in running state Applications need to be able to get monitoring information.

Application specific scheduling: basic ideas Monitor resources (CPU, Main Memory, I/O and network traffic) on an event type and analysis type basis. Use this information for forecasting in scheduling decisions. Application needs to be able to access Grid monitoring information on the fly.

Application specific scheduling Potential synergy effects WP1: scheduling  HEP-CG Scheduling Architecture (Lars Schley) WP2: monitoring  talks of Ralph Müller-Pfefferkorn  Job Execution Monitor (Markus Mechtel)  Online steering of HEP applications (Daniel Lorenz)

Status and Outview GAP Analysis: advanced Interactive Analysis: advanced Application specific job scheduling: started Application specific job scheduling will come after Interactive Analysis due to non existing standard monitoring interface in the middleware.