ROOT-CORE Team 1 PROOF xrootd Fons Rademakers Maarten Ballantjin Marek Biskup Derek Feichtinger (ARDA) Gerri Ganis Guenter Kickinger Andreas Peters (ARDA)

Slides:



Advertisements
Similar presentations
LEAD Portal: a TeraGrid Gateway and Application Service Architecture Marcus Christie and Suresh Marru Indiana University LEAD Project (
Advertisements

FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
Proof/Proof-on-Demand at Atlas T3 Shuwei Ye (BNL) Aug 20, 2014.
June, 20013rd ROOT Workshop1 PROOF and ROOT Grid Features Fons Rademakers.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
1 PROOF & GRID Update Fons Rademakers. 2 Parallel ROOT Facility The PROOF system allows: parallel execution of scripts parallel analysis of trees in a.
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL March 25, 2003 CHEP 2003 Data Analysis Environment and Visualization.
June 21, PROOF - Parallel ROOT Facility Maarten Ballintijn, Rene Brun, Fons Rademakers, Gunter Roland Bring the KB to the PB.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Grids and Globus at BNL Presented by John Scott Leita.
What is adaptive web technology?  There is an increasingly large demand for software systems which are able to operate effectively in dynamic environments.
Makrand Siddhabhatti Tata Institute of Fundamental Research Mumbai 17 Aug
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
PROOF - Parallel ROOT Facility Kilian Schwarz Robert Manteufel Carsten Preuß GSI Bring the KB to the PB not the PB to the KB.
BaBar WEB job submission with Globus authentication and AFS access T. Adye, R. Barlow, A. Forti, A. McNab, S. Salih, D. H. Smith on behalf of the BaBar.
DIRAC Web User Interface A.Casajus (Universitat de Barcelona) M.Sapunov (CPPM Marseille) On behalf of the LHCb DIRAC Team.
AliEn uses bbFTP for the file transfers. Every FTD runs a server, and all the others FTD can connect and authenticate to it using certificates. bbFTP implements.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
Scalable Systems Software Center Resource Management and Accounting Working Group Face-to-Face Meeting June 13-14, 2002.
Grid Job and Information Management (JIM) for D0 and CDF Gabriele Garzoglio for the JIM Team.
Interactive Data Analysis with PROOF Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers CERN.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
GUI For A Virtual Pipeline Simulation Testbed By, Revathi Manni Ranganathan Major Professor: Dr.Virgil Wallentine.
Scalable Systems Software Center Resource Management and Accounting Working Group Face-to-Face Meeting October 10-11, 2002.
INFSO-RI Enabling Grids for E-sciencE Workload Management System Mike Mineter
Can we use the XROOTD infrastructure in the PROOF context ? The need and functionality of a PROOF Master coordinator has been discussed during the meeting.
Evolution of Parallel Programming in HEP F. Rademakers – CERN International Workshop on Large Scale Computing VECC, Kolkata.
1 Marek BiskupACAT2005PROO F Parallel Interactive and Batch HEP-Data Analysis with PROOF Maarten Ballintijn*, Marek Biskup**, Rene Brun**, Philippe Canal***,
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
ROOT for Data Analysis1 Intel discussion meeting CERN 5 Oct 2003 Ren é Brun CERN Distributed Data Analysis.
Trip Report SC’04 Pittsburgh Nov 6-12 Fons Rademakers.
Testing the dynamic per-query scheduling (with a FIFO queue) Jan Iwaszkiewicz.
ROOT and Federated Data Stores What Features We Would Like Fons Rademakers CERN CC-IN2P3, Nov, 2011, Lyon, France.
Design and Implementation of PARK (PARallel Kernel for data fitting) Paul KIENZLE, Wenwu CHEN and Ziwen FU Reflectometry Group.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Data reprocessing for DZero on the SAM-Grid Gabriele Garzoglio for the SAM-Grid Team Fermilab, Computing Division.
1 PROOF The Parallel ROOT Facility Gerardo Ganis / CERN CHEP06, Computing in High Energy Physics 13 – 17 Feb 2006, Mumbai, India Bring the KB to the PB.
Getting started DIRAC Project. Outline  DIRAC information system  Documentation sources  DIRAC users and groups  Registration with DIRAC  Getting.
ARDA Prototypes Andrew Maier CERN. ARDA WorkshopAndrew Maier, CERN2 Overview ARDA in a nutshell –Experiments –Middleware Experiment prototypes (basic.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
Interactive Workflows Branislav Šimo, Ondrej Habala, Ladislav Hluchý Institute of Informatics, Slovak Academy of Sciences.
Summary Distributed Data Analysis Track F. Rademakers, S. Dasu, V. Innocente CHEP06 TIFR, Mumbai.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
PROOF and ALICE Analysis Facilities Arsen Hayrapetyan Yerevan Physics Institute, CERN.
Tanenbaum & Van Steen, Distributed Systems: Principles and Paradigms, 2e, (c) 2007 Prentice-Hall, Inc. All rights reserved DISTRIBUTED SYSTEMS.
A prototype for an extended PROOF What is PROOF ? ROOT analysis model … … on a multi-tier architecture Status New development Prototype based on XRD Demo.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
March, PROOF - Parallel ROOT Facility Maarten Ballintijn Bring the KB to the PB not the PB to the KB.
Super Scaling PROOF to very large clusters Maarten Ballintijn, Kris Gulbrandsen, Gunther Roland / MIT Rene Brun, Fons Rademakers / CERN Philippe Canal.
JAliEn Java AliEn middleware A. Grigoras, C. Grigoras, M. Pedreira P Saiz, S. Schreiner ALICE Offline Week – June 2013.
1 Status of PROOF G. Ganis / CERN Application Area meeting, 24 May 2006.
March 13, 2006PROOF Tutorial1 Distributed Data Analysis with PROOF Fons Rademakers Bring the KB to the PB not the PB to the KB.
ATLAS-specific functionality in Ganga - Requirements for distributed analysis - ATLAS considerations - DIAL submission from Ganga - Graphical interfaces.
Overview, Major Developments, Directions1 ROOT Project Status Major developments Directions NSS05 Conference 25 October Ren é Brun CERN Based on my presentation.
By Nitin Bahadur Gokul Nadathur Department of Computer Sciences University of Wisconsin-Madison Spring 2000.
September, 2002CSC PROOF - Parallel ROOT Facility Fons Rademakers Bring the KB to the PB not the PB to the KB.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Sept. 2000CERN School of Computing1 PROOF and ROOT Grid Features Fons Rademakers.
ROOT and PROOF Tutorial Arsen HayrapetyanMartin Vala Yerevan Physics Institute, Yerevan, Armenia; European Organization for Nuclear Research (CERN)
Core and Framework DIRAC Workshop October Marseille.
PROOF on multi-core machines G. GANIS CERN / PH-SFT for the ROOT team Workshop on Parallelization and MultiCore technologies for LHC, CERN, April 2008.
Dynamic Deployment of VO Specific Condor Scheduler using GT4
PROOF – Parallel ROOT Facility
Grid2Win: Porting of gLite middleware to Windows XP platform
LCG middleware and LHC experiments ARDA project
Support for ”interactive batch”
PROOF - Parallel ROOT Facility
Job Application Monitoring (JAM)
Presentation transcript:

ROOT-CORE Team 1 PROOF xrootd Fons Rademakers Maarten Ballantjin Marek Biskup Derek Feichtinger (ARDA) Gerri Ganis Guenter Kickinger Andreas Peters (ARDA) BaBar

ROOT-CORE Team 2 Original Goals Interactive parallel analysis on local cluster Transparency same selectors, same chain Draw(), etc. on PROOF as in local session Scalability quite good and well understood up to 1000 nodes (most extreme case) extensive monitoring capabilities MLM (Multi-Level-Master) improves scalability on wide area clusters Adaptability partly achieved, system handles varying load on cluster nodes MLM allows much better latencies on wide area clusters not yet coming and going of worker nodes

ROOT-CORE Team 3 New Additional Goals Support for “interactive” batch mode allow submission of long running queries allow client/master disconnect and reconnect Support “hostile” grid environments startup of agents via Grid job scheduler agents calling out to master (firewalls, NAT) dynamic master-worker setup

ROOT-CORE Team 4 Interactive/Batch queries GUI Commands scripts Batch statefull or stateless stateless

ROOT-CORE Team 5 Analysis session snapshot what we want to do AQ1: 1s query produces a local histogram AQ2: a 10mn query submitted to PROOF1 AQ3->AQ7: short queries AQ8: a 10h query submitted to PROOF2 BQ1: browse results of AQ2 BQ2: browse temporary results of AQ8 BQ3->BQ6: submit 4 10mn queries to PROOF1 CQ1: Browse results of AQ8, BQ3->BQ6 Monday at 10h15 ROOT session On my laptop Monday at 16h25 ROOT session On my laptop Wednesday at 8h40 Carrot session on any web browser

ROOT-CORE Team 6 Disconnect / Reconnect Authentication, sessions token State issues Transparency issues TSelector::Begin(), Terminate() objects created in Terminate(), output lists, etc. storage of intermediate results

ROOT-CORE Team 7 Performance Issues Read ahead interface in (x)rootd Using and understanding xrootd Cache index on master, optimizes repeat queries assign different sessions with same data set to same worker nodes, reuse in memory files Monitoring Proofbench Query estimator

ROOT-CORE Team 8 Authentication, Authorization New xrootd authentication plugins Certificates (login and user name) single experiment wide login user name used for sandbox Authorization to sandbox and shared global space not to other user’s sandboxes under same account

ROOT-CORE Team 9 Robustness Get rid of OOB Split in two processes protocol, authentication processor (proofd) proofserv Communicate via priority based message queue Threaded or forked? xrootd code reuse? Periodic output list reporting to the master

ROOT-CORE Team 10 Usability Selector with dynamic variables Selector proxy support (TTree::MakeProxy – better name) Tree friends Event lists Compressed bit slice indices All possible draw options of the Draw() interface

ROOT-CORE Team 11 Usability Browser interface to: proof sessions on demand histograms (single shot, timer driven feedback) monitoring histograms other input/output list objects Stop / resume capability

ROOT-CORE Team 12 Usability Limiting processing to specific part of the data set allow meta data in the TDSet (time, run, conditions, etc) be independent from catalog allows creation of new reduced data sets Remote display package, all canvases created in the master are send back to the client

ROOT-CORE Team 13 Grid Interfacing Grid catalog data set creation meta data, #events, time, run, etc. proofd agent creation agents call out to (no-incoming connection) Config file generation / fully dynamic Coming and going of worker nodes Grid aware packetizer Scheduled execution Limiting processing to specific part of the data set

ROOT-CORE Team 14 Interactive Analysis with PROOF on the Grid PROOF USER SESSION PROOF SLAVE SERVERS PROOF MASTER SERVER PROOF SLAVE SERVERS Guaranteed site access through PROOF Sub-Masters calling out to Master (agent technology) PROOF SUB-MASTER SERVERS PROOF PROOF PROOF Grid/Root Authentication Grid Access Control Service TGrid UI/Queue UI Proofd Startup Grid Service Interfaces Grid File/Metadata Catalogue Client retrieves list of logical files (LFN + MSN)

ROOT-CORE Team 15 PROOF Grid Session Diagram Client GRID query Client sends analysis request to PROOF Master PROOF Sub-Masters register with the PROOF Master Client runs analysis on GRID query dataset PROOF master forwards analysis and collects results ClientMaster PROOF Master populates GRID or batch queue with new slaves or discovers static slaves GRID UI GRID Queue GRID Phase I: Grid MW dependent Phase II: Grid MW independent ! Remote Slaves

ROOT-CORE Team 16 Conclusion We have started an acceleration in the PROOF development towards achieving the new exciting goals that will hugely enhance the data analysis experience of very large data sets A first firm milestone is a demo at SC’05