Transformation System report Luisa Arrabito 1, Federico Stagni 2 1) LUPM CNRS/IN2P3, France 2) CERN 5 th DIRAC User Workshop 27 th – 29 th May 2015, Ferrara.

Slides:



Advertisements
Similar presentations
RunJob in CMS Greg Graham Discussion Slides. RunJob in CMS RunJob is an Application Configuration and Job Creation Tool –RunJob uses metadata to abstract.
Advertisements

1 Microsoft Access 2002 Tutorial 9 – Automating Tasks With Macros.
VAMDC Registry Portal Proof of Concept. Registry VAMDC Registry is available at – ex.jsp
1 Grid services based architectures Growing consensus that Grid services is the right concept for building the computing grids; Recent ARDA work has provoked.
Evaluation of NoSQL databases for DIRAC monitoring and beyond
Stuart K. PatersonCHEP 2006 (13 th –17 th February 2006) Mumbai, India 1 from DIRAC.Client.Dirac import * dirac = Dirac() job = Job() job.setApplication('DaVinci',
Asynchronous Web Services Approach Enrique de Andrés Saiz.
Bookkeeping data Monitoring info Get jobs Site A Site B Site C Site D Agent Production service Monitoring service Bookkeeping service Agent © Andrei Tsaregorodtsev.
Pilots 2.0: DIRAC pilots for all the skies Federico Stagni, A.McNab, C.Luzzi, A.Tsaregorodtsev On behalf of the DIRAC consortium and the LHCb collaboration.
The ATLAS Production System. The Architecture ATLAS Production Database Eowyn Lexor Lexor-CondorG Oracle SQL queries Dulcinea NorduGrid Panda OSGLCG The.
Web Application Architecture and Communication. Displaying a Web page in a Browser
K. Harrison CERN, 20th April 2004 AJDL interface and LCG submission - Overview of AJDL - Using AJDL from Python - LCG submission.
Marianne BargiottiBK Workshop – CERN - 6/12/ Bookkeeping Meta Data catalogue: present status Marianne Bargiotti CERN.
David Adams ATLAS AJDL: Analysis Job Description Language David Adams BNL December 15, 2003 PPDG Collaboration Meeting LBL.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
Bookkeeping Tutorial. Bookkeeping & Monitoring Tutorial2 Bookkeeping content  Contains records of all “jobs” and all “files” that are created by production.
DIRAC Review (13 th December 2005)Stuart K. Paterson1 DIRAC Review Exposing DIRAC Functionality.
Stuart Wakefield Imperial College London Evolution of BOSS, a tool for job submission and tracking W. Bacchi, G. Codispoti, C. Grandi, INFN Bologna D.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
ALICE, ATLAS, CMS & LHCb joint workshop on
Giuseppe Codispoti INFN - Bologna Egee User ForumMarch 2th BOSS: the CMS interface for job summission, monitoring and bookkeeping W. Bacchi, P.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks WMSMonitor: a tool to monitor gLite WMS/LB.
LHCb Software Week November 2003 Gennady Kuznetsov Production Manager Tools (New Architecture)
Getting started DIRAC Project. Outline  DIRAC information system  Documentation sources  DIRAC users and groups  Registration with DIRAC  Getting.
DDM Monitoring David Cameron Pedro Salgado Ricardo Rocha.
A university for the world real R © 2009, Chapter 9 The Runtime Environment Michael Adams.
David Adams ATLAS DIAL/ADA JDL and catalogs David Adams BNL December 4, 2003 ATLAS software workshop Production session CERN.
A PanDA Backend for the Ganga Analysis Interface J. Elmsheuser 1, D. Liko 2, T. Maeno 3, P. Nilsson 4, D.C. Vanderster 5, T. Wenaus 3, R. Walker 1 1: Ludwig-Maximilians-Universität.
Module: Software Engineering of Web Applications Chapter 2: Technologies 1.
High-Performance Computing Lab Overview: Job Submission in EDG & Globus November 2002 Wei Xing.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
SAM Sensors & Tests Judit Novak CERN IT/GD SAM Review I. 21. May 2007, CERN.
Bookkeeping Tutorial. 2 Bookkeeping content  Contains records of all “jobs” and all “files” that are produced by production jobs  Job:  In fact technically.
1 DIRAC Job submission A.Tsaregorodtsev, CPPM, Marseille LHCb-ATLAS GANGA Workshop, 21 April 2004.
David Adams ATLAS ATLAS distributed data management David Adams BNL February 22, 2005 Database working group ATLAS software workshop.
DIRAC Review (12 th December 2005)Stuart K. Paterson1 DIRAC Review Workload Management System.
K. Harrison CERN, 22nd September 2004 GANGA: ADA USER INTERFACE - Ganga release status - Job-Options Editor - Python support for AJDL - Job Builder - Python.
Batch Jobs Using the batch job functions. Use [Bulk Changes][Batch Job Utility] to start. Read the information panel. Check with TAMS Technical Support.
SPI NIGHTLIES Alex Hodgkins. SPI nightlies  Build and test various software projects each night  Provide a nightlies summary page that displays all.
David Adams ATLAS ATLAS-ARDA strategy and priorities David Adams BNL October 21, 2004 ARDA Workshop.
ATLAS-specific functionality in Ganga - Requirements for distributed analysis - ATLAS considerations - DIAL submission from Ganga - Graphical interfaces.
DIRAC Project A.Tsaregorodtsev (CPPM) on behalf of the LHCb DIRAC team A Community Grid Solution The DIRAC (Distributed Infrastructure with Remote Agent.
1 Cherenkov Telescope Array: a production system prototype L. Arrabito 1 C. Barbier 2, J. Bregeon 1, A. Haupt 3, N. Neyroud 2 for the CTA Consortium 1.
Pavel Nevski DDM Workshop BNL, September 27, 2006 JOB DEFINITION as a part of Production.
LHCbDirac and Core Software. LHCbDirac and Core SW Core Software workshop, PhC2 Running Gaudi Applications on the Grid m Application deployment o CVMFS.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Practical using WMProxy advanced job submission.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
The GridPP DIRAC project DIRAC for non-LHC communities.
David Adams ATLAS AJDL: Abstract Job Description Language David Adams BNL June 29, 2004 PPDG Collaboration Meeting Williams Bay.
1 DIRAC Data Management Components A.Tsaregorodtsev, CPPM, Marseille DIRAC review panel meeting, 15 November 2005, CERN.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
LHCb/DIRAC week A.Tsaregorodtsev, CPPM 7 April 2011.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
Core and Framework DIRAC Workshop October Marseille.
DIRAC for Grid and Cloud Dr. Víctor Méndez Muñoz (for DIRAC Project) LHCb Tier 1 Liaison at PIC EGI User Community Board, October 31st, 2013.
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
1 DIRAC Project Status A.Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille 10 March, DIRAC Developer meeting.
Modularization of Geant4 Dynamic loading of modules Configurable build using CMake Pere Mato Witek Pokorski
Practical using C++ WMProxy API advanced job submission
GridPP DIRAC Daniela Bauer & Simon Fayer.
Design rationale and status of the org.glite.overlay component
DIRAC Production Manager Tools
BOSS: the CMS interface for job summission, monitoring and bookkeeping
BOSS: the CMS interface for job summission, monitoring and bookkeeping
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
BOSS: the CMS interface for job summission, monitoring and bookkeeping
DUCKS – Distributed User-mode Chirp-Knowledgeable Server
Production Manager Tools (New Architecture)
Presentation transcript:

Transformation System report Luisa Arrabito 1, Federico Stagni 2 1) LUPM CNRS/IN2P3, France 2) CERN 5 th DIRAC User Workshop 27 th – 29 th May 2015, Ferrara

Plan  What’s the Transformation System?  Transformation System architecture  How it works in practice?  Proposal for a new design 2

What’s the Transformation System?  A DIRAC System as usually comprising:  MySQL tables, Services, Agents, Clients, Scripts and Plugins  A system for handling “repetitive work”, i.e. many identical tasks with a varying parameter  2 main usages:  Productions: the “same” job – i.e. the same workflow - is executed  Client for the Workload Management System  Data handling: replications, removal  Client for Request Management System  Handles input datasets (if present)  It interacts with Replica and Metadata catalogs (e.g. DFC or external catalogs)  Plugins are grouping input files into tasks according to various criteria  LHCb ‘Production System’ is built on top of it and CTA is going to do same 3

Transformation System architecture Production Manager defines the transformations TransformationAgent processes the transformations and creates tasks given a Transformation Plugin InputDataAgent queries the Catalog to obtain files to be ‘transformed’ WorkflowTaskAgent transforms tasks into job workflows, given a TaskManager Plugin RequestTaskAgent transforms tasks into requests WMS Production Manager Transformation System Catalog Plugins Transformations Transformation Agent InputData Agent Workflow Task Agent Database table Agent Tasks Files InputDataQuery Request Task Agent RMS Application 4

Standard(): o Group files by replicas (tasks created based on the file location) BySize(): o Group files until they reach a certain size (Input size in Gb) ByShare() o Group files given the share (specified in the CS) and location For replication: Broadcast() o Takes files at the source SE and broadcast to a given number of locations Transformation Plugins 5

6 TaskManager Plugins (from v6r13) BySE(): o Default plugin o Set jobs destination depending from the location of its input data ByJobType(): o By default, all sites are allowed to do every job o The actual rules are freely specified in the CS Operation JobTypeMapping section

How it works in practice (I)?  See documentation at:  mation/index.html  Installation  Need to have the Transformation System installed and running. The minimum is:  Service: TransformationManagerHandler  Database: TransformationDB  Agents:  TransformationAgent  WorkflowTaskAgent  RequestTaskAgent  InputDataAgent  TransformationCleaningAgent 7

How it works in practice (II)?  Configuration  Add the transformation types in the Operations/[VO]/Transformations section, e.g.:  Eventually configure the WorkflowTaskAgent and the RequestTaskAgent to treat a particular transformation type Transformations { DataProcessing = MCSimulation DataProcessing += Merge DataProcessing += Analysis DataProcessing += DataReprocessing DataManipulation = Removal DataManipulation += Replication } 8

How it works in practice (III)?  Create your transformation defining:  Type (e.g.: MCSimulation, DataReprocessing, Replication)  Body (the job workflow to execute, or the request type to execute)  Plugin (e.g.: ByReplica, BySize, Broadcast, default is Standard)  Example for a “processing” transformation: set Type set Body transformation is created here set Inputdata 9

How it works in practice (IV)?  Monitor (and manage) your transformation 10

Proposal for a new design (I)  See RFC #21:   Motivations for improvement:  Large catalog queries may be a bottleneck (experience from LHCb)  Proposal to make the TS fully ‘data-driven’ by implemeting ‘meta-filters’ (see next slide)  Job submission could be improved using bulk submission as done for ‘parametric jobs’  Need to support ‘chained transformations’  Example: in LHCb chained transformations, e.g. Re-processing -> Merging -> Removal, are handled by a dedicated Production System  Proposal to extend the TS to support chained transformations as basis for each community to build its own 'Production System’  Agents in the TS work in ‘polling’ mode  Proposal to use a Message Queueing System complementary to polling 11

Proposal for a new design (II) WMS Production Manager Transformation System Catalog Plugins Transformations Transformation Agent InputData Agent Workflow Task Agent Database table Agent Tasks Files InputDataQuery Request Task Agent RMS Application Filter by meta-data Use the Catalog interface of the TS When new files are registered, a filter based on meta-data is applied No need anymore to perform large Catalog queries 12

Conclusions  The Transformation System allows to handle massive ‘production’ operations (large number of jobs or requests)  Successfully used by LHCb, ILC, CTA…  LHCb experience shows some scalability problem, essentially due to large queries on the catalog  Development work has started to make the TS fully ‘data-driven’  RFC #21 waits for your comments! 13

BACKUP 14

Job Workflows Job description Application Step 1 Application Step 2 Finalization Step (for users AND production jobs) Job description format Enables running “complex” jobs o e.g. multiple applications, linked together via input/output data o I/O chaining description in different formats: XML, JDL, python o JDL executable: dirac-jobexec o Argument: jobDescription.xml (which is in the Input Sandbox) A workflow is composed of steps o that are made of modules o workflow modules are instantiated by python modules P that do the real job o parameters at any level 15

Task_1  Job_1 Task_2  Job_2 … Task_n  Job_n App Step 1 App Step 2 Finalization Step Transformation Job workflow 16