Virtual Data Management for CMS Simulation Production A GriPhyN Prototype.

Slides:



Advertisements
Similar presentations
Semantics Static semantics Dynamic semantics attribute grammars
Advertisements

Virtual Data and the Chimera System* Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer Science.
A. Arbree, P. Avery, D. Bourilkov, R. Cavanaugh, S. Katageri, G. Graham, J. Rodriguez, J. Voeckler, M. Wilde CMS & GriPhyN Conference in High Energy Physics,
GriPhyN & iVDGL Architectural Issues GGF5 BOF Data Intensive Applications Common Architectural Issues and Drivers Edinburgh, 23 July 2002 Mike Wilde Argonne.
WP 1 Grid Workload Management Massimo Sgaravatto INFN Padova.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Three Flavors of Data Science Data Simulations and Sensor Readings Catalog Data Metadata; descriptors of datasets, data products and other processing artifacts.
R. Cavanaugh GriPhyN Analysis Workshop Caltech, June, 2003 Virtual Data Toolkit.
Status of Globus activities within INFN (update) Massimo Sgaravatto INFN Padova for the INFN Globus group
Software Issues Derived from Dr. Fawcett’s Slides Phil Pratt-Szeliga Fall 2009.
Operating Systems Concepts 1. A Computer Model An operating system has to deal with the fact that a computer is made up of a CPU, random access memory.
SS ZG653Second Semester, Topic Architectural Patterns Pipe and Filter.
CSE 1301 J Lecture 2 Intro to Java Programming Richard Gesick.
Resource Management Reading: “A Resource Management Architecture for Metacomputing Systems”
Python Mini-Course University of Oklahoma Department of Psychology Day 1 – Lesson 2 Fundamentals of Programming Languages 4/5/09 Python Mini-Course: Day.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
Grappa: Grid access portal for physics applications Shava Smallen Extreme! Computing Laboratory Department of Physics Indiana University.
XCAT Science Portal Status & Future Work July 15, 2002 Shava Smallen Extreme! Computing Laboratory Indiana University.
Chapter 33 CGI Technology for Dynamic Web Documents There are two alternative forms of retrieving web documents. Instead of retrieving static HTML documents,
Track 1: Cluster and Grid Computing NBCR Summer Institute Session 2.2: Cluster and Grid Computing: Case studies Condor introduction August 9, 2006 Nadya.
The Grid is a complex, distributed and heterogeneous execution environment. Running applications requires the knowledge of many grid services: users need.
ITEC224 Database Programming
Don Quijote Data Management for the ATLAS Automatic Production System Miguel Branco – CERN ATC
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
1 A pattern language for security models Eduardo B. Fernandez and Rouyi Pan Presented by Liping Cai 03/15/2006.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
A Metadata Based Approach For Supporting Subsetting Queries Over Parallel HDF5 Datasets Vignesh Santhanagopalan Graduate Student Department Of CSE.
Distribution After Release Tool Natalia Ratnikova.
Pegasus-a framework for planning for execution in grids Ewa Deelman USC Information Sciences Institute.
GriPhyN Status and Project Plan Mike Wilde Mathematics and Computer Science Division Argonne National Laboratory.
Modeling and simulation of systems Model building Slovak University of Technology Faculty of Material Science and Technology in Trnava.
ILDG Middleware Status Chip Watson ILDG-6 Workshop May 12, 2005.
Stuart Wakefield Imperial College London Evolution of BOSS, a tool for job submission and tracking W. Bacchi, G. Codispoti, C. Grandi, INFN Bologna D.
Lecture 1 Introduction Figures from Lewis, “C# Software Solutions”, Addison Wesley Richard Gesick.
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
 Three-Schema Architecture Three-Schema Architecture  Internal Level Internal Level  Conceptual Level Conceptual Level  External Level External Level.
The european ITM Task Force data structure F. Imbeaux.
Giuseppe Codispoti INFN - Bologna Egee User ForumMarch 2th BOSS: the CMS interface for job summission, monitoring and bookkeeping W. Bacchi, P.
Pipes & Filters Architecture Pattern Source: Pattern-Oriented Software Architecture, Vol. 1, Buschmann, et al.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
Virtual Batch Queues A Service Oriented View of “The Fabric” Rich Baker Brookhaven National Laboratory April 4, 2002.
Grid Scheduler: Plan & Schedule Adam Arbree Jang Uk In.
The GriPhyN Planning Process All-Hands Meeting ISI 15 October 2001.
LIGO-G E Generic LabelLIGO Laboratory at Caltech 1 Distributed Computing for LIGO Data Analysis The Aspen Winter Conference on Gravitational Waves,
Metadata Mòrag Burgon-Lyon University of Glasgow.
CS 127 Introduction to Computer Science. What is a computer?  “A machine that stores and manipulates information under the control of a changeable program”
DGC Paris WP2 Summary of Discussions and Plans Peter Z. Kunszt And the WP2 team.
Replica Management Kelly Clynes. Agenda Grid Computing Globus Toolkit What is Replica Management Replica Management in Globus Replica Management Catalog.
VDT 1 The Virtual Data Toolkit Todd Tannenbaum (Alain Roy)
Peter Couvares Computer Sciences Department University of Wisconsin-Madison Condor DAGMan: Introduction &
Grid Compute Resources and Job Management. 2 Grid middleware - “glues” all pieces together Offers services that couple users with remote resources through.
STAR Scheduling status Gabriele Carcassi 9 September 2002.
Status of Globus activities Massimo Sgaravatto INFN Padova for the INFN Globus group
Application Web Service Toolkit Allow users to quickly add new applications GGF5 Edinburgh Geoffrey Fox, Marlon Pierce, Ozgur Balsoy Indiana University.
STAR Scheduler Gabriele Carcassi STAR Collaboration.
David Adams ATLAS ATLAS Distributed Analysis and proposal for ATLAS-LHCb system David Adams BNL March 22, 2004 ATLAS-LHCb-GANGA Meeting.
10 March Andrey Grid Tools Working Prototype of Distributed Computing Infrastructure for Physics Analysis SUNY.
Chimera Workshop September 4 th, Outline l Install Chimera l Run Chimera –Hello world –Convert simple shell pipeline –Some diamond, etc. l Get.
COP-3330: Object Oriented Programming Course Introduction May 14, 2012 Eng. Hector M Lugo-Cordero, MS.
Preservation Data Services Persistent Archive Research Group Reagan W. Moore October 1, 2003.
Magda Distributed Data Manager Torre Wenaus BNL October 2001.
BOSS: the CMS interface for job summission, monitoring and bookkeeping
BOSS: the CMS interface for job summission, monitoring and bookkeeping
Some ideas for the future
Data, Databases, and DBMSs
Pegasus and Condor Gaurang Mehta, Ewa Deelman, Carl Kesselman, Karan Vahi Center For Grid Technologies USC/ISI.
Weaving Abstractions into Workflows
Mats Rynge USC Information Sciences Institute
Presentation transcript:

Virtual Data Management for CMS Simulation Production A GriPhyN Prototype

10/20012 Goals Explore –virtual data dependency tracking –data derivability –integrate virtual data catalog functionality –use of DAGs in virtual data production Identify –architectural issues: planners, catalogs, interfaces –hard issues in executing real production physics applications Create prototypes –tools that can go into the VDT Test virtual data concepts on something “real”

10/20013 Which Part of GriPhyN Application Planner Executor Catalog Services Info Services Policy/Security Monitoring Repl. Mgmt. Reliable Transfer Service Compute ResourceStorage Resource cDAG aDAG DAGMAN, Kangaroo GRAMGridFTP; GRAM; SRM GSI, CAS MDS MCAT; GriPhyN catalogs GDMP MDS Globus = initial solution is operational

10/20014 What Was Done Created: –A virtual data catalog A catalog scheme for a RDBMS –A “virtual data language”  VDL –A VDL command interpreter –Simple DAGs for the CMS pipeline –Complex DAGs for a canonical test application Kanonical executable for GriPhyN  keg These DAGs actually execute on a Condor- Globus Grid

10/20015 The CMS Challenge Remember Rick’s slides and the complexity! Types of executables (4) Parameters, inputs, and outputs Templates of parameter lists Sensitivities of binaries –Dynamic libraries –Environment variables –Condor-related environment issues – less obvious

10/20016 Assumptions Grid activity takes place as sub-jobs Some subset of subjects create tracked, durable data products – these are tracked in the virtual data catalog Job execution mechanisms execute VDL functions to describe their virtual data manipulations – dependencies and derivations The product of sub-jobs are physical instances of logical files (i.e., physical files) Planners decide where physical files should be created Physical copies of logical files are tracked in replica catalog

10/20017 The VDL begin v /bin/cat arg –n file i filename1 file i filename2 stdout filename3 env key=value end setenv … /bin/cat -n filename1 filename2 filename3

10/20018 Dependent Programs begin v /bin/phys1 arg –n file i f1 file i f2 stdout f3 env key=value end begin v /bin/phys2 arg –m file i f1 file i f3 file o f4 env key=value end …note that dependencies can be complex graphs

10/20019 The Interpreter How program invocations are formed –Environment variables –Regular Parameters –Input files –Output file How DAGs are formed –Recursive determination of dependencies –Parallel execution How scripts are formed –Recursive determination of dependencies –Serial execution (now); parallel is possible

10/ Virtual Data Catalog Relational Database Structure: As Implemented

10/ Virtual Data Catalog Conceptual Data Structure TRANSFORMATION /bin/physapp1 version 1.2.3b(2) created on 12 Oct 1998 owned by physbld.orca DERIVATION ^ paramlist ^ transformation FILE LFN=filename1 PFN1=/store1/ PFN2=/store9/ PFN3=/store4/ ^derivation FILE LFN=filename2 PFN1=/store1/ PFN2=/store9/ ^derivation PARAMETER LIST PARAMETER i filename1 PARAMETER O filename2 PARAMETER E PTYPE=muon PARAMETER p -g

10/ DAGs & Data Structures DAGMan Example –TOP generates even random number –LEFT and RIGHT divide number by 2 –BOTTOM sums Diamond DAG random half sum f.a f.bf.c f.d

10/ DAGs & Data Structures II begin v random stdout f.a end begin v half stdin f.a stdout f.b end begin v half stdin f.a stdout f.c end begin v sum file i f.b file i f.c stdout f.d end rc f.a out.a rc f.b out.b rc f.c out.c rc f.d out.d random half sum f.a f.bf.c

10/ DAGs & Data Structures III XFORMPARAMDERIVEDRC xidcuprgpidvaluexidpidddidposflgpidpfn 1vrnd1f.a111O1out.a 2vhalf2f.b212I2out.b 3vsum3f.c222O3out.c 4f.d213I4out.d 233O 3240i 3341i 344 O

10/ Abstract and Concrete DAGs Abstract DAGs –Resource locations unspecified –File names are logical –Data destinations unspecified Concrete DAGs –Resource locations determined –Physical file names specified –Data delivered to and returned from physical locations Translation is the job of the “planner”

10/ What We Tested DAG structures –Diamond DAG –Canonical “keg” app in complex DAGs –The CMS pipeline Execution environments –Local execution –Grid execution via DAGMan

10/ Generality simple fabric  very powerful DAGs DAGs of this pattern with >260 nodes were run.

10/ What We Have Learned UNIX program execution semantics is messy but manageable –Command line execution is manageable –File accesses can be trapped and tracked Dynamic loading makes reproducibility more difficult – should be avoided if possible Object handling *obviously* needs concentrated research effort

10/ Future Work Working with OO Databases Handling navigational access Refining notion of signatures Dealing with fuzzy dependencies and equivalence Cost tracking and calculations (w/ planner) Automating the cataloging process –Integration with portals –Uniform execution language –Analysis of scripts (shell, Perl, Python, Tcl) Refinement of data staging paradigms Handling shell details –Pipes, 3>&1 (fd #s)

10/ Future Work II Design of Staging Semantics –What files need to be moved where to start a computation –How do you know (exactly) where the computation will run, and how to get the file “there” (NFS, local etc) –How/when to get the results back –How/when to trust the catalog –Double-check file’s existence/ safe arrival when you get there to use it –DB marking of files existence – schema, timing –Mechanisms to audit and correct consistency of catalog vs. reality