R. Cavanaugh GriPhyN Analysis Workshop Caltech, June, 2003 Virtual Data Toolkit.

Slides:



Advertisements
Similar presentations
FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
Advertisements

Grid Resource Allocation Management (GRAM) GRAM provides the user to access the grid in order to run, terminate and monitor jobs remotely. The job request.
NorduGrid Grid Manager developed at NorduGrid project.
CERN LCG Overview & Scaling challenges David Smith For LCG Deployment Group CERN HEPiX 2003, Vancouver.
A. Arbree, P. Avery, D. Bourilkov, R. Cavanaugh, S. Katageri, G. Graham, J. Rodriguez, J. Voeckler, M. Wilde CMS & GriPhyN Conference in High Energy Physics,
GriPhyN & iVDGL Architectural Issues GGF5 BOF Data Intensive Applications Common Architectural Issues and Drivers Edinburgh, 23 July 2002 Mike Wilde Argonne.
Sphinx Server Sphinx Client Data Warehouse Submitter Generic Grid Site Monitoring Service Resource Message Interface Current Sphinx Client/Server Multi-threaded.
CMS HLT production using Grid tools Flavia Donno (INFN Pisa) Claudio Grandi (INFN Bologna) Ivano Lippi (INFN Padova) Francesco Prelz (INFN Milano) Andrea.
GRID workload management system and CMS fall production Massimo Sgaravatto INFN Padova.
Virtual Data in CMS Analysis A.Arbree, P.Avery, D.Bourilkov, R.Cavanaugh, G.Graham, J.Rodriguez, M.Wilde, Y.Zhao CMS & GriPhyN CHEP03, La Jolla, California.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Experience with ATLAS Data Challenge Production on the U.S. Grid Testbed Kaushik De University of Texas at Arlington CHEP03 March 27, 2003.
GRID Workload Management System Massimo Sgaravatto INFN Padova.
Globus activities within INFN Massimo Sgaravatto INFN Padova for the INFN Globus group
Workload Management Massimo Sgaravatto INFN Padova.
First steps implementing a High Throughput workload management system Massimo Sgaravatto INFN Padova
Status of Globus activities within INFN (update) Massimo Sgaravatto INFN Padova for the INFN Globus group
Evaluation of the Globus GRAM Service Massimo Sgaravatto INFN Padova.
Managing Workflows with the Pegasus Workflow Management System
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
Vladimir Litvin, Harvey Newman Caltech CMS Scott Koranda, Bruce Loftis, John Towns NCSA Miron Livny, Peter Couvares, Todd Tannenbaum, Jamie Frey Wisconsin.
OSG End User Tools Overview OSG Grid school – March 19, 2009 Marco Mambelli - University of Chicago A brief summary about the system.
Grid Testbed Activities in US-CMS Rick Cavanaugh University of Florida 1. Infrastructure 2. Highlight of Current Activities 3. Future Directions NSF/DOE.
Don Quijote Data Management for the ATLAS Automatic Production System Miguel Branco – CERN ATC
10/20/05 LIGO Scientific Collaboration 1 LIGO Data Grid: Making it Go Scott Koranda University of Wisconsin-Milwaukee.
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
OSG Services at Tier2 Centers Rob Gardner University of Chicago WLCG Tier2 Workshop CERN June 12-14, 2006.
OSG Middleware Roadmap Rob Gardner University of Chicago OSG / EGEE Operations Workshop CERN June 19-20, 2006.
ESP workshop, Sept 2003 the Earth System Grid data portal presented by Luca Cinquini (NCAR/SCD/VETS) Acknowledgments: ESG.
Job Submission Condor, Globus, Java CoG Kit Young Suk Moon.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
1 st December 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
Pegasus-a framework for planning for execution in grids Ewa Deelman USC Information Sciences Institute.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
GriPhyN Status and Project Plan Mike Wilde Mathematics and Computer Science Division Argonne National Laboratory.
CSF4 Meta-Scheduler Name: Zhaohui Ding, Xiaohui Wei
Grid Workload Management Massimo Sgaravatto INFN Padova.
Resource Brokering in the PROGRESS Project Juliusz Pukacki Grid Resource Management Workshop, October 2003.
David Adams ATLAS ADA, ARDA and PPDG David Adams BNL June 28, 2004 PPDG Collaboration Meeting Williams Bay, Wisconsin.
Giuseppe Codispoti INFN - Bologna Egee User ForumMarch 2th BOSS: the CMS interface for job summission, monitoring and bookkeeping W. Bacchi, P.
CPT Demo May Build on SC03 Demo and extend it. Phase 1: Doing Root Analysis and add BOSS, Rendezvous, and Pool RLS catalog to analysis workflow.
Grid Security: Authentication Most Grids rely on a Public Key Infrastructure system for issuing credentials. Users are issued long term public and private.
Grid Scheduler: Plan & Schedule Adam Arbree Jang Uk In.
July 11-15, 2005Lecture3: Grid Job Management1 Grid Compute Resources and Job Management.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
The GriPhyN Planning Process All-Hands Meeting ISI 15 October 2001.
GriPhyN Virtual Data System Grid Execution of Virtual Data Workflows Mike Wilde Argonne National Laboratory Mathematics and Computer Science Division.
US CMS Centers & Grids – Taiwan GDB Meeting1 Introduction l US CMS is positioning itself to be able to learn, prototype and develop while providing.
Pegasus-a framework for planning for execution in grids Karan Vahi USC Information Sciences Institute May 5 th, 2004.
Alain Roy Computer Sciences Department University of Wisconsin-Madison Condor & Middleware: NMI & VDT.
Planning Ewa Deelman USC Information Sciences Institute GriPhyN NSF Project Review January 2003 Chicago.
VDT 1 The Virtual Data Toolkit Todd Tannenbaum (Alain Roy)
Virtual Data Management for CMS Simulation Production A GriPhyN Prototype.
Korea Workshop May GAE CMS Analysis (Example) Michael Thomas (on behalf of the GAE group)
GriPhyN Project Paul Avery, University of Florida, Ian Foster, University of Chicago NSF Grant ITR Research Objectives Significant Results Approach.
Peter Couvares Computer Sciences Department University of Wisconsin-Madison Condor DAGMan: Introduction &
Grid Compute Resources and Job Management. 2 Grid middleware - “glues” all pieces together Offers services that couple users with remote resources through.
STAR Scheduling status Gabriele Carcassi 9 September 2002.
Status of Globus activities Massimo Sgaravatto INFN Padova for the INFN Globus group
Grid Workload Management (WP 1) Massimo Sgaravatto INFN Padova.
STAR Scheduler Gabriele Carcassi STAR Collaboration.
Managing LIGO Workflows on OSG with Pegasus Karan Vahi USC Information Sciences Institute
OSG Status and Rob Gardner University of Chicago US ATLAS Tier2 Meeting Harvard University, August 17-18, 2006.
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
Workload Management Workpackage
U.S. ATLAS Grid Production Experience
Peter Kacsuk – Sipos Gergely MTA SZTAKI
Pegasus and Condor Gaurang Mehta, Ewa Deelman, Carl Kesselman, Karan Vahi Center For Grid Technologies USC/ISI.
Presentation transcript:

R. Cavanaugh GriPhyN Analysis Workshop Caltech, June, 2003 Virtual Data Toolkit

Caltech Analysis Workshop2 Very Early GriPhyN Data Grid Architecture Application Planner Executor Catalog Services Info Services Monitoring Repl. Mgmt. Reliable Transfer Service Compute ResourceStorage Resource DAGMAN, Condor-G GRAMGridFTP; GRAM; SRM GSI, CAS MDS MCAT; GriPhyN catalogs GDMP MDS Globus = initial solution is operational Policy/Security

Caltech Analysis Workshop3 Currently Evolved GriPhyN Picture Picture Taken from Mike Wilde

Caltech Analysis Workshop4 Current VDT Emphasis l Current reality –Easy grid construction >Strikes a balance between flexibility and “easibility” >purposefully errs (just a little bit) on the side of “easibility” –Long running, high-throughput, file-based computing –Abstract description of complex workflows –Virtual Data Request Planning –Partial provenance tracking of workflows l Future directions (current research) including: –Policy based scheduling >With notions of Quality of Service (advanced reservation of resources, etc) –Dataset based (arbitrary type structures) –Full provenance tracking of workflows –Several others…

Caltech Analysis Workshop5 Current VDT Flavors l Client –Globus Toolkit 2 >GSI >globusrun >GridFTP Client –CA signing policies for DOE and EDG –Condor-G / DAGMan –RLS Client –MonALISA Client (soon) l Chimera l SDK –Globus –ClassAds –RLS Client –Netlogger l Server –Globus Toolkit >GSI >Gatekeeper >job-managers and GASS Cache >MDS >GridFTP Server –MyProxy –CA signing policies for DOE and EDG –EDG Certificate Revocation List –Fault Tolerant Shell –GLUE Schema –mkgridmap –Condor / DAGMan –RLS Server –MonALISA Server (soon)

Caltech Analysis Workshop6 Chimera Virtual Data System l Virtual Data Language –textual –XML l Virtual Data Catalog –MySQL or PostGreSQL based –File based version available

Caltech Analysis Workshop7 Virtual Data Language TR CMKIN( out a2, in a1 ) { argument file = ${a1}; argument file = ${a2}; } TR CMSIM( out a2, in a1 ) { argument file = ${a1}; argument file = ${a2}; } DV x1->CMKIN( DV x2->CMSIM( file1 file2 file3 x1 x2 Picture Taken from Mike Wilde

Caltech Analysis Workshop8 Virtual Data Request Planning l Abstract Planner –Graph traversal of (virtual) data dependencies –Generates the graph with maximal data dependencies –Somewhat analogous to Build Style l Concrete (Pegasus) Planner –Prunes execution steps for which data already exists (RLS lookup) –Binds all execution steps in the graph to a site –Adds “housekeeping” steps >Create environment, stage-in data, stage-out data, publish data, clean-up environment, etc –Generates a graph with minimal execution steps –Somewhat analogous to Make Style

Caltech Analysis Workshop9 Chimera Virtual Data System: Mapping Abstract Workflows onto Concrete Environments l Abstract DAGs (virtual workflow) –Resource locations unspecified –File names are logical –Data destinations unspecified –build style l Concrete DAGs (stuff for submission) –Resource locations determined –Physical file names specified –Data delivered to and returned from physical locations –make style Abs. Plan VDC RLS C. Plan. DAX DAGMan DAG VDL Logical Physical XML In general there is a full range of planning steps between abstract workflows and concrete workflows Picture Taken from Mike Wilde

Caltech Analysis Workshop10 mass = 200 decay = WW stability = 1 event = 8 mass = 200 decay = WW stability = 1 plot = 1 mass = 200 decay = WW plot = 1 mass = 200 decay = WW event = 8 mass = 200 decay = WW stability = 1 mass = 200 decay = WW stability = 3 mass = 200 decay = WW mass = 200 decay = ZZ mass = 200 decay = bb mass = 200 plot = 1 mass = 200 event = 8 A virtual space of simulated data is generated for future use by scientists... Supercomputing 2002

Caltech Analysis Workshop11 mass = 200 decay = WW stability = 1 LowPt = 20 HighPt = mass = 200 decay = WW stability = 1 event = 8 mass = 200 decay = WW stability = 1 plot = 1 mass = 200 decay = WW plot = 1 mass = 200 decay = WW event = 8 mass = 200 decay = WW stability = 1 mass = 200 decay = WW stability = 3 mass = 200 decay = WW mass = 200 decay = ZZ mass = 200 decay = bb mass = 200 plot = 1 mass = 200 event = 8 Scientists may add new derived data branches... Supercomputing 2002

Caltech Analysis Workshop12 POOL Generator Simulator Formator writeESD writeAOD writeTAG writeESD writeAOD writeTAG Analysis Scripts Digitiser Calib. DB Example CMS Data/ Workflow

Caltech Analysis Workshop13 POOL Generator Simulator Formator writeESD writeAOD writeTAG writeESD writeAOD writeTAG Analysis Scripts Digitiser Calib. DB Online Teams (Re)processing Team MC Production Team Physics Groups Data/workflow is a collaborative endeavour!

Caltech Analysis Workshop14 A “Concurrent Analysis Versioning System:” Complex Data Flow and Data Provenance in HEP Raw ESD AOD TAG Plots, Tables, Fits Comparisons Plots, Tables, Fits Real Data Simulated Data l Family History of a Data Analysis l Collaborative Analysis Development Environment l "Check-point" a Data Analysis l Analysis Development Environment (like CVS) l Audit a Data Analysis

Caltech Analysis Workshop15 Current Prototype GriPhyN “Architecture” (Picture) Picture Taken from Mike Wilde

Caltech Analysis Workshop16 Post-talk: My wandering mind… Typical VDT Configuration l Single public head-node (gatekeeper) –VDT-server installed l Many private worker-nodes –Local scheduler software installed –No grid-middleware installed l Shared file system (e.g. NFS) –User area shared between head-node and worker-nodes –One or many raid systems typically shared

Caltech Analysis Workshop17 compute machines Condor-G Chimera DAGman gahp_server submit hostremote host gatekeeper Local Scheduler (Condor, PBS, etc.) Default middleware configuration from the Virtual Data Toolkit

Caltech Analysis Workshop18 EDG Configuration (for comparison) l CPU separate from Storage –CE: single gatekeeper for access to cluster –SE: single gatekeeper for access to storage l Many public worker-nodes (at least NAT) –Local scheduler installed (LSF or PBS) –Each worker-node runs a GridFTP Client l No assumed shared file system –Data access is accomplished via globus-url-copy to local disk on worker-node

Caltech Analysis Workshop19 Why Care? l Data Analyses would benefit from being fabric independent! l But…the devil is (still) in the details! –Assumptions in job descriptions/requirements currently lead to direct fabric-level consequences and vice versa. l Are existing middleware configurations sufficient for Data Analysis (“scheduled” and “interactive”)? –Really need input from groups like here! –What kind of fabric layer is necessary for “interactive” data analysis using PROOF, JAS? l Does the VDT need multiple configuration flavors? –Production, batch oriented (current default) –Analysis, interactive oriented