CMS Applications Towards Requirements for Data Processing and Analysis on the Open Science Grid Greg Graham FNAL CD/CMS for OSG Deployment 16-Dec-2004.

Slides:



Advertisements
Similar presentations
Plateforme de Calcul pour les Sciences du Vivant SRB & gLite V. Breton.
Advertisements

McFarm: first attempt to build a practical, large scale distributed HEP computing cluster using Globus technology Anand Balasubramanian Karthik Gopalratnam.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
GRID DATA MANAGEMENT PILOT (GDMP) Asad Samar (Caltech) ACAT 2000, Fermilab October , 2000.
Data Management for Physics Analysis in PHENIX (BNL, RHIC) Evaluation of Grid architecture components in PHENIX context Barbara Jacak, Roy Lacey, Saskia.
Workload Management Massimo Sgaravatto INFN Padova.
F Run II Experiments and the Grid Amber Boehnlein Fermilab September 16, 2005.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Open Science Grid Software Stack, Virtual Data Toolkit and Interoperability Activities D. Olson, LBNL for the OSG International.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
HEP Experiment Integration within GriPhyN/PPDG/iVDGL Rick Cavanaugh University of Florida DataTAG/WP4 Meeting 23 May, 2002.
1 Dynamic Application Installation (Case of CMS on OSG) Introduction CMS Software Installation Overview Software Installation Issues Validation Considerations.
OSG Services at Tier2 Centers Rob Gardner University of Chicago WLCG Tier2 Workshop CERN June 12-14, 2006.
OSG Middleware Roadmap Rob Gardner University of Chicago OSG / EGEE Operations Workshop CERN June 19-20, 2006.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
VOX Project Status T. Levshina. Talk Overview VOX Status –Registration –Globus callouts/Plug-ins –LRAS –SAZ Collaboration with VOMS EDG team Preparation.
INFSO-RI Enabling Grids for E-sciencE SA1: Cookbook (DSA1.7) Ian Bird CERN 18 January 2006.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
A. Sim, CRD, L B N L 1 OSG Applications Workshop 6/1/2005 OSG SRM/DRM Readiness and Plan Alex Sim / Jorge Rodriguez Scientific Data Management Group Computational.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
Distribution After Release Tool Natalia Ratnikova.
Apr 30, 20081/11 VO Services Project – Stakeholders’ Meeting Gabriele Garzoglio VO Services Project Stakeholders’ Meeting Apr 30, 2008 Gabriele Garzoglio.
1 st December 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
DOSAR Workshop, Sao Paulo, Brazil, September 16-17, 2005 LCG Tier 2 and DOSAR Pat Skubic OU.
ILDG Middleware Status Chip Watson ILDG-6 Workshop May 12, 2005.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
Mar 28, 20071/9 VO Services Project Gabriele Garzoglio The VO Services Project Don Petravick for Gabriele Garzoglio Computing Division, Fermilab ISGC 2007.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
David Adams ATLAS ADA, ARDA and PPDG David Adams BNL June 28, 2004 PPDG Collaboration Meeting Williams Bay, Wisconsin.
CPT Demo May Build on SC03 Demo and extend it. Phase 1: Doing Root Analysis and add BOSS, Rendezvous, and Pool RLS catalog to analysis workflow.
09/02 ID099-1 September 9, 2002Grid Technology Panel Patrick Dreher Technical Panel Discussion: Progress in Developing a Web Services Data Analysis Grid.
CERN Using the SAM framework for the CMS specific tests Andrea Sciabà System Analysis WG Meeting 15 November, 2007.
1 DØ Grid PP Plans – SAM, Grid, Ceiling Wax and Things Iain Bertram Lancaster University Monday 5 November 2001.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Stefano Belforte INFN Trieste 1 Middleware February 14, 2007 Resource Broker, gLite etc. CMS vs. middleware.
Overview of Privilege Project at Fermilab (compilation of multiple talks and documents written by various authors) Tanya Levshina.
GridPP11 Liverpool Sept04 SAMGrid GridPP11 Liverpool Sept 2004 Gavin Davies Imperial College London.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
VO Privilege Activity. The VO Privilege Project develops and implements fine-grained authorization to grid- enabled resources and services Started Spring.
Storage and Data Movement at FNAL D. Petravick CHEP 2003.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Sep 25, 20071/5 Grid Services Activities on Security Gabriele Garzoglio Grid Services Activities on Security Gabriele Garzoglio Computing Division, Fermilab.
David Adams ATLAS ATLAS-ARDA strategy and priorities David Adams BNL October 21, 2004 ARDA Workshop.
Status of Globus activities Massimo Sgaravatto INFN Padova for the INFN Globus group
April 25, 2006Parag Mhashilkar, Fermilab1 Resource Selection in OSG & SAM-On-The-Fly Parag Mhashilkar Fermi National Accelerator Laboratory Condor Week.
US ATLAS – new grid initiatives John Huth Harvard University US ATLAS Software Meeting: BNL Aug 03.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
VOX Project Tanya Levshina. 05/17/2004 VOX Project2 Presentation overview Introduction VOX Project VOMRS Concepts Roles Registration flow EDG VOMS Open.
VOX Project Status T. Levshina. 5/7/2003LCG SEC meetings2 Goals, team and collaborators Purpose: To facilitate the remote participation of US based physicists.
10 March Andrey Grid Tools Working Prototype of Distributed Computing Infrastructure for Physics Analysis SUNY.
Towards deploying a production interoperable Grid Infrastructure in the U.S. Vicky White U.S. Representative to GDB.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
OSG Status and Rob Gardner University of Chicago US ATLAS Tier2 Meeting Harvard University, August 17-18, 2006.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
Monitoring and Information Services Core Infrastructure (MIS-CI) Service Description Mark L. Green OSG Integration Workshop at UC Feb 15-17, 2005.
April 18, 2006FermiGrid Project1 FermiGrid Project Status April 18, 2006 Keith Chadwick.
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
GGF OGSA-WG, Data Use Cases Peter Kunszt Middleware Activity, Data Management Cluster EGEE is a project funded by the European.
A Model for Grid User Management
LCG middleware and LHC experiments ARDA project
Patrick Dreher Research Scientist & Associate Director
Leigh Grundhoefer Indiana University
Status of Grids for HEP and HENP
Presentation transcript:

CMS Applications Towards Requirements for Data Processing and Analysis on the Open Science Grid Greg Graham FNAL CD/CMS for OSG Deployment 16-Dec-2004

12/16/04Greg Graham FNAL CD/CMS2 CMS Applications - General Requirements Access to large datasets at a few “central” sites Access to small datasets at many distributed sites Ability to move large datasets between sites Ability to create jobs to run against these datasets Ability to submit jobs and track progress Ability to control/restrict access to sites/resources Ability to lookup information about datasets and jobs

12/16/04Greg Graham FNAL CD/CMS3 Specific Application Examples CMS Distributed Processing Environment (DPE) –VDT/Grid2003 based software package to provide CMS specific software on top of Grid software. Monte Carlo Production –MCRunjob (CMS Tool) to create jobs, MOP (PPDG) to submit jobs using Condor-G, ConfMon to provide site parameters for MOP. Large Scale Data Transfer –srmcp to transfer results of production from one site(transient) to another (permanent) –Phedex to transfer data with metadata (GridFTP)

12/16/04Greg Graham FNAL CD/CMS4 Monte Carlo Processing Service A Clarens based system for generating, processing, and analyzing Monte Carlo data. –Runjob, MOP, DAR software repository, and MOPDb deployed behind Clarens Web services –SC2004 demo: point and click MC generation and analysis (Root tuples also served by Clarens) Currently deployed on top of DPE; and it requires Clarens. Status: Deployed now, need to groom CMS users –Needed: a parameter service to accept and store arbitrary job configuration parameters AND a context service

12/16/04Greg Graham FNAL CD/CMS5 CMS History with Grid Using Condor-G/Globus based technology to do real CMS MC production since –Shook out bugs and performance issues, used MOP Using Grid2003 technology to do real production since –Stakeholder in security (SAZ), registration (VOMRS), and data transfer protocols We plan to migrate to an OSG product based on the current Grid2003 –Must meet requirements and we are working to discover those –In the meantime, we assume it will work like it does currently for DPE running on top of Grid2003 cache

12/16/04Greg Graham FNAL CD/CMS6 Current CMS Deployment Activities for OSG Within the DPE scope: –Moving to VDT to be consistent with Grid3-dev –Testing latest versions of SRM We are able to run MOP production with older versions of srm. Craig Prescott is investigating later versions with Timur. 12/13/04 OSGD milestone. –MCPS rollout on OSG 3/1/05 OSGD Milestone –Testing Condor-C and providing feedback to the Condor team; and also testing VDT 1.3 No milestone listed in the OSG deployment doc Keeping up will help us be ready for OSG “turn on”

12/16/04Greg Graham FNAL CD/CMS7 Conclusion The requirements for CMS applications running on OSG can be gleaned from looking at the current requirements for running on Grid2003. The requirements laid out here should be concretized in two documents –CMS Requirements for OSG Deployment To track the current requirements –Impact of OSG Deployment on CMS Software To track evolution of the requirements CMS has a lot of experience running on the Grid –Procedures are in place to deal with an evolving middleware environment.

Summary of Known Requirements for OSG Deployment from CMS

12/16/04Greg Graham FNAL CD/CMS9 Infrastructure: MC Production Support for MOP style job submission –Condor-G/Globus from VDT 1.14 or better But we are exploring use of Condor-C –Information service ConfMon MDS based hack of the Glue Schema to tell MOP where to find software remotely, where to deposit output files. But we are (hopefully) moving to GridCat –Space to drop in CMS application software and hold the output temporarily –Servers to move the data off of the remote site. srmcp is preferred, GridFTP is default right now.

12/16/04Greg Graham FNAL CD/CMS10 Infrastructure: Data Access The requirements are less well known at the moment Directory based lookup of data products –Since this is CMS based data, we would expect that CMS clients would be used to do lookup. Are there any common lookup operations? Then CLARENS may be required on the client side. Data movement from “large” central sites to/from “small” sites –srmcp, GridFTP clients are required. Data movement between all sites and “push” from large sites –srmcp and GridFTP servers are required.

12/16/04Greg Graham FNAL CD/CMS11 Infrastructure: Common Security: –Middleware needs to have strong authentication Kerberos tickets or equivalent VO authentication –Middleware needs to support the callouts or other mechanisms used by SAZ database and GUMS We are now dependent upon gridmapfiles, but I am not sure if this is required –Participating sites need to support the interfaces and provide information needed by VOMRS. Required to submit jobs to Fermilab, maybe not required to accept jobs from Fermilab ;-)

12/16/04Greg Graham FNAL CD/CMS12 Infrastructure: Common Information services –Real-time information about running jobs and resource usage –Historical information and accounting (soft requirement) –Remote viewing of selected logfiles would also be useful (soft requirement - satisfied by operations staff?) Catalog services –CMS will initially come in with its own file and metadata catalogs. In the future we may rely on Globus RLS for file replicas. Open question if common cataloging services would be useful. Service Discovery –CMS will initially come in with its own service discovery method (ie- the null one;-). In the future, we may rely on CLARENS based services.

12/16/04Greg Graham FNAL CD/CMS13 LCG Interoperability We currently have a job creation and submission tool that can submit to either LCG or Grid2003 resources. Interoperability at a lower level may also be required to satisfy simultaneously the needs of the CMS collaboration and the institutional needs of Fermilab. –This is currently under development and we are very interested in the results.