22 nd September 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.

Slides:



Advertisements
Similar presentations
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
Advertisements

Physics with SAM-Grid Stefan Stonjek University of Oxford 6 th GridPP Meeting 30 th January 2003 Coseners House.
4/2/2002HEP Globus Testing Request - Jae Yu x Participating in Globus Test-bed Activity for DØGrid UTA HEP group is playing a leading role in establishing.
Rod Walker IC 13th March 2002 SAM-Grid Middleware  SAM.  JIM.  RunJob.  Conclusions. - Rod Walker,ICL.
A Computation Management Agent for Multi-Institutional Grids
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
The Sam-Grid project Gabriele Garzoglio ODS, Computing Division, Fermilab PPDG, DOE SciDAC ACAT 2002, Moscow, Russia June 26, 2002.
Workload Management Massimo Sgaravatto INFN Padova.
JIM Deployment for the CDF Experiment M. Burgon-Lyon 1, A. Baranowski 2, V. Bartsch 3,S. Belforte 4, G. Garzoglio 2, R. Herber 2, R. Illingworth 2, R.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Grid Job, Information and Data Management for the Run II Experiments at FNAL Igor Terekhov et al (see next slide) FNAL/CD/CCF, D0, CDF, Condor team, UTA,
SAMGrid – A fully functional computing grid based on standard technologies Igor Terekhov for the JIM team FNAL/CD/CCF.
Vladimir Litvin, Harvey Newman Caltech CMS Scott Koranda, Bruce Loftis, John Towns NCSA Miron Livny, Peter Couvares, Todd Tannenbaum, Jamie Frey Wisconsin.
Connecting OurGrid & GridSAM A Short Overview. Content Goals OurGrid: architecture overview OurGrid: short overview GridSAM: short overview GridSAM: example.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
5 November 2001GridPP Collaboration Meeting1 CDF and the Grid Requirements and Anti-Requirements CDF-o-Centric View Proposal Conclusion: CDF/D0 Deliverables.
Grid Job and Information Management (JIM) for D0 and CDF Gabriele Garzoglio for the JIM Team.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
CDF Grid Status Stefan Stonjek 05-Jul th GridPP meeting / Durham.
Grid Resource Allocation and Management (GRAM) Execution management Execution management –Deployment, scheduling and monitoring Community Scheduler Framework.
Deploying and Operating the SAM-Grid: lesson learned Gabriele Garzoglio for the SAM-Grid Team Sep 28, 2004.
28 April 2003Lee Lueking, PPDG Review1 BaBar and DØ Experiment Reports DOE Review of PPDG January 28-29, 2003 Lee Lueking Fermilab Computing Division D0.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
CHEP 2003Stefan Stonjek1 Physics with SAM-Grid Stefan Stonjek University of Oxford CHEP th March 2003 San Diego.
CHEP'07 September D0 data reprocessing on OSG Authors Andrew Baranovski (Fermilab) for B. Abbot, M. Diesburg, G. Garzoglio, T. Kurca, P. Mhashilkar.
Installation Sites. Active SAM-Grid Sites  The following sites are actively used for different types of jobs Monte Carlo – production of simulated events.
China Grid Activity on SIG Presented by Guoqing Li At WGISS-21, Budapest 8 May, 2006.
INFSO-RI Enabling Grids for E-sciencE Project Gridification: the UNOSAT experience Patricia Méndez Lorenzo CERN (IT-PSS/ED) CERN,
SAMGrid for CDF MC (and beyond) Igor Terekhov, FNAL/CD/CCF/SAM for JIM team.
1 st December 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
Interactive Job Monitor: CafMon kill CafMon tail CafMon dir CafMon log CafMon top CafMon ps LcgCAF: CDF submission portal to LCG resources Francesco Delli.
INFSO-RI Enabling Grids for E-sciencE Workload Management System Mike Mineter
SAMGrid as a Stakeholder of FermiGrid Valeria Bartsch Computing Division Fermilab.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
- Distributed Analysis (07may02 - USA Grid SW BNL) Distributed Processing Craig E. Tull HCG/NERSC/LBNL (US) ATLAS Grid Software.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
GridPP18 Glasgow Mar 07 DØ – SAMGrid Where’ve we come from, and where are we going? Evolution of a ‘long’ established plan Gavin Davies Imperial College.
Resource Brokering in the PROGRESS Project Juliusz Pukacki Grid Resource Management Workshop, October 2003.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
The SAM-Grid and the use of Condor-G as a grid job management middleware Gabriele Garzoglio for the SAM-Grid Team Fermilab, Computing Division.
Dzero MC production on LCG How to live in two worlds (SAM and LCG)
16 September GridPP 5 th Collaboration Meeting D0&CDF SAM and The Grid Act I: Grid, Sam and Run II Rick St. Denis – Glasgow University Act II: Sam4CDF.
The SAM-Grid / LCG Interoperability Test Bed Gabriele Garzoglio ( ) Speaker: Pierre Girard (
The Experiments – progress and status Roger Barlow GridPP7 Oxford 2 nd July 2003.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Data reprocessing for DZero on the SAM-Grid Gabriele Garzoglio for the SAM-Grid Team Fermilab, Computing Division.
…building the next IT revolution From Web to Grid…
Metadata Mòrag Burgon-Lyon University of Glasgow.
Evolution of a High Performance Computing and Monitoring system onto the GRID for High Energy Experiments T.L. Hsieh, S. Hou, P.K. Teng Academia Sinica,
GridPP11 Liverpool Sept04 SAMGrid GridPP11 Liverpool Sept 2004 Gavin Davies Imperial College London.
VO Privilege Activity. The VO Privilege Project develops and implements fine-grained authorization to grid- enabled resources and services Started Spring.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Grid2Win : gLite for Microsoft Windows Roberto.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Grid Job, Information and Data Management for the Run II Experiments at FNAL Igor Terekhov et al FNAL/CD/CCF, D0, CDF, Condor team.
April 25, 2006Parag Mhashilkar, Fermilab1 Resource Selection in OSG & SAM-On-The-Fly Parag Mhashilkar Fermi National Accelerator Laboratory Condor Week.
Victoria A. White Head, Computing Division, Fermilab Fermilab Grid Computing – CDF, D0 and more..
A Data Handling System for Modern and Future Fermilab Experiments Robert Illingworth Fermilab Scientific Computing Division.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
G. Russo, D. Del Prete, S. Pardi Kick Off Meeting - Isola d'Elba, 2011 May 29th–June 01th A proposal for distributed computing monitoring for SuperB G.
Sept Wyatt Merritt Run II Computing Review1 Status of SAMGrid / Future Plans for SAMGrid  Brief introduction to SAMGrid  Status and deployments.
DØ Grid Computing Gavin Davies, Frédéric Villeneuve-Séguier Imperial College London On behalf of the DØ Collaboration and the SAMGrid team The 2007 Europhysics.
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
Presentation transcript:

22 nd September 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow

22 nd September 2003 JIM for CDF 2 Contents What is JIM? What is SAMGrid? How does JIM relate to SAMGrid? How does JIM work? Using JIM What will JIM do once complete? Deployment Plan

22 nd September 2003 JIM for CDF 3 What is JIM? Job Management Infrastructure and Information Management Job Management Infrastructure is the framework allowing job submissions for execution on a cluster that has enough resources to complete the job. Information Management is knowing what resources are available for use.

22 nd September 2003 JIM for CDF 4 What is SAMGrid? SAMGrid is a grid infrastructure whose goal is to allow globally distributed computing for current experiments, CDF and DØ. Minos is strongly considering SAM Communicating with LHC experiments CMS (Fermilab) and Atlas (Brookhaven)

22 nd September 2003 JIM for CDF 5 How does JIM relate to SAMGrid? Complements Sequencial Access via Metadata to provide complete grid services Job Management Information and Monitoring Data Handling

22 nd September 2003 JIM for CDF 6 How does JIM relate to SAMGrid?

22 nd September 2003 JIM for CDF 7 How does JIM relate to SAMGrid? JIM allows a user to submit jobs to SAMGrid and to access the output files on completion. JIM chooses which resources will be used to execute a job. This decision is based on the nearest location of the required data files. SAM executes the submitted job on the chosen cluster providing the I/O data management for the files.

22 nd September 2003 JIM for CDF 8 How does JIM relate to SAMGrid Condor MMS was expanded for SAMGrid –Broker can query SAM station to see how much data is already present. Globus toolkit is used for job transfer and monitoring

22 nd September 2003 JIM for CDF 9 How does JIM work? JOB Computing Element Submission Client User Interface Queuing System User Interface Broker Match Making Service Information Collector Execution Site #1 Submission Client Match Making Service Computing Element Grid Sensors Execution Site #n Queuing System Grid Sensors Storage Element Computing Element Storage Element Data Handling System Storage Element Informatio n Collector Grid Sensor s Computin g Element Data Handling System

22 nd September 2003 JIM for CDF 10 How does JIM work? Distinguishes grid level (global) scheduling (selection of a cluster to run) from local scheduling (distribution of the job within the cluster). Distinguishes structured jobs (where the details are known to Grid middleware) from unstructured jobs (where the whole job is mapped onto a single cluster).

22 nd September 2003 JIM for CDF 11 Using JIM – Submitting a job User creates a jdl file such as the example shown: sam_dataset = jpmm08-1file executable = retrieve.sh input_sandbox = /home_scotgrid/m/mlyon/test/testjob cpu-per-event = 1s job_manager = sam job_type = sam_analysis sam_universe = prd sam_experiment = cdf log = testjob.log output = testjob.out error = testjob.err arguments = Download Output Yet? group = test instances = 1

22 nd September 2003 JIM for CDF 12 Using JIM – Submitting a job The job is submitted by typing: samg submit testjobfile.jdf The progress of the job can be viewed by selecting the submission site from the list: Select the job from the list. Details of the job state are displayed Once the job has been completed the output may be downloaded

22 nd September 2003 JIM for CDF 13 Using JIM – Viewing Map

22 nd September 2003 JIM for CDF 14 Using JIM – Viewing submission sites

22 nd September 2003 JIM for CDF 15 Using JIM – Viewing submitted jobs

22 nd September 2003 JIM for CDF 16 Using JIM – Downloading output

22 nd September 2003 JIM for CDF 17 What will JIM do once complete? Current functionality allows job submission and retrieval Next steps: –Making the web-download secure –User support and defect fixing through testing phase –Roll out of SAMGrid to all CDF sites with available resources –Add more brokering criteria Directors review of Run II computing has recommended the expansion of SAM to be a lab- wide product

22 nd September 2003 JIM for CDF 18 Deployment Plan Two plans running in parallel: Testing JIM with non-trivial submissions: –Large Monte Carlo Simulation –Reconstruction –Analysis JIM software installations: –Glasgow ScotGrid IBM and CDFG Dell cluster –Oxford Dell cluster –Liverpool Dell cluster –UCL Dell cluster –Wisconsin In progress Complete Planned for October

22 nd September 2003 JIM for CDF 19 Deployment Plan Glasgow University has a complete installation of SAMGrid on both ScotGrid and the CDF cluster. This installation is being tested with Monte Carlo simulation. Oxford University will be installed in the first week of October. Other UK institutions and Wisconsin to follow

22 nd September 2003 JIM for CDF 20 Credits Thanks to Gabriele Garzoglio and the JIM team at Fermilab for providing material for this presentation.