6-Mar-2003Grids NOW 1 Grids Now: The USCMS Integration Grid Testbed and the European Data Grid Michael Ernst.

Slides:



Advertisements
Similar presentations
Claudio Grandi INFN Bologna DataTAG WP4 meeting, Bologna 14 jan 2003 CMS Grid Integration Claudio Grandi (INFN – Bologna)
Advertisements

CERN LCG Overview & Scaling challenges David Smith For LCG Deployment Group CERN HEPiX 2003, Vancouver.
CMS Applications Towards Requirements for Data Processing and Analysis on the Open Science Grid Greg Graham FNAL CD/CMS for OSG Deployment 16-Dec-2004.
Réunion DataGrid France, Lyon, fév CMS test of EDG Testbed Production MC CMS Objectifs Résultats Conclusions et perspectives C. Charlot / LLR-École.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
June 2003 Iosif Legrand MONitoring Agents using a Large Integrated Services Architecture Iosif Legrand California Institute of Technology.
GRID DATA MANAGEMENT PILOT (GDMP) Asad Samar (Caltech) ACAT 2000, Fermilab October , 2000.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Experience with ATLAS Data Challenge Production on the U.S. Grid Testbed Kaushik De University of Texas at Arlington CHEP03 March 27, 2003.
October 2003 Iosif Legrand Iosif Legrand California Institute of Technology.
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
US CMS Testbed A Grid Computing Case Study Alan De Smet Condor Project University of Wisconsin at Madison
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
Vladimir Litvin, Harvey Newman Caltech CMS Scott Koranda, Bruce Loftis, John Towns NCSA Miron Livny, Peter Couvares, Todd Tannenbaum, Jamie Frey Wisconsin.
CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 CMS Applications Progress towards GridPP milestones Data management.
Dave Newbold, University of Bristol24/6/2003 CMS MC production tools A lot of work in this area recently! Context: PCP03 (100TB+) just started Short-term.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
Grid Testbed Activities in US-CMS Rick Cavanaugh University of Florida 1. Infrastructure 2. Highlight of Current Activities 3. Future Directions NSF/DOE.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
HEP Experiment Integration within GriPhyN/PPDG/iVDGL Rick Cavanaugh University of Florida DataTAG/WP4 Meeting 23 May, 2002.
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
ACAT 2003 Iosif Legrand Iosif Legrand California Institute of Technology.
Ramiro Voicu December Design Considerations  Act as a true dynamic service and provide the necessary functionally to be used by any other services.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Claudio Grandi INFN Bologna CHEP'03 Conference, San Diego March 27th 2003 Plans for the integration of grid tools in the CMS computing environment Claudio.
Wenjing Wu Andrej Filipčič David Cameron Eric Lancon Claire Adam Bourdarios & others.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
Grid Workload Management Massimo Sgaravatto INFN Padova.
CMS Stress Test Report Marco Verlato (INFN-Padova) INFN-GRID Testbed Meeting 17 Gennaio 2003.
MAGDA Roger Jones UCL 16 th December RWL Jones, Lancaster University MAGDA  Main authors: Wensheng Deng, Torre Wenaus Wensheng DengTorre WenausWensheng.
Resource Brokering in the PROGRESS Project Juliusz Pukacki Grid Resource Management Workshop, October 2003.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
D0RACE: Testbed Session Lee Lueking D0 Remote Analysis Workshop February 12, 2002.
Giuseppe Codispoti INFN - Bologna Egee User ForumMarch 2th BOSS: the CMS interface for job summission, monitoring and bookkeeping W. Bacchi, P.
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
US CMS Centers & Grids – Taiwan GDB Meeting1 Introduction l US CMS is positioning itself to be able to learn, prototype and develop while providing.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
29 Sept 2004 CHEP04 A. Fanfani INFN Bologna 1 A. Fanfani Dept. of Physics and INFN, Bologna on behalf of the CMS Collaboration Distributed Computing Grid.
April 2003 Iosif Legrand MONitoring Agents using a Large Integrated Services Architecture Iosif Legrand California Institute of Technology.
PPDG February 2002 Iosif Legrand Monitoring systems requirements, Prototype tools and integration with other services Iosif Legrand California Institute.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
10 May 2001WP6 Testbed Meeting1 WP5 - Mass Storage Management Jean-Philippe Baud PDP/IT/CERN.
Peter Couvares Computer Sciences Department University of Wisconsin-Madison Condor DAGMan: Introduction &
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
CMS Production Management Software Julia Andreeva CERN CHEP conference 2004.
U.S. ATLAS Grid Production Experience
ALICE Physics Data Challenge 3
CMS report from FNAL demo week Marco Verlato (INFN-Padova)
US CMS Testbed.
Stephen Burke, PPARC/RAL Jeff Templon, NIKHEF
Report on GLUE activities 5th EU-DataGRID Conference
The LHCb Computing Data Challenge DC06
Presentation transcript:

6-Mar-2003Grids NOW 1 Grids Now: The USCMS Integration Grid Testbed and the European Data Grid Michael Ernst

6-Mar-2003Grids NOW 2

6-Mar-2003Grids NOW 3 Why Grids?  Efficient sharing of distributed heterogeneous compute and storage resources.  Virtual Organizations and Institutional resource sharing  Dynamic reallocation of resources to target specific problems  Collaboration-wide data access and analysis environments  Grid solutions need to be scalable and robust  Must handle petabytes per year  Tens of thousands of CPUs  Tens of thousands of jobs  Grid solutions presented here are supported in part by the EDG, DataTag, PPDG, GriPhyN, and iVDGL.  We are learning a lot from these current efforts for LCG-1!

6-Mar-2003Grids NOW 4 Introduction  US CMS is positioning itself to be able to learn, prototype and develop while providing a production environment to cater to CMS, US CMS and LCG demands  R&D  Integration  Production  3 phased approach is not a new idea, but is mission critical!  Development Grid Testbed (DGT), Integration Grid Testbed (IGT), and Production Grid

6-Mar-2003Grids NOW 5 Commissioning with CMS Monte Carlo Production  USCMS IGT is producing:  1M Egamma “BigJets” events from generation stage all the way through analysis Ntuples Time dominated by GEANT simulation, but no intermediate data is kept Must run on RedHat Linux 6.1 resources for Objectivity license  500K Egamma “BigJets” events at CMSIM stage only  EDG is producing:  1M Egamma “BigJets” events at CMSIM stage only  Comparison of requests:  CMSIM-only involves large intermediate data (~ 1.5 TB) transfers  Full simulation through to Ntuples involves more complex workflow

6-Mar-2003Grids NOW 6 Special IGT Software: MOP  MOP is a system for packaging production processing jobs into DAGManjob descriptions  DAGMan jobs are Directed Acyclic Graphs (DAGs)  MOP uses the following DAG Nodes for each job: Stage-in: Stages in needed application files, scripts, data from the submit host Run: The application(s) run on the remote host Stage-out: The produced data is staged out from the execution site back to the submit host Clean-up: Temporary areas on the remote site are cleansed Publish: Data is published to a GDMP replica catalogue after it is returned Master Site Remote Site 1 IMPALAmop_submitter DAGMan Condor-G GridFTP Batch Queue GridFTP Remote Site N Batch Queue GridFTP

6-Mar-2003Grids NOW 7 Grid Resources  The USCMS Integration Grid Testbed (IGT) comprises:  About 230 CPU (750 MHz equivalent, RedHat Linux 6.1) Additional 80 CPU at 2.4 GHz running RedHat Linux 7.X  About 5 TB local disk space plus Enstore Mass storage at FNAL using dCache  This is the combined USCMS Tier-1/Tier-2 resources: Caltech, Fermilab, U Florida, UC SanDiego, (UW Madison)  The European Data Grid (EDG) comprises:  About 350 CPU ( MHz equivalent, mostly RHL 6.2) Up to 400 additional CPU at Lyon that are shared CERN, CNAF, RAL, (NIKHEF), Legnaro, Lyon  About 3.7 TB local Disk space plus CASTOR Mass Storage at CERN and HPSS at Lyon

6-Mar-2003Grids NOW 8 Current Grid Software  Both grids use Globus and Condor core middleware  USCMS Integration Grid Testbed (IGT)  Using Virtual Data Toolkit (VDT)  Bottom-up” approach Advantage: Bugs have been shaken out of the core middleware products  European Data Grid  Using EDG release 1.3  Enhanced functionality Eg- Using multiple Resource Brokers relying on real time monitoring information to schedule jobs  “Top-down” approach Advantage: More functionality has been added

6-Mar-2003Grids NOW 9 The CMS IGT “Stack” The CMS IGT “stack” comprises nine layers. The Application layer contains only CMS executables. The Job Creation layer comprises CMS provided tools MCRunJob and Impala. Neither MCRunJob nor Impala are specifically “grid aware.” Then there is a DAG Creation layer and a Job Submission layer. Both functionalities are provided by MOP. Jobs are submitted to DAGMAN which, through Condor-G, manages jobs run on remote Globus Job Managers. Finally, there is a local Farm or Batch System used by Globus GRAM to manage jobs. In the case of the IGT, the local Batch manager was always FBSNG or Condor. Scheduling and Integrated monitoring are not present.

6-Mar-2003Grids NOW 10 Monitoring  MonaLisa is used as the primary IGT Grid-wide monitor. Physical parameters: CPU load, network usage, disk space, etc.  Dynamic discovery of monitoring targets and schema  Implemented with Java/Jini with SNMP local monitors  Interfaces to/from other monitoring packages  EDG uses two sources of Monitoring:  EDG Monitoring System (based on Globus Metadata Directory Service) Physical Parameters: CPU Load, network usage, disk space, etc.  BOSS (Batch Object Submission Service) Application level: Running Time, Size of Output, Job Progress, etc Stores information in MySQL DB in real time

6-Mar-2003Grids NOW 11 ML Design Considerations Act as a true dynamic service and provide the necessary functionally to be used by any other services that require such information (Jini, UDDI - WSDL / SOAP) - - mechanism to dynamically discover all the "Farm Units" used by a community - remote event notification for changes in the any system - lease mechanism for each registered unit Allow dynamic configuration and the list of monitor parameters. Integrate existing monitoring tools ( SNMP, Ganglia, Hawkeye …) To provide: - single-farm values and details for each node - network aspect - real time information - historical data and extracted trend information - listener subscription / notification - (mobile) agent filters (algorithms for prediction and decision-support)

6-Mar-2003Grids NOW 12 IGT Results (so far) Time to process 1 event: MHz Speedup: Avg factor of 100 speedup during current run Resources: Approximately 230 MHz equiv. Sustained efficiency: about 43.5% Oct 25

6-Mar-2003Grids NOW 13 EDG CMSIM events vs. time Avg. CPU Utilization was “about the same”

6-Mar-2003Grids NOW 14 Analysis of Problems on IGT  The usual servers die and need to be restarted  But nothing that seems to be related to the load...  Failure semantics currently lean heavily towards automatic resubmission of failed jobs  Sometimes failures are not recognized right away  Need better system for spotting chronic problems BOSS does this already, we aren’t using it because it was still under development when we were planning the IGT  Problems must be better routed to the “right” people  At one point, an application problem was misdiagnosed early as a middleware issue Partly because IGT is currently run by middleware developers!  Once the application expert looked into it, the problem was solved in 90 minutes

6-Mar-2003Grids NOW 15 Analysis of Problems on EDG  The biggest problems related to the Information System:  Symptom: no resources are found Cause: instability of the MDS when it is overloaded  Solution: submitting jobs at a lower rate improves the chances of success  Symptom: the RB gets stuck (no job ever starts) Cause: investigating...  Symptom: grid elements disappear from the II Cause: services on some machines stopped working Solution: restart the services  Symptom: timeouts when copying the input sandbox  Symptom: log file lost (“Stdout does not contain useful data”) Cause: several (no free files/inodes, broken connect. between CE & RB, …)  Problems related to the replica manager:  Symptom: file registration in the RC fails from time to time  None of these problems is a show-stopper and they happen just in a fraction of the jobs!  Fixes are already there for some of them (but not yet deployed)

6-Mar-2003Grids NOW 16 The Integration Grid Testbed (IGT) Resource Allocations (1 GHz equiv. CPU) in 2002/2003 for IGT and Production Grid. (R&D Grid not included.) New resources for Tier-2 are from iVDGL.

6-Mar-2003Grids NOW 17 Comparison to CMS Spring 02  This is really apples and oranges, but...  Average CPU utilization was about 20-25% during Spring02 over all participating CMS Regional Centers  Though it is impossible to extract a concrete number... It really isn’t known how many Spring02 CPUs should be counted in the denominator, estimated CPUs File transfers were much more complicated during Spring02 Objectivity data was kept and archived Different events were processed at different steps  But the current efforts show that the Grids are in the same ballpark!  We still need a factor of ~2.5 for DC04 production  want slightly better than 1 evt/sec

6-Mar-2003Grids NOW 18 Manpower Results (so far)  Estimates of Manpower for the IGT  2.65 FTE equivalent during initial phase and debugging Reported voluntarily in response to a general query  1.1 FTE equivalent during smooth running periods The manager plus periodic small file transfers  Expected to be less than 1 FTE when regular shift procedures are adopted  Caveats:  This is the “second” attempt for the IGT. The first attempt last Spring needed more manpower.  We STILL have a rapidly changing middleware environment  This is not counting general sysadmin support  This is really saying that production ops is becoming less specialized  EDG began in early December, first attempts,  No manpower estimates yet.  A task force has been set up including reps from EDG, CMS, and LCG

6-Mar-2003Grids NOW 19 Next steps  For the EDG:  Deploy the fixes for the problems encountered so far  Put the online monitoring in place  For the IGT:  Deploy fixes for problems uncovered so far  Deploy more functionality  IGT and EDG are preparation for LCG-1  Recently, LCG started participating in the IGT 50 nodes running CMSIM-only production  Getting “top” and “bottom” to meet in the middle

6-Mar-2003Grids NOW 20 Conclusion l Our approach to developing the software systems for the distributed data processing environment adopts “rolling prototyping” –Analyze current practices –Prototyping of the distributed processing environment –Software Support and Transitioning –Servicing external milestones l Next prototype system to be delivered is the US CMS contribution to the LCG Production Grid (June 2003) –CMS will run a large Data Challenge on that system to prove the computing systems (including new object storage solution ?) l This scheme allows us to flexibly react to technology developments AND to changing and developing external requirements