CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 CMS Applications Progress towards GridPP milestones Data management.

Slides:



Advertisements
Similar presentations
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
Advertisements

CMS Applications – Status and Near Future Plans
CMS Report – GridPP Collaboration Meeting IX Peter Hobson, Brunel University4/2/2004 CMS Status Progress towards GridPP milestones Data management – the.
IEEE NSS 2003 Performance of the Relational Grid Monitoring Architecture (R-GMA) CMS data challenges. The nature of the problem. What is GMA ? And what.
CMS Grid Batch Analysis Framework
RLS Production Services Maria Girone PPARC-LCG, CERN LCG-POOL and IT-DB Physics Services 10 th GridPP Meeting, CERN, 3 rd June What is the RLS -
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
ATLAS-Specific Activity in GridPP EDG Integration LCG Integration Metadata.
Dave Newbold, University of Bristol24/6/2003 CMS MC production tools A lot of work in this area recently! Context: PCP03 (100TB+) just started Short-term.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
WP3 RGMA Deployment Laurence Field / RAL Steve Fisher / RAL.
Use of R-GMA in BOSS Henry Nebrensky (Brunel University) VRVS 26 April 2004 Some slides stolen from various talks at EDG 2 nd Review (
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
CHEP – Mumbai, February 2006 The LCG Service Challenges Focus on SC3 Re-run; Outlook for 2006 Jamie Shiers, LCG Service Manager.
INFSO-RI Enabling Grids for E-sciencE SA1: Cookbook (DSA1.7) Ian Bird CERN 18 January 2006.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Cosener’s House – 30 th Jan’031 LHCb Progress & Plans Nick Brook University of Bristol News & User Plans Technical Progress Review of deliverables.
Nick Brook Current status Future Collaboration Plans Future UK plans.
LCG Service Challenge Phase 4: Piano di attività e impatto sulla infrastruttura di rete 1 Service Challenge Phase 4: Piano di attività e impatto sulla.
Claudio Grandi INFN Bologna CHEP'03 Conference, San Diego March 27th 2003 Plans for the integration of grid tools in the CMS computing environment Claudio.
Wenjing Wu Andrej Filipčič David Cameron Eric Lancon Claire Adam Bourdarios & others.
CMS Report – GridPP Collaboration Meeting V Peter Hobson, Brunel University16/9/2002 CMS Status and Plans Progress towards GridPP milestones Workload management.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
CMS Stress Test Report Marco Verlato (INFN-Padova) INFN-GRID Testbed Meeting 17 Gennaio 2003.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
Δ Storage Middleware GridPP10 What’s new since GridPP9? CERN, June 2004.
13 May 2004EB/TB Middleware meeting Use of R-GMA in BOSS for CMS Peter Hobson & Henry Nebrensky Brunel University, UK Some slides stolen from various talks.
Giuseppe Codispoti INFN - Bologna Egee User ForumMarch 2th BOSS: the CMS interface for job summission, monitoring and bookkeeping W. Bacchi, P.
Claudio Grandi INFN Bologna CHEP'03 Conference, San Diego March 27th 2003 BOSS: a tool for batch job monitoring and book-keeping Claudio Grandi (INFN Bologna)
GridPP Presentation to AstroGrid 13 December 2001 Steve Lloyd Queen Mary University of London.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
EGEE is a project funded by the European Union under contract IST R-GMA: Production Services for Information and Monitoring in the Grid John.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
Michael Doherty RAL UK e-Science AHM 2-4 September 2003 SRB in Action.
Tier1A Status Andrew Sansum 30 January Overview Systems Staff Projects.
Presenter Name Facility Name UK Testbed Status and EDG Testbed Two. Steve Traylen GridPP 7, Oxford.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
2-Sep-02Steve Traylen, RAL WP6 Test Bed Report1 RAL and UK WP6 Test Bed Report Steve Traylen, WP6
Partner Logo A Tier1 Centre at RAL and more John Gordon eScience Centre CLRC-RAL HEPiX/HEPNT - Catania 19th April 2002.
 CMS data challenges. The nature of the problem.  What is GMA ?  And what is R-GMA ?  Performance test description  Performance test results  Conclusions.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
WP3 Information and Monitoring Rob Byrom / WP3
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
David Stickland CMS Core Software and Computing
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
A UK Computing Facility John Gordon RAL October ‘99HEPiX Fall ‘99 Data Size Event Rate 10 9 events/year Storage Requirements (real & simulated data)
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Enabling Grids for E-sciencE CMS/ARDA activity within the CMS distributed system Julia Andreeva, CERN On behalf of ARDA group CHEP06.
INFSO-RI Enabling Grids for E-sciencE gLite Test and Certification Effort Nick Thackray CERN.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
The Grid Storage System Deployment Working Group 6 th February 2007 Flavia Donno IT/GD, CERN.
CMS Production Management Software Julia Andreeva CERN CHEP conference 2004.
J Jensen/J Gordon RAL Storage Storage at RAL Service Challenge Meeting 27 Jan 2005.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
EDG Project Conference – Barcelona 13 May 2003 – n° 1 A.Fanfani INFN Bologna – CMS WP8 – Grid Planning in CMS Outline  CMS Data Challenges  CMS Production.
Real Time Fake Analysis at PIC
Moving the LHCb Monte Carlo production system to the GRID
Tim Barrass Split ( ?) between BaBar and CMS projects.
Sergio Fantinel, INFN LNL/PD
UK GridPP Tier-1/A Centre at CLRC
Scalability Tests With CMS, Boss and R-GMA
Simulation use cases for T2 in ALICE
Presentation transcript:

CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 CMS Applications Progress towards GridPP milestones Data management (Bristol) Monitoring (Brunel + Imperial) Bristol, Brunel and Imperial (1.5 GRIDPP FTE in total)

CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 DC04 Pre-Challenge Production Data Challenge: March 2004 nominal (see Hugh Tallini’s talk) An end-to-end test of the CMS offline computing system ‘Play back’ digi data, emulating CMS DAQ -> storage, reconstruction, calibration, data reduction and analysis at T0 & external T1’s Pre-challenge production >70M fully simulated, hit-formatted, digitised events required for DC04 Using both Geant3 and Geant4 simulation; based on POOL persistency UK status In production for ~3 months at RAL T1 (Bristol-managed) & Imperial UK has contributed ~25% of production so far RAL is also a major data store and hosts the central catalogue for current data management solution (SRB) [v. high-profile contribution] Next steps: Digitisation of simulated data – much more demanding of farms Production will continue for the rest of 2003 (though not at all sites) Large-scale replication of digis to CERN (Castor) via WAN

CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 Production Stats We are late! => Need to maximise use of RAL farm until the last minute (November) Hand over resources to Atlas as they are required (i.e. keep the queues full, ramp down CMS production as Atlas ramps up through queue policies). Migrate to LCG farm…?

CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 Data management (wide-area) Short-term solution (PCP03) Using SRB for all data management, across ~20 sites After considerable effort by RAL e-science staff and CMS people: it works very nicely (deployed across all sites in ~10 days). RAL doing a highly professional job of hosting central MCAT The medium term (DC04) Move towards LCG ASAP; Introduce middleware components into the running production, as they are released and tested (LCG timescale?) Potential problem with data management MSS interface timescales (ask technical gurus for details); currently discussing our approach One possibility: Integrate SRB (incl. MSS interface) below LCG RLS Of potential interest to BaBar, Belle, US Grid projects – will discuss at SLAC Abstract submitted to ACAT ’03 Alternative: Each T1 implements its own MSS interface At RAL, will probably be SRB-ADS anyway, since this is tested and working The longer term (analysis of data for Physics TDR): LCG Will need a transparent migration of current catalogues, etc

CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 Data Management (local) Digitisation setup Data (hits) serving for realistic full-pileup (25 overlapping events) digitisation is very demanding Current RAID disk servers + LAN don’t scale to 100’s of CPUs Performance scales roughly as number of spindles, so bigger disks don’t gain us much. Solution: use distributed disk resources, localised in ‘sub-farms’ Use Dcache as the local data management solution FNAL and RAL are the testbeds for this approach POOL POOL release 1.3 now integrated within the CMS COBRA framework Functional / performance testing & development of catalogue handling approach under way within CMS (incl. Bristol) Full integration of POOL catalogue with local + wide area data management is the next step (work within LCG + CMS) Also re-examining data clustering strategy for wide-area optimisation

CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 Stress testing BOSS with RGMA The CMS job submission and monitoring system BOSS is now GRID enabled using the R-GMA middleware from EDG WP3.

CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 Stress testing BOSS with RGMA Static information is relayed via an information service architecture, operating system, CPU details, disk capacity, access policy and application version. query/ response semantics only Dynamic information is relayed via a monitoring service CPU load, fraction of disk used, network speed and application trace data. both query/ response semantics and publish/ subscribe semantics.

CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 Stress testing BOSS with RGMA Use Case Name:Production The production coordinator submits 10,000 production jobs using BOSS ( from a single Grid node. Each job takes of the order of 10 hours to run on a CPU with speed of the order of 1GHz and produces output files of the order of 500Mb. The jobs are likely to be distributed to around 10 sites. Each job may contain up to 20 messages inserted by the physicist for the purposes of alert, or, more rarely, alarm.

CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 Stress testing BOSS with RGMA BOSS DB UI IMPALA/BOSS WN Sandbox BOSS wrapper Job Tee OutFile R-GMA API Farm servlets Receiver servlets Registry Receiver a5b 6 Tested on CMS-LCG0 testbed at IC and Brunel

CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 Stress testing BOSS with RGMA Plausible sensor data volume for a single BOSS job Plan: Submit 50 real production jobs to a local batch system, and deduce an approximation to the distribution of intervals between sensor messages and the size of those messages. The sensor data produced will be fed directly into R-GMA to investigate scaling and failure modes. Results will be presented at the IEEE NSS conference in Oregon in October 2003

CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 Summary Sucesses 1/4 of all pre-production data produced in the UK. SRB for pre-production challenge data management has worked well. POOL release 1.3 now integrated within the CMS COBRA framework. Problems Late start to the pre-production challenge. Some concerns over the stability and scalability of RGMA.