GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.

Slides:



Advertisements
Similar presentations
S.L.LloydATSE e-Science Visit April 2004Slide 1 GridPP – A UK Computing Grid for Particle Physics GridPP 19 UK Universities, CCLRC (RAL & Daresbury) and.
Advertisements

Particle physics – the computing challenge CERN Large Hadron Collider –2007 –the worlds most powerful particle accelerator –10 petabytes (10 million billion.
Level 1 Components of the Project. Level 0 Goal or Aim of GridPP. Level 2 Elements of the components. Level 2 Milestones for the elements.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Torsten Antoni – LCG Operations Workshop, CERN 02-04/11/04 Global Grid User Support - GGUS -
EGEE SA1 Operations Workshop Stockholm, 13-15/06/2007 Enabling Grids for E-sciencE Service Level Agreement Metrics SLA SA1 Working Group Łukasz Skitał.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 15 th April 2009 Visit of Spanish Royal Academy.
08/11/908 WP2 e-NMR Grid deployment and operations Technical Review in Brussels, 8 th of December 2008 Marco Verlato.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Romanian SA1 report Alexandru Stanciu ICI.
INFSO-RI Enabling Grids for E-sciencE GLOBAL GRID USER SUPPORT THE MODEL AND EXPERIENCE IN LCG/EGEE Gilles Mathieu(1), Torsten Antoni(2),
INFSO-RI Enabling Grids for E-sciencE SA1: Cookbook (DSA1.7) Ian Bird CERN 18 January 2006.
GGF12 – 20 Sept LCG Incident Response Ian Neilson LCG Security Officer Grid Deployment Group CERN.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
Monitoring the Grid at local, national, and Global levels Pete Gronbech GridPP Project Manager ACAT - Brunel Sept 2011.
1 st EGEE Conference – April UK and Ireland Partner Dave Kant Deputy ROC Manager.
11/30/2007 Overview of operations at CC-IN2P3 Exploitation team Reported by Philippe Olivero.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
3 June 2004GridPP10Slide 1 GridPP Dissemination Sarah Pearce Dissemination Officer
Organisation Management and Policy Group (MPG): Responsible for setting and policy decisions and resolving any issues concerning fractional usage, acceptable.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
Project Management Sarah Pearce 3 September GridPP21.
John Gordon CCLRC e-Science Centre LCG Deployment in the UK John Gordon GridPP10.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
Responsibilities of ROC and CIC in EGEE infrastructure A.Kryukov, SINP MSU, CIC Manager Yu.Lazin, IHEP, ROC Manager
The ILC And the Grid Andreas Gellrich DESY LCWS2007 DESY, Hamburg, Germany
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow GridPP Computing for Particle.
11 March 2008 GridPP20 Collaboration meeting David Britton - University of Glasgow GridPP Status GridPP20 Collaboration Meeting, Dublin David Britton,
GridPP Deployment Status GridPP14 Jeremy Coles 6 th September 2005.
GridPP Building a UK Computing Grid for Particle Physics Professor Steve Lloyd, Queen Mary, University of London Chair of the GridPP Collaboration Board.
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
…building the next IT revolution From Web to Grid…
Tony Doyle - University of Glasgow 8 July 2005Collaboration Board Meeting GridPP Report Tony Doyle.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks SA1: Grid Operations Maite Barroso (CERN)
INFSO-RI Enabling Grids for E-sciencE EGEE SA1 in EGEE-II – Overview Ian Bird IT Department CERN, Switzerland EGEE.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks The EGEE User Support Infrastructure Torsten.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGI Operations Tiziana Ferrari EGEE User.
Rutherford Appleton Lab, UK VOBox Considerations from GridPP. GridPP DTeam Meeting. Wed Sep 13 th 2005.
INFSO-RI Enabling Grids for E-sciencE An overview of EGEE operations & support procedures Jules Wolfrat SARA.
LCG WLCG Accounting: Update, Issues, and Plans John Gordon RAL Management Board, 19 December 2006.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Deliverable DSA1.4 Jules Wolfrat ARM-9 –
EGEE-III INFSO-RI Enabling Grids for E-sciencE COD20. June 2009 Helsinki R-COD in UKI Claire Devereux, Jeremy Coles & Co. COD-20,
Accounting in LCG/EGEE Can We Gauge Grid Usage via RBs? Dave Kant CCLRC, e-Science Centre.
INFSO-RI Enabling Grids for E-sciencE User and Virtual Organisation Support in EGEE Flavia Donno, CERN Torsten Antoni, FZK Alistair.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Operations Automation Team Kickoff Meeting.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE Operations: Evolution of the Role of.
J Jensen/J Gordon RAL Storage Storage at RAL Service Challenge Meeting 27 Jan 2005.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE Operational Procedures (Contacts, procedures,
UK Grid Operations Support Centre All slides stolen by P.Clarke from a talk given by: Dr Neil Geddes CCLRC Head of e-Science Director of the UK Grid Operations.
A Computing Tier 2 Node Eric Fede – LAPP/IN2P3. 2 Eric Fede – 1st Chinese-French Workshop Plan What is a Tier 2 –Context and definition To be a Tier 2.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks CYFRONET site report Marcin Radecki CYFRONET.
Slide § David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow GridPP delivering The UK Grid.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Services for Distributed e-Infrastructure Access Tiziana Ferrari on behalf.
Scuola Grid - Martina Franca, Thursday 08 November Il Sistema di Supporto INFNGrid & GGUS ( Global Grid User.
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
18/12/03PPD Christmas Lectures 2003 Grid in the Department A Guide for the Uninvolved PPD Computing Group Christmas Lecture 2003 Chris Brew.
Bob Jones EGEE Technical Director
Regional Operations Centres Core infrastructure Centres
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
User Support Workflow in EGEE
Ian Bird GDB Meeting CERN 9 September 2003
Brief overview on GridICE and Ticketing System
Long-term Grid Sustainability
UK GridPP Tier-1/A Centre at CLRC
UK Status and Plans Scientific Computing Forum 27th Oct 2017
The CCIN2P3 and its role in EGEE/LCG
Pierre Girard ATLAS Visit
Presentation transcript:

GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres in the UK. It includes a LHC Tier-1 regional centre at the Rutherford Appleton Laboratory and 4 regional Tier-2s: ScotGrid, NorthGrid, SouthGrid and London Tier-2. Over the next year the number of CPUs will increase to over 10,000 as the LHC begins operation. The EGEE Computing Grid GridPP manages the UK contribution to the Large Hadron Collider (LHC) computing Grid (LCG) and other UK HEP grid activities. LCG is a major contributor to the Enabling Grids for E-sciencE (EGEE) project. The EGEE grid is currently the largest functioning Grid in the world, with over 40,000 CPUs and over 30 million GB of storage at over 150 sites in 40 countries. Thousands of jobs are run every day on the EGEE production Grid by a wide selection of Virtual Organisation users. Organisation The 19 sites in the UK have formed into 4 regional Tier-2s: ScotGrid, NorthGrid, SouthGrid and the London Tier-2. RAL hosts another site, the GridPP Tier-1 which is managed independently of the other resources. Each Tier-2 has a technical coordinator and a manager to direct the work of the system administrators and monitor provision of resources. There is also a core deployment team who support and advise in areas such as storage, networking and monitoring.. Middleware releases Core grid services are run by the ROCs. They provide information about resources available, location of data and who is allowed to use what. The sites run a specific set of software, the middleware, to provide data for information systems and interpret grid job requests. As this is updated the deployment team work with the sites to ensure problems are solved quickly. Periods of downtime are scheduled in a Grid Operations Centre Database, which is linked into the mechanism users use to submit jobs to prevent a site that is unavailable from receiving any jobs. This database also serves as a useful repository of information about a site which can be used for accounting and monitoring.. Preparing for LCG GridPP deployment and operations needs to ensure the sites are prepared for the volume of data expected when the LHC experiments enter full operation. LCG has focussed on Tier-0 to Tier-1 testing. GridPP has extended this to Tier-1 to Tier-2 testing, to understand bottlenecks in the connecting networks and the optimisation of storage resource management arrangements at sites. Site Monitoring The deployment team is responsible for monitoring and solving problems with the sites on a daily basis. This area is developing quickly with support coming from across the EGEE project. Most of the tools which have been made available rely on information from the Site Availability Monitoring (SAM) tests. The SAM tests are run every 3 hours for each site to test basic functionality. All results from these and other tests are reviewed on a regular basis by a team of operators who spot areas of concern and submit trouble tickets to the Regional Operations Centre (ROC), site or user via the Global Grid User Support (GGUS) centre based in Karlsruhe, Germany. The UK & Ireland helpdesk interfaces directly to the GGUS system. Resource usage GridPP (together with LCG and EGEE) has developed a number of metrics to assist with the ongoing challenge of ensuring that resources are being utilised effectively. They cover areas like percentage of successful jobs, the number of users, and trouble tickets issued to a site. These reveal information which help improve performance and reliability. Computer racks full of worker nodes – a typical scene at each of the GridPP sites A snapshot from the UK developed Real Time Grid Monitor showing sites and jobs on EGEE the grid. GridPP sites and their grouping within regional Tier-2s Screen shots of two site monitoring tools with tests run every few hours to identify problems The results of a set of sustained inter-site data transfer tests during The graph shows the rates achieved first copying data into the Edinburgh storage element across the wide area network and then reading it back out again. All GridPP sites have been tested with such single direction transfers as well as simultaneous inbound and outbound. Example usage views: relative usage of EGEE resources by VO (pie chart) and successful hours of GridPP CPU time by the LHC experiments for Q