Tony Doyle - University of Glasgow 10 January 2005Oversight Committee GridPP: status report Andy Parker (Taken from GridPP report at OSC by Tony Doyle)

Slides:



Advertisements
Similar presentations
S.L.LloydATSE e-Science Visit April 2004Slide 1 GridPP – A UK Computing Grid for Particle Physics GridPP 19 UK Universities, CCLRC (RAL & Daresbury) and.
Advertisements

Tony Doyle GridPP – From Prototype To Production, HEPiX Meeting, Edinburgh, 25 May 2004.
The Quantum Chromodynamics Grid James Perry, Andrew Jackson, Matthew Egbert, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Tony Doyle - University of Glasgow GridPP EDG - UK Contributions Architecture Testbed-1 Network Monitoring Certificates & Security Storage Element R-GMA.
Tony Doyle Executive Summary, PPARC, MRC London, 15 May 2003.
Your university or experiment logo here What is it? What is it for? The Grid.
Tony Doyle GridPP2 Specification Process Grid Steering Committee Meeting, MRC, London, 18 February 2004.
GridPP Building a UK Computing Grid for Particle Physics A PPARC funded project.
Tony Doyle - University of Glasgow 3 February 2005Science Committee Meeting GridPP Status Report Tony Doyle.
Tony Doyle GridPP2 Proposal, BT Meeting, Imperial, 23 July 2003.
GridPP From Prototype to Production David Britton 21/Sep/06 1.Context – Introduction to GridPP 2.Performance of the GridPP/EGEE/wLCG Grid 3.Some Successes.
15 May 2006Collaboration Board GridPP3 Planning Executive Summary Steve Lloyd.
Andrew McNab - Manchester HEP - 2 May 2002 Testbed and Authorisation EU DataGrid Testbed 1 Job Lifecycle Software releases Authorisation at your site Grid/Web.
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
Plateforme de Calcul pour les Sciences du Vivant SRB & gLite V. Breton.
Tony Doyle “GridPP2 Proposal”, GridPP7 Collab. Meeting, Oxford, 1 July 2003.
S.L.LloydGridPP CB 29 Oct 2002Slide 1 Agenda 1.Introduction – Steve Lloyd 2.Minutes of Previous Meeting (23 Oct 2001) 3.Matters Arising 4.Project Leader's.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow GridPP Vendor Day 30 th April.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
GridPP Steve Lloyd, Chair of the GridPP Collaboration Board.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
Tony Doyle - University of Glasgow 12 January 2005Collaboration Board Beyond GridPP2 Tony Doyle.
Tony Doyle GridPP – From Prototype To Production, GridPP10 Meeting, CERN, 2 June 2004.
Tony Doyle - University of Glasgow 12 January 2005Collaboration Board GridPP: Executive Summary Tony Doyle.
INFSO-RI Enabling Grids for E-sciencE SA1: Cookbook (DSA1.7) Ian Bird CERN 18 January 2006.
May 8, 20071/15 VO Services Project – Status Report Gabriele Garzoglio VO Services Project – Status Report Overview and Plans May 8, 2007 Computing Division,
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
Robin Middleton RAL/PPD DG Co-ordination Rome, 23rd June 2001.
1 st EGEE Conference – April UK and Ireland Partner Dave Kant Deputy ROC Manager.
3 June 2004GridPP10Slide 1 GridPP Dissemination Sarah Pearce Dissemination Officer
Dan Tovey, University of Sheffield GridPP: Experiment Status & User Feedback Dan Tovey University Of Sheffield.
Neil Geddes GridPP-10, June 2004 UK e-Science Grid Dr Neil Geddes CCLRC Head of e-Science Director of the UK Grid Operations Support Centre.
Tony Doyle - University of Glasgow 6 September 2005Collaboration Meeting GridPP Overview (emphasis on beyond GridPP) Tony Doyle.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
Δ Storage Middleware GridPP10 What’s new since GridPP9? CERN, June 2004.
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
LCG LHC Computing Grid Project – LCG CERN – European Organisation for Nuclear Research Geneva, Switzerland LCG LHCC Comprehensive.
GridPP Presentation to AstroGrid 13 December 2001 Steve Lloyd Queen Mary University of London.
EGEE is a project funded by the European Union under contract IST Middleware Planning for LCG/EGEE Bob Jones EGEE Technical Director e-Science.
GridPP Building a UK Computing Grid for Particle Physics Professor Steve Lloyd, Queen Mary, University of London Chair of the GridPP Collaboration Board.
LCG EGEE is a project funded by the European Union under contract IST LCG PEB, 7 th June 2004 Prototype Middleware Status Update Frédéric Hemmer.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
…building the next IT revolution From Web to Grid…
Tony Doyle - University of Glasgow 8 July 2005Collaboration Board Meeting GridPP Report Tony Doyle.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
UK Tier 1 Centre Glenn Patrick LHCb Software Week, 28 April 2006.
Grid Security work in 2004 Andrew McNab Grid Security Research Fellow University of Manchester.
Sep 25, 20071/5 Grid Services Activities on Security Gabriele Garzoglio Grid Services Activities on Security Gabriele Garzoglio Computing Division, Fermilab.
DataGrid is a project funded by the European Commission under contract IST rd EU Review – 19-20/02/2004 The EU DataGrid Project Three years.
The National Grid Service Mike Mineter.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Status of gLite-3.0 deployment and uptake Ian Bird CERN IT LCG-LHCC Referees Meeting 29 th January 2007.
J Jensen/J Gordon RAL Storage Storage at RAL Service Challenge Meeting 27 Jan 2005.
GridPP2 Data Management work area J Jensen / RAL GridPP2 Data Management Work Area – Part 2 Mass storage & local storage mgmt J Jensen
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
Slide § David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow GridPP delivering The UK Grid.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
18/12/03PPD Christmas Lectures 2003 Grid in the Department A Guide for the Uninvolved PPD Computing Group Christmas Lecture 2003 Chris Brew.
Bob Jones EGEE Technical Director
Stephen Pickles Technical Director, GOSC
Middleware Development and Deployment Status
UK GridPP Tier-1/A Centre at CLRC
Building a UK Computing Grid for Particle Physics
LHC Data Analysis using a worldwide computing grid
Collaboration Board Meeting
Presentation transcript:

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee GridPP: status report Andy Parker (Taken from GridPP report at OSC by Tony Doyle)

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Contents What was GridPP1? What is GridPP2? Vision Challenges LCG –Data Challenges –Issues Deployment Status (9/1/05) –Tier-1/A, Tier-2, NGS M/S/N EGEE Middleware Applications Dissemination What lies ahead? Beyond GridPP2 Grid and e-Science Support in 2008 Executive Summary

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee What was GridPP1? A team that built a working prototype grid of significant scale > 2,000 (9,000) CPUs > 1,000 (5,000) TB of available storage > 1,000 (6,000) simultaneous jobs A complex project where 88% of the milestones were completed and all metrics were within specification A Success “The achievement of something desired, planned, or attempted”

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee What is GridPP2? Structures agreed and in place (except LCG phase-2) 253 Milestones, 112 Monitoring Metrics at present. Must deliver a “Production Grid”: robust, reliable, resilient, secure, stable service delivered to end-user applications. The Collaboration aims to develop, deploy and operate a very large Production Grid in the UK for use by the worldwide particle physics community.

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Vision 1.SCALE: GridPP will deliver Grid middleware and hardware infrastructure to enable the construction of a UK Production Grid for the LHC of significant scale. 2.INTEGRATION: The GridPP project is designed to integrate with the existing Particle Physics programme within the UK, thus enabling full use of Grid technology and efficient use of shared resources. 3.DISSEMINATION: The project will disseminate the GridPP deliverables in the multi- disciplinary e-Science environment and will seek to build collaborations with emerging non-PPARC Grid activities both nationally and internationally. 4.UK LHC COMPUTING: The main aim is to provide a computing environment for the UK Particle Physics Community capable of meeting the challenges posed by the unprecedented data, processing and analysis requirements of the LHC experiments. 5.OTHER UK PARTICLE PHYSICS COMPUTING: The process of creating and testing the computing environment for the LHC will naturally support the current and next generation of highly data intensive Particle Physics experiments. 6.EGEE: Grid technology is the framework used to develop the required capability: key components will be developed as part of the EGEE project and elsewhere. 7.LCG: The collaboration builds on the strong computing traditions of the UK at CERN. GridPP will make a strong contribution to the LCG deployment and operations programme. 8.INTEROPERABILITY: The project is integrated with national and international developments from other Grid projects and the GGF in order to ensure a common set of principles, protocols and standards that can support a wide range of applications. 9.INFRASTRUCTURE: Provision is made for a Tier-1 facility at RAL and four Regional Tier-2s, encompassing the collaborating Institutes. 10.OTHER FUNDING: The Tier-1 and Tier-2s will provide a focus for dissemination to the academic and commercial sector and will attract additional funds such that the full programme can be realised.

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Must share data between thousands of scientists with multiple interests link major (Tier-0 [Tier-1]) and minor (Tier-1 [Tier-2]) computer centres ensure all data accessible anywhere, anytime grow rapidly, yet remain reliable for more than a decade cope with different management policies of different centres ensure data security be up and running routinely by 2007 What are the Grid challenges?

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee What are the Grid challenges? Data Management, Security and Sharing 1. Software process 2. Software efficiency 3. Deployment planning 4. Link centres 5. Share data 6. Manage data7. Install software 8. Analyse data9. Accounting 10. Policies

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee LCG Overview By 2007: - 100,000 CPUs - More than 100 institutes worldwide - building on complex middleware being developed in advanced Grid technology projects, both in Europe (Glite) and in the USA (VDT) - prototype went live in September 2003 in 12 countries - Extensively tested by the LHC experiments during this summer

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Data Challenges Ongoing.. Grid and non-Grid Production Grid now significant CMS - 75 M events and 150 TB: first of this year’s Grid data challenges ALICE - 35 CPU Years Phase 1 done Phase 2 ongoing LCG Entering Grid Production Phase..

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Data Challenge 7.7 M GEANT4 events and 22 TB UK ~20% of LCG Ongoing.. (3) Grid Production ~150 CPU years so far Largest total computing requirement Small fraction of what ATLAS need.. Entering Grid Production Phase..

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee LHCb Data Challenge 424 CPU years (4,000 kSI2k months), 186M events UK’s input significant (>1/4 total) LCG(UK) resource: –Tier-1 7.7% –Tier-2 sites: –London 3.9% –South 2.3% –North 1.4% DIRAC: –Imperial 2.0% –L'pool 3.1% –Oxford 0.1% –ScotGrid 5.1% DIRAC alone LCG in action /day LCG paused Phase 1 Completed /day LCG restarted 186 M Produced Events Entering Grid Production Phase..

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Paradigm Shift Transition to Grid… Jun: 80%:20% 25% of DC’04 Aug: 27%:73% 42% of DC’04 May: 89%:11% 11% of DC’04 Jul: 77%:23% 22% of DC’ CPU · Years

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Issues First large-scale Grid production problems being addressed… at all levels “LCG-2 MIDDLEWARE PROBLEMS AND REQUIREMENTS FOR LHC EXPERIMENT DATA CHALLENGES”

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Is GridPP a Grid? 1.Coordinates resources that are not subject to centralized control 2.… using standard, open, general-purpose protocols and interfaces 3.… to deliver nontrivial qualities of service 1.YES. This is why development and maintenance of LCG is important. 2.YES. VDT (Globus/Condor-G) + EDG/EGEE(Glite) ~meet this requirement. 3.YES. LHC experiments data challenges over the summer of

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee GridPP Deployment Status (9/1/05) Three Grids on Global scale in HEP (similar functionality) sitesCPUs LCG (GridPP)90 (16)9000 (2029) Grid3 [USA] NorduGrid GridPP deployment is part of LCG (Currently the largest Grid in the world) The future Grid in the UK is dependent upon LCG releases totalCPUfreeCPUrunJobwaitJobseAvail TBseUsed TBmaxCPUavgCPU Total

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee UK Tier-1/A Centre Rutherford Appleton Laboratory High quality data services National and international role UK focus for international Grid development 1000 CPU 200 TB Disk 60 TB Tape (Capacity 1PB) Grid Resource Discovery Time = 8 Hours 2004 CPU Utilisation2004 Disk Use

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Tier-1/A Capacity Planning Resource planning estimates required for all experiments, esp. BaBar and each of the 4 LHC experiments for Int’l. MoUs ( ) UK author cost model implemented

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee UK Tier-2 Centres ScotGrid Durham, Edinburgh, Glasgow NorthGrid Daresbury, Lancaster, Liverpool, Manchester, Sheffield SouthGrid Birmingham, Bristol, Cambridge, Oxford, RAL PPD, Warwick LondonGrid Brunel, Imperial, QMUL, RHUL, UCL

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Level-2 Grid In future will include services to facilitate collaborative (grid) computing Authentication (PKI X509) Job submission/batch service Resource brokering Authorisation Virtual Organisation management Certificate management Information service Data access/integration (SRB/OGSA-DAI/DQPS) National Registry (of registry’s) Data replication Data caching Grid monitoring Accounting * Leeds Manchester * * Oxford RAL * * DL

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Middleware Development Configuration Management Storage Interfaces Network Monitoring Security Information Services Grid Data Management

Enabling Grids for E-sciencE INFSO-RI LHCC Comprehensive Review – November Prototype Middleware Status & Plans (I) Workload Management –AliEn TaskQueue –EDG WMS (plus new TaskQueue and Information Supermarket) –EDG L&B Computing Element –Globus Gatekeeper + LCAS/LCMAPS  Dynamic accounts (from Globus) –CondorC –Interfaces to LSF/PBS (blahp) –“Pull components”  AliEn CE  gLite CEmon (being configured) Blue: deployed on development testbed Red: proposed

Enabling Grids for E-sciencE INFSO-RI LHCC Comprehensive Review – November Prototype Middleware Status & Plans (II) Storage Element –Existing SRM implementations  dCache, Castor, …  FNAL & LCG DPM –gLite-I/O (re-factored AliEn-I/O) Catalogs –AliEn FileCatalog – global catalog –gLite Replica Catalog – local catalog –Catalog update (messaging) –FiReMan Interface –RLS (globus) Data Scheduling –File Transfer Service (Stork+GridFTP) –File Placement Service –Data Scheduler Metadata Catalog –Simple interface defined (AliEn+BioMed) Information & Monitoring –R-GMA web service version; multi-VO support

Enabling Grids for E-sciencE INFSO-RI LHCC Comprehensive Review – November Prototype Middleware Status & Plans (III) Security –VOMS as Attribute Authority and VO mgmt –myProxy as proxy store –GSI security and VOMS attributes as enforcement  fine-grained authorization (e.g. ACLs)  globus to provide a set-uid service on CE Accounting –EDG DGAS (not used yet) User Interface –AliEn shell –CLIs and APIs –GAS  Catalogs  Integrate remaining services Package manager –Prototype based on AliEn backend –evolve to final architecture agreed with ARDA team

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Application Development ATLAS LHCbCMS BaBar (SLAC) SAMGrid (FermiLab) QCDGridPhenoGrid

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee More Applications ZEUS uses LCG needs the Grid to respond to increasing demand for MC production 5 million Geant events on Grid since August 2004 QCDGrid For UKQCD Currently a 4-site data grid Key technologies used - Globus Toolkit European DataGrid - eXist XML database managing a few hundred gigabytes of data

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Dissemination much has happened.. more people are reading about it.. LHCb-UK members get up to speed with the GridWed 5 Jan 2005 GridPP in PittsburghThu 9 Dec 2004 GridPP website busier than everMon 6 Dec 2004 Optorsim 2.0 releasedWed 24 Nov 2004 ZEUS produces 5 million Grid eventsMon 15 Nov 2004 CERN 50th anniversary receptionTue 26 Oct 2004 GridPP at CHEP'04Mon 18 Oct 2004 LHCb data challenge first phase a success for LCG and UKMon 4 Oct 2004 Networking in Nottingham - GLIF launch meetingMon 4 Oct 2004 GridPP going for Gold - website award at AHMMon 6 Sep 2004 GridPP at the All Hands MeetingWed 1 Sep 2004 R-GMA included in latest LCG releaseWed 18 Aug 2004 LCG2 administrators learn tips and tricks in OxfordTue 27 Jul 2004 Take me to your (project) leaderFri 2 Jul 2004 ScotGrid's 2nd birthday: ScotGrid clocks up 1 million CPU hoursFri 25 Jun 2004 Meet your production managerFri 18 Jun 2004 GridPP10 report and photographsWed 9 Jun 2004 CERN recognizes UK's outstanding contribution to Grid computingWed 2 Jun 2004 UK particle physics Grid takes shapeWed 19 May 2004 A new monitoring map for GridPPMon 10 May 2004 Press reaction to EGEE launchTue 4 May 2004 GridPP at the EGEE launch conferenceTue 27 Apr 2004 LCG2 releasedThu 8 Apr 2004 University of Warwick joins GridPPThu 8 Apr 2004 Grid computing steps up a gear: the start of EGEEThu 1 Apr 2004 EDG gets glowing final reviewMon 22 Mar 2004 Grids and Web Services meeting, 23 April, LondonTue 16 Mar 2004 EU DataGrid Software License approved by OSIFri 27 Feb 2004 GridPP Middleware workshop, March , UCLFri 20 Feb 2004 Version 1.0 of the Optorsim grid simulation tool released by EU DataGridTue 17 Feb 2004 Summary and photographs of the 9th GridPP Collaboration MeetinThu 12 Feb ,976 hits in December

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Annual data storage: PetaBytes per year 100 Million SPECint2000  100,000 PCs (3 GHz Pentium 4) Concorde (15 km) CD stack with 1 year LHC data (~ 20 km) What lies ahead? Some mountain climbing.. Quantitatively, we’re ~9% of the way there in terms of CPU (9,000 ex 100,000) and disk (3 ex 12-14*3 years)… In production terms, we’ve made base camp We are here (1 km) Importance of step-by-step planning… Pre-plan your trip, carry an ice axe and crampons and arrange for a guide…

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Grid and e-Science Support in 2008 What areas require support? IV Running the Tier-1 Data Centre IVHardware annual upgrade IVContribution to Tier-2 Sysman effort  (non-PPARC) hardware IVFrontend Tier-2 hardware IVContribution to Tier-0 support IIIOne M/S/N expert in each of 6 areas IIIProduction manager and four Tier-2 coordinators IIApplication/Grid experts (UK support) IATLAS Computing MoU commitments and support ICMS Computing MoU commitments and support ILHCb Core Tasks and Computing Support IALICE Computing support IFuture experiments adopt e-Infrastructure methods No GridPP management: (assume production mode established + devolved management to Institutes) III. Grid Middleware I. Experiment Layer II. Application Middleware IV. Facilities and Fabrics

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Priorities in context of a financial snapshot in 2008 Grid (£5.6m p.a.) and e-Science (£2.7m p.a.) Assumes no GridPP project management Savings? –EGEE Phase 2 ( ) may contribute –UK e-Science context is 1.NGS (National Grid Service) 2.OMII (Open Middleware Infrastructure Institute) 3.DCC (Digital Curation Centre) Timeline? Grid and e-Science funding requirements To be compared with Road Map: Not a Bid - Preliminary Input

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Grid and e-Science Exploitation Timeline? PPAP initial inputOct 2004 Science Committee initial input PPARC call assessment ( )2005 Science Committee outcomeOct 2005 PPARC call Jan 2006 PPARC close of call May 2006 Assessment Jun-Dec 2006 PPARC outcome Dec 2006 Institute Recruitment/RetentionJan-Aug 2007 Grid and e-Science Exploitation Sep …. Note if the assessment from PPARC internal planning differs significantly from this preliminary advice from PPAP and SC, then earlier planning is required.

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee Executive Summary GRIDPP-PMB-40-EXEC The Grid is a reality A project was/is needed Under control LCG2 support: SC case being written 16 UK sites are on the Grid –MoUs, planning, deployment, monitoring –each underway as part of GridPP2 Developments estd.,R-GMA deployed Glite designed inc. web services Interfaces developed, testing phase Area transformed Initial ideas.. consultation reqd. Introduction Project Management Resources LCG Deployment –Tier-1/A production + Tier-2 resources M/S/N EGEE Applications Dissemination Beyond GridPP2

Tony Doyle - University of Glasgow 10 January 2005Oversight Committee OSC actions Plan for support of physics analysis community in 2007 Better understanding of capacity limits at Tier1 and strategic priorities £1M bid for LCG2 to go to PPARC Science Committee Plan to raise job completion efficiency Statement on gLite OSC views project as excellently managed and very successful