London Tier 2 Status Report GridPP 12, Brunel, 1 st February 2005 Owen Maroney.

Slides:



Advertisements
Similar presentations
London Tier2 Status O.van der Aa. Slide 2 LT 2 21/03/2007 London Tier2 Status Current Resource Status 7 GOC Sites using sge, pbs, pbspro –UCL: Central,
Advertisements

Workload Management Status of current activity GridPP 13, Durham, 6 th July 2005.
The London Tier 2 31 st Jan 2003David Colling, London Tier 21 Institutes: Brunel, Imperial, Queen Mary, Royal Holloway, University College. Core e-Science.
Applications Area Issues RWL Jones GridPP13 – 5 th June 2005.
LondonGrid Status Duncan Rand. Slide 2 GridPP 21 Swansea LondonGrid Status LondonGrid Five Universities with seven GOC sites –Brunel University –Imperial.
NorthGrid status Alessandra Forti Gridpp12 Brunel, 1 February 2005.
Steve Traylen Particle Physics Department Experiences of DCache at RAL UK HEP Sysman, 11/11/04 Steve Traylen
HTCondor and the European Grid Andrew Lahiff STFC Rutherford Appleton Laboratory European HTCondor Site Admins Meeting 2014.
Quarterly report ScotGrid Quarter Fraser Speirs.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
NorthGrid status Alessandra Forti Gridpp13 Durham, 4 July 2005.
London Tier 2 Status Report GridPP 13, Durham, 4 th July 2005 Owen Maroney, David Colling.
Andrew McNab - Manchester HEP - 22 April 2002 UK Rollout and Support Plan Aim of this talk is to the answer question “As a site admin, what are the steps.
Cambridge Site Report Cambridge Site Report HEP SYSMAN, RAL th June 2010 Santanu Das Cavendish Laboratory, Cambridge Santanu.
Southgrid Status Report Pete Gronbech: February 2005 GridPP 12 - Brunel.
RHUL1 Site Report Royal Holloway Sukhbir Johal Simon George Barry Green.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
Quarterly report SouthernTier-2 Quarter P.D. Gronbech.
Southgrid Technical Meeting Pete Gronbech: 16 th March 2006 Birmingham.
Issues in Milan Two main problems (details in the next slides): – Site excluded from analysis due to corrupted installation of some releases (mainly )
LT 2 London Tier2 Status Olivier van der Aa LT2 Team M. Aggarwal, D. Colling, A. Fage, S. George, K. Georgiou, W. Hay, P. Kyberd, A. Martin, G. Mazza,
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
Quarterly report ScotGrid Quarter Fraser Speirs.
Organisation Management and Policy Group (MPG): Responsible for setting and policy decisions and resolving any issues concerning fractional usage, acceptable.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
GridPP3 Project Management GridPP20 Sarah Pearce 11 March 2008.
Steve Traylen Particle Physics Department EDG and LCG Status 9 th December 2003
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
Southgrid Technical Meeting Pete Gronbech: 26 th August 2005 Oxford.
CCRC’08 Weekly Update Jamie Shiers ~~~ LCG MB, 1 st April 2008.
UK middleware deployment GridPP27 - CERN 15 th September 2011 GridPP27 - CERN 15 th September 2011 Status & plans Jeremy Coles.
Steve Traylen PPD Rutherford Lab Grid Operations PPD Christmas Lectures Steve Traylen RAL Tier1 Grid Deployment
GLite – An Outsider’s View Stephen Burke RAL. January 31 st 2005gLite overview Introduction A personal view of the current situation –Asked to be provocative!
GridPP Building a UK Computing Grid for Particle Physics Professor Steve Lloyd, Queen Mary, University of London Chair of the GridPP Collaboration Board.
London Tier 2 Status Report GridPP 11, Liverpool, 15 September 2004 Ben Waugh on behalf of Owen Maroney.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
INFSO-RI Enabling Grids for E-sciencE Enabling Grids for E-sciencE Pre-GDB Storage Classes summary of discussions Flavia Donno Pre-GDB.
Production Manager’s Report PMB Jeremy Coles 13 rd September 2004.
University of Bristol 5th GridPP Collaboration Meeting 16/17 September, 2002Owen Maroney University of Bristol 1 Testbed Site –EDG 1.2 –LCFG GridPP Replica.
Presenter Name Facility Name UK Testbed Status and EDG Testbed Two. Steve Traylen GridPP 7, Oxford.
Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
IHEP(Beijing LCG2) Site Report Fazhi.Qi, Gang Chen Computing Center,IHEP.
2-Sep-02Steve Traylen, RAL WP6 Test Bed Report1 RAL and UK WP6 Test Bed Report Steve Traylen, WP6
Andrew McNab - Manchester HEP - 17 September 2002 UK Testbed Deployment Aim of this talk is to the answer the questions: –“How much of the Testbed has.
Rutherford Appleton Lab, UK VOBox Considerations from GridPP. GridPP DTeam Meeting. Wed Sep 13 th 2005.
Tier-1 Andrew Sansum Deployment Board 12 July 2007.
Last update 29/01/ :01 LCG 1Maria Dimou- cern-it-gd Maria Dimou IT/GD CERN VOMS server deployment LCG Grid Deployment Board
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
RAL PPD Tier 2 (and stuff) Site Report Rob Harper HEP SysMan 30 th June
BaBar Cluster Had been unstable mainly because of failing disks Very few (
Accounting in LCG/EGEE Can We Gauge Grid Usage via RBs? Dave Kant CCLRC, e-Science Centre.
1 Update at RAL and in the Quattor community Ian Collier - RAL Tier1 HEPiX FAll 2010, Cornell.
INFSO-RI Enabling Grids for E-sciencE gLite Certification and Deployment Process Markus Schulz, SA1, CERN EGEE 1 st EU Review 9-11/02/2005.
SL5 Site Status GDB, September 2009 John Gordon. LCG SL5 Site Status ASGC T1 - will be finished before mid September. Actually the OS migration process.
Enabling Grids for E-sciencE INFSO-RI Enabling Grids for E-sciencE Gavin McCance GDB – 6 June 2007 FTS 2.0 deployment and testing.
CERN IT Department CH-1211 Genève 23 Switzerland t SL(C) 5 Migration at CERN CHEP 2009, Prague Ulrich SCHWICKERATH Ricardo SILVA CERN, IT-FIO-FS.
INFSO-RI Enabling Grids for E-sciencE gLite Test and Certification Effort Nick Thackray CERN.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE Operations: Evolution of the Role of.
J Jensen/J Gordon RAL Storage Storage at RAL Service Challenge Meeting 27 Jan 2005.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
II EGEE conference Den Haag November, ROC-CIC status in Italy
Dissemination and User Feedback Castor deployment team Castor Readiness Review – June 2006.
UK Status and Plans Catalin Condurache – STFC RAL ALICE Tier-1/Tier-2 Workshop University of Torino, February 2015.
18/12/03PPD Christmas Lectures 2003 Grid in the Department A Guide for the Uninvolved PPD Computing Group Christmas Lecture 2003 Chris Brew.
London Tier-2 Quarter Owen Maroney
Update on Plan for KISTI-GSDC
Presentation transcript:

London Tier 2 Status Report GridPP 12, Brunel, 1 st February 2005 Owen Maroney

1 st February 2005GridPP 12: London Tier 2 Status LT2 Sites Brunel University Imperial College London –(including London e-Science Centre) Queen Mary University of London Royal Holloway University of London University College London

1 st February 2005GridPP 12: London Tier 2 Status LT2 Management Management board had first meeting on 3 rd December 2004 –Next meeting 9 th March 2005 Members of management board: –Brunel: Paul Kyberd –IC: John Darlington (& Steve McGough) –RHUL: Michael Green –QMUL: Alex Martin –UCL: Ben Waugh –Chair: David Colling –Secretary: Owen Maroney

1 st February 2005GridPP 12: London Tier 2 Status Brunel 1 WN LCG-2_2_0 –R-GMA installed but not APEL In process of adding 60 WN’s –Issues with private networking attempted to resolve with LCG-2_2_0 –Will now proceed directly to LCG-2_3 –Investigating installation of SL on nodes If goes well will use YAIM If goes badly will use RH7.3 LCFG

1 st February 2005GridPP 12: London Tier 2 Status Imperial College London 66 CPU PBS HEP LCG-2_2_0 –APEL installed –Upgrading to LCG-2_3_0 (this week!) –Will still use RH7.3 LFCGng HEP computing undergoing re-organisation LCG nodes will be incorporated into SGE cluster, and made available to LCG (dependancy on LeSC SGE integration) Will re-install as RHEL OS at that time. London e-Science Centre –Problems over internal re-organisation –SGE farm, 64bit RHEL Problems with default installation tool (APT) supplied by LCG Also LCG-2_3 not supported on 64bit systems Working on deploying LCG-2_3 on 32bit frontend nodes using YUM and RHEL Tarball install on WN. Hope this is binary compatible! –Then need to work on SGE information provider

1 st February 2005GridPP 12: London Tier 2 Status Queen Mary 320 CPU Torque farm OS is Fedora 2 –Currently running LCG-2_1_1 on frontend, LCG-2_0_0 on WN. More up-to-date versions of LCG were not binary compatible with Fedora –Trinity College Dublin have recently provided Fedora port of LCG-2_2_0 and are working on port of LCG-2_3_0 –Will install LCG-2_3_0 frontend as SL3 machines, using yaim. Install LCG-2_2_0 on Fedora WN Upgrade to 2_3_0 on WN when TCD ready.

1 st February 2005GridPP 12: London Tier 2 Status Royal Holloway Little change: 148 CPU PBS farm –APEL installed –But no data reported! Very little manpower available Currently running LCG-2_2_0 –Hoped to upgrade to LCG-2_3_0 during February Late breaking news…. RHUL PBS server hacked and taken offline….

1 st February 2005GridPP 12: London Tier 2 Status University College London UCL-HEP 20 CPU PBS LCG-2_2_0 –In process of upgrading to LCG-2_3_0 –Frontends SL3 using YAIM –WN stay on RH7.3 UCL-CCC 88 CPU PBS LCG-2_2_0 –Running APEL –Upgrade to LCG-2_3_0 SL3 during February

1 st February 2005GridPP 12: London Tier 2 Status Contribution to GridPP Promised vs. Delivered : No change since GridPP11 SitePromisedDelivered CPUkSI2KTBCPUkSI2KTB Brunel IC (HEP) IC (LeSC)916* QMUL444* * RHUL UCL-HEP UCL-CCC192* Total *CPU count includes shared resources where CPU’s are not 100% dedicated to Grid/HEP kSI2K value takes this sharing into account

1 st February 2005GridPP 12: London Tier 2 Status Usage by VO (APEL) Jobs Nov 2004Dec 2004Jan 2005 alice000 atlas cms000 dteam lhcb zeus CPU Nov 2004Dec 2004Jan 2005 alice000 atlas02,710,2764,120,225 cms000 dteam0204,7775,379 lhcb036,337,22714,322,962 zeus0124,5852,983,753

1 st February 2005GridPP 12: London Tier 2 Status Usage by VO (Jobs)

1 st February 2005GridPP 12: London Tier 2 Status Usage by VO (CPU)

1 st February 2005GridPP 12: London Tier 2 Status Site Experiences (I) Storage Elements are all ‘classic’ gridftp servers –Still waiting for deployment release of SRM solution Problems with experiments use of Tier 2 Storage –Assumption: Tier 2 SE used as a import/export buffer for local farm Input data staged in for jobs on farm Output data staged out to long term storage at Tier 0/1 Tier 2 not permanent storage: no backup! –In practice: Grid does not distinguish between SE’s. No automatic data migration tools. No SE “clean-up” tools. All SE’s advertised as “Permanent” by default. –“Volatile” and “Durable” settings only appropriate for SRM? SE’s fill up with data: become ‘read-only’ data servers –Some datafiles left on SE without entry in RLS: dead-space! –One VO can fill an SE blocking all other VO’s »Disk quota integration with information provider Clean-up tools needed to deal with files older than “x” weeks? –Delete from SE and entry in RLS, if another copy exists –Migrate to different (nearest Tier 1?) SE if only copy –But site admin needs to be in all VO’s to do this!

1 st February 2005GridPP 12: London Tier 2 Status Site Experiences (II) Timing and release of LCG-2_3_0 still could have been improved –Information flow (pre-)release still a problem. –But at least a long upgrade period was allowed! –Structure of documentation changed Generally an improvement Some documents clearly not proof read before release BUT: NO LT2 sites have managed to upgrade yet! WHY NOT? –Lot’s of absence over Christmas/New Year period: not really 2 months –Perception that YAIM installation tool was not mature: lots of ‘bugs’ Bugs fixed quickly, but still the temptation to let other sites ‘go first’ YAIM did not originally handle separate CE and PBS server –Most common configuration in LT2! –Still need to schedule time against other constraints Hardware support posts still not appointed Sites still supported on unfunded ‘best-effort’ basis. –Uncertainty at sites if experiments were ready to use SL New release schedule proposed by LCG Deployment at CERN should help –As should appointment of hardware support posts

1 st February 2005GridPP 12: London Tier 2 Status Summary Little change since GridPP11 –R-GMA and APEL installations –Additional resources (Brunel, LeSC) still to come online –Failure to upgrade to LCG-2_3_0 rapidly Significant effort over Summer 2004 put a lot of resources into LCG –But manpower was coming from unfunded ‘best-effort’ –When term-time starts, much less effort available! Maintenance manageable Upgrades difficult Major upgrades very difficult! Use of resources in practice is turning out to be different to expectations!