Cambridge Site Report Cambridge Site Report HEP SYSMAN, RAL 10-11 th June 2010 Santanu Das Cavendish Laboratory, Cambridge Santanu.

Slides:



Advertisements
Similar presentations
Southgrid Status Pete Gronbech: 21 st March 2007 GridPP 18 Glasgow.
Advertisements

SouthGrid Status Pete Gronbech: 12 th March 2008 GridPP 20 Dublin.
Andrew McNab - Manchester HEP - 17 September 2002 Putting Existing Farms on the Testbed Manchester DZero/Atlas and BaBar farms are available via the Testbed.
GLite Status Stephen Burke RAL GridPP 13 - Durham.
UCL HEP Computing Status HEPSYSMAN, RAL,
Steve Traylen Particle Physics Department Experiences of DCache at RAL UK HEP Sysman, 11/11/04 Steve Traylen
Edinburgh (ECDF) Update Wahid Bhimji On behalf of the ECDF Team HepSysMan,10 th June 2010 June-10 Hepsysman1Wahid Bhimji - ECDF  Edinburgh Setup  Hardware.
HTCondor and the European Grid Andrew Lahiff STFC Rutherford Appleton Laboratory European HTCondor Site Admins Meeting 2014.
Chris Brew RAL PPD Site Report Chris Brew SciTech/PPD.
Birmingham site report Lawrie Lowe: System Manager Yves Coppens: SouthGrid support HEP System Managers’ Meeting, RAL, May 2007.
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
BEIJING-LCG2 Tire2 Grid. Outline  Introduction about IHEP Grid  Fabric infrastructure of BeiJing-Lcg2  Middleware grid-service and supported VO’s.
Wahid Bhimji Andy Washbrook And others including ECDF systems team Not a comprehensive update but what ever occurred to me yesterday.
- CENTRU DE RESURSE GRID ICI - Bucharest 1 Grid site RO-12-ICI Grid site RO-12-ICI Bildea Ana ICI Bucharest, December 5, 2008.
Site Report HEPHY-UIBK Austrian federated Tier 2 meeting
IFIN-HH LHCB GRID Activities Eduard Pauna Radu Stoica.
Andrew McNab - Manchester HEP - 22 April 2002 UK Rollout and Support Plan Aim of this talk is to the answer question “As a site admin, what are the steps.
SouthGrid Status Pete Gronbech: 4 th September 2008 GridPP 21 Swansea.
UKI-SouthGrid Overview Face-2-Face Meeting Pete Gronbech SouthGrid Technical Coordinator Oxford June 2013.
Gareth Smith RAL PPD HEP Sysman. April 2003 RAL Particle Physics Department Site Report.
1 INDIACMS-TIFR TIER-2 Grid Status Report IndiaCMS Meeting, Sep 27-28, 2007 Delhi University, India.
London Tier 2 Status Report GridPP 12, Brunel, 1 st February 2005 Owen Maroney.
Southgrid Status Report Pete Gronbech: February 2005 GridPP 12 - Brunel.
Quarterly report SouthernTier-2 Quarter P.D. Gronbech.
RAL PPD Site Update and other odds and ends Chris Brew.
Southgrid Technical Meeting Pete Gronbech: 16 th March 2006 Birmingham.
David Hutchcroft on behalf of John Bland Rob Fay Steve Jones And Mike Houlden [ret.] * /.\ /..‘\ /'.‘\ /.''.'\ /.'.'.\ /'.''.'.\ ^^^[_]^^^ * /.\ /..‘\
Group Computing Strategy Introduction and BaBar Roger Barlow June 28 th 2005.
BINP/GCF Status Report BINP LCG Site Registration Oct 2009
SouthGrid Status Pete Gronbech: 2 nd April 2009 GridPP22 UCL.
Quarterly report ScotGrid Quarter Fraser Speirs.
UKI-SouthGrid Overview and Oxford Status Report Pete Gronbech SouthGrid Technical Coordinator HEPSYSMAN RAL 30 th June 2009.
Oxford Update HEPix Pete Gronbech GridPP Project Manager October 2014.
INDIACMS-TIFR Tier 2 Grid Status Report I IndiaCMS Meeting, April 05-06, 2007.
Romanian Tier-2 Federation One site for all: RO-07-NIPNE Mihai Ciubancan on behalf of IT Department.
23 April 2002HEP SYSMAN meeting1 Cambridge HEP Group - site report April 2002 John Hill.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
Southgrid Technical Meeting Pete Gronbech: 26 th August 2005 Oxford.
UKI-SouthGrid Update Hepix Pete Gronbech SouthGrid Technical Coordinator April 2012.
CSCS Status Peter Kunszt Manager Swiss Grid Initiative CHIPP, 21 April, 2006.
Tier2 Centre in Prague Jiří Chudoba FZU AV ČR - Institute of Physics of the Academy of Sciences of the Czech Republic.
Southgrid Technical Meeting Pete Gronbech: May 2005 Birmingham.
University of Bristol 5th GridPP Collaboration Meeting 16/17 September, 2002Owen Maroney University of Bristol 1 Testbed Site –EDG 1.2 –LCFG GridPP Replica.
9 May 2006HEP SYSMAN meeting1 Cambridge HEP Group - site report May 2006 John Hill.
Presenter Name Facility Name UK Testbed Status and EDG Testbed Two. Steve Traylen GridPP 7, Oxford.
HEPSYSMAN May 2007 Oxford & SouthGrid Computing Status (Ian McArthur), Pete Gronbech May 2007 Physics IT Services PP Computing.
Grid DESY Andreas Gellrich DESY EGEE ROC DECH Meeting FZ Karlsruhe, 22./
HEP Computing Status Sheffield University Matt Robinson Paul Hodgson Andrew Beresford.
Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.
HEP SYSMAN 23 May 2007 National Grid Service Steven Young National Grid Service Manager Oxford e-Research Centre University of Oxford.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
2-Sep-02Steve Traylen, RAL WP6 Test Bed Report1 RAL and UK WP6 Test Bed Report Steve Traylen, WP6
UK Tier 1 Centre Glenn Patrick LHCb Software Week, 28 April 2006.
Partner Logo A Tier1 Centre at RAL and more John Gordon eScience Centre CLRC-RAL HEPiX/HEPNT - Catania 19th April 2002.
UKI-SouthGrid Overview and Oxford Status Report Pete Gronbech SouthGrid Technical Coordinator HEPSYSMAN – RAL 10 th June 2010.
Status of India CMS Grid Computing Facility (T2-IN-TIFR) Rajesh Babu Muda TIFR, Mumbai On behalf of IndiaCMS T2 Team July 28, 20111Status of India CMS.
Site Report: Prague Jiří Chudoba Institute of Physics, Prague WLCG GridKa+T2s Workshop.
RAL PPD Tier 2 (and stuff) Site Report Rob Harper HEP SysMan 30 th June
BaBar Cluster Had been unstable mainly because of failing disks Very few (
15-Feb-02Steve Traylen, RAL WP6 Test Bed Report1 RAL/UK WP6 Test Bed Report Steve Traylen, WP6 PPGRID/RAL, UK
SL5 Site Status GDB, September 2009 John Gordon. LCG SL5 Site Status ASGC T1 - will be finished before mid September. Actually the OS migration process.
INRNE's participation in LCG Elena Puncheva Preslav Konstantinov IT Department.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
Tier2 Centre in Prague Jiří Chudoba FZU AV ČR - Institute of Physics of the Academy of Sciences of the Czech Republic.
NERSC/LBNL at LBNL in Berkeley October 2009 Site Report Roberto Gomezel INFN 1.
Cambridge Site Report John Hill 20 June 20131SouthGrid Face to Face.
Moroccan Grid Infrastructure MaGrid
Installation/Configuration
Kashif Mohammad VIPUL DAVDA
Presentation transcript:

Cambridge Site Report Cambridge Site Report HEP SYSMAN, RAL th June 2010 Santanu Das Cavendish Laboratory, Cambridge Santanu Das Cavendish Laboratory, Cambridge

HEP, Cavendish Laboratory 2 Man Power: For Group  John Hill – the main sys-admin  Steve Wotton – deputy sys-admin  Kavitha Nirmaladevi (half-time)  Around 1.2FTE of effort (part of Steve and John) For Grid system  Just myself For Group  John Hill – the main sys-admin  Steve Wotton – deputy sys-admin  Kavitha Nirmaladevi (half-time)  Around 1.2FTE of effort (part of Steve and John) For Grid system  Just myself

HEP, Cavendish Laboratory 3 Group System: Hardware/OS  Around 70 desktops (roughly 2:1 Linux:Windows)  Mainly SLC-5.5 and Windows XP (still some SLC4 desktops)  Servers are SLC4 at present, serving ~27TB of storage Present Work  18TB of clustered storage to replace ~7TB of the old storage mentioned earlier  Migrating from Windows 2000 to Windows 2008 domain  Buy ~35TB of storage for LHC n-tuple (and equivalent) Hardware/OS  Around 70 desktops (roughly 2:1 Linux:Windows)  Mainly SLC-5.5 and Windows XP (still some SLC4 desktops)  Servers are SLC4 at present, serving ~27TB of storage Present Work  18TB of clustered storage to replace ~7TB of the old storage mentioned earlier  Migrating from Windows 2000 to Windows 2008 domain  Buy ~35TB of storage for LHC n-tuple (and equivalent)

HEP, Cavendish Laboratory 4 Group System: Network  Gigabit backbone with a 1Gbps connection onto the University network  10Gbps (??) University connection on to JANET Future Plan  Nothing big  Buy ~35TB of storage for LHC n-tuple (and equivalent)  Traffic levels are rising and we may be forced to consider a upgrade Network  Gigabit backbone with a 1Gbps connection onto the University network  10Gbps (??) University connection on to JANET Future Plan  Nothing big  Buy ~35TB of storage for LHC n-tuple (and equivalent)  Traffic levels are rising and we may be forced to consider a upgrade

HEP, Cavendish Laboratory 5 Grid Status [hardware] : Head nodes  Dell PE-1050s (quad-core/8Gb) for CE, SE and UI Worker nodes  33 x Dell PE1950 (2*dual-core GHz; shared with CamGrid)  4 x Viglan (2*quad-core 2.50Ghz)  4 x SunFires (2*quad-core 2.27GHz)  4 x Dell R410 (2*quad-core 2.53Ghz) Storage  108TB online (~100TB reserved for atlas).  ~30TB being used by local project; will be added to the grid soon. Head nodes  Dell PE-1050s (quad-core/8Gb) for CE, SE and UI Worker nodes  33 x Dell PE1950 (2*dual-core GHz; shared with CamGrid)  4 x Viglan (2*quad-core 2.50Ghz)  4 x SunFires (2*quad-core 2.27GHz)  4 x Dell R410 (2*quad-core 2.53Ghz) Storage  108TB online (~100TB reserved for atlas).  ~30TB being used by local project; will be added to the grid soon.

HEP, Cavendish Laboratory 6 Grid Status [middleware] :  gLite 3.2 for the WNs.  gLite 3.1. for MON and UI.  gLite 3.2 for the CE, SE, site-BDII.  DPM v1.7.2 on the head node.  DPM v1.7.3 on of the DPM disk servers  XFS file system for the storage  Condor (v7.2.4) is used as the batch system.  Supported VOs: Mainly Atlas, LHCb and Camont  Additional VO support: Alice, Biomed, Calice, CMS, dteam, euindia, gridpp and obviously ops  gLite 3.2 for the WNs.  gLite 3.1. for MON and UI.  gLite 3.2 for the CE, SE, site-BDII.  DPM v1.7.2 on the head node.  DPM v1.7.3 on of the DPM disk servers  XFS file system for the storage  Condor (v7.2.4) is used as the batch system.  Supported VOs: Mainly Atlas, LHCb and Camont  Additional VO support: Alice, Biomed, Calice, CMS, dteam, euindia, gridpp and obviously ops

HEP, Cavendish Laboratory 7 Grid System: Network  Gigabit backbone with separate 1Gbps connection onto the University network  All the WNs are on the gigabit-network  10Gbps (??) University connection on to JANET Future Plan  Nothing big  Buy ~35TB of storage for LHC n-tuple (and equivalent)  But traffic levels are rising and we may be forced to consider a upgrade Network  Gigabit backbone with separate 1Gbps connection onto the University network  All the WNs are on the gigabit-network  10Gbps (??) University connection on to JANET Future Plan  Nothing big  Buy ~35TB of storage for LHC n-tuple (and equivalent)  But traffic levels are rising and we may be forced to consider a upgrade

HEP, Cavendish Laboratory 8 Grid System [issues]: Network  Middleware is too buggy for Condor  No proper/practical support, yet  All the previously written scripts are almost no longer maintained  Most of the “info-provider” scripts rewritten /modified locally  Every new release breaks the condor-glite integration  Cannot use yaim on CE  Spending too much time on fixing glite scripts rather trying new things.  GridPP4 money, of course. Network  Middleware is too buggy for Condor  No proper/practical support, yet  All the previously written scripts are almost no longer maintained  Most of the “info-provider” scripts rewritten /modified locally  Every new release breaks the condor-glite integration  Cannot use yaim on CE  Spending too much time on fixing glite scripts rather trying new things.  GridPP4 money, of course.

HEP, Cavendish Laboratory 9 Grid System [plans]: Network  Upgrade Condor  More job-slots and disk space  Condor on Cream-CE  Install Scas,glexec Network  Upgrade Condor  More job-slots and disk space  Condor on Cream-CE  Install Scas,glexec

HEP, Cavendish Laboratory 10 Questions??