Your university or experiment logo here Tier1 Deployment Steve Traylen.

Slides:



Advertisements
Similar presentations
Andrew McNab - Manchester HEP - 17 September 2002 Putting Existing Farms on the Testbed Manchester DZero/Atlas and BaBar farms are available via the Testbed.
Advertisements

12th September 2002Tim Adye1 RAL Tier A Tim Adye Rutherford Appleton Laboratory BaBar Collaboration Meeting Imperial College, London 12 th September 2002.
Steve Traylen Particle Physics Department Experiences of DCache at RAL UK HEP Sysman, 11/11/04 Steve Traylen
HTCondor and the European Grid Andrew Lahiff STFC Rutherford Appleton Laboratory European HTCondor Site Admins Meeting 2014.
Martin Bly RAL CSF Tier 1/A RAL Tier 1/A Status HEPiX-HEPNT NIKHEF, May 2003.
Scheduling under LCG at RAL UK HEP Sysman, Manchester 11th November 2004 Steve Traylen
Andrew McNab - Manchester HEP - 6 November Old version of website was maintained from Unix command line => needed (gsi)ssh access.
HTCondor within the European Grid & in the Cloud
Pilots 2.0: DIRAC pilots for all the skies Federico Stagni, A.McNab, C.Luzzi, A.Tsaregorodtsev On behalf of the DIRAC consortium and the LHCb collaboration.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
1 Linux in the Computer Center at CERN Zeuthen Thorsten Kleinwort CERN-IT.
Integration and Sites Rob Gardner Area Coordinators Meeting 12/4/08.
Monitoring the Grid at local, national, and Global levels Pete Gronbech GridPP Project Manager ACAT - Brunel Sept 2011.
Steve Traylen Particle Physics Department EDG and LCG Status 9 th December 2003
Project Management Sarah Pearce 3 September GridPP21.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Multi-core jobs at the RAL Tier-1 Andrew Lahiff, Alastair Dewhurst, John Kelly February 25 th 2014.
Steve Traylen PPD Rutherford Lab Grid Operations PPD Christmas Lectures Steve Traylen RAL Tier1 Grid Deployment
WLCG GridKa+T2s Workshop Site Report --- Presenter, Site, Country.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Installing, running, and maintaining large Linux Clusters at CERN Thorsten Kleinwort CERN-IT/FIO CHEP
Southgrid Technical Meeting Pete Gronbech: May 2005 Birmingham.
Resource Management Task Report Thomas Röblitz 19th June 2002.
Managing the CERN LHC Tier0/Tier1 centre Status and Plans March 27 th 2003 CERN.ch.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.
2-Sep-02Steve Traylen, RAL WP6 Test Bed Report1 RAL and UK WP6 Test Bed Report Steve Traylen, WP6
Rutherford Appleton Lab, UK VOBox Considerations from GridPP. GridPP DTeam Meeting. Wed Sep 13 th 2005.
Automatic Resource & Usage Monitoring Steve Traylen/Flavia Donno CERN/IT.
Jan 2010 OSG Update Grid Deployment Board, Feb 10 th 2010 Now having daily attendance at the WLCG daily operations meeting. Helping in ensuring tickets.
SAM Sensors & Tests Judit Novak CERN IT/GD SAM Review I. 21. May 2007, CERN.
WLCG Software Lifecycle First ideas for a post EMI approach 0.
LCG Report from GDB John Gordon, STFC-RAL MB meeting February24 th, 2009.
Workload management, virtualisation, clouds & multicore Andrew Lahiff.
LCG CERN David Foster LCG WP4 Meeting 20 th June 2002 LCG Project Status WP4 Meeting Presentation David Foster IT/LCG 20 June 2002.
The GridPP DIRAC project DIRAC for non-LHC communities.
LCG Pilot Jobs and glexec John Gordon.
+ Publishing Your First Post USING WORDPRESS. + A CMS (content management system) is an application that allows you to publish, edit, modify, organize,
Tier1A Status Martin Bly 28 April CPU Farm Older hardware: –108 dual processors (450, 600 and 1GHz) –156 dual processor 1400MHz PIII Recent delivery:
BaBar Cluster Had been unstable mainly because of failing disks Very few (
UKI-LT2-RHUL ATLAS STEP09 report Duncan Rand on behalf of the RHUL Grid team.
15-Feb-02Steve Traylen, RAL WP6 Test Bed Report1 RAL/UK WP6 Test Bed Report Steve Traylen, WP6 PPGRID/RAL, UK
Monitoring with InfluxDB & Grafana
Scientific Computing in PPD and other odds and ends Chris Brew.
1 Update at RAL and in the Quattor community Ian Collier - RAL Tier1 HEPiX FAll 2010, Cornell.
Your university or experiment logo here User Board Glenn Patrick GridPP20, 11 March 2008.
Testing CernVM-FS scalability at RAL Tier1 Ian Collier RAL Tier1 Fabric Team WLCG GDB - September
SL5 Site Status GDB, September 2009 John Gordon. LCG SL5 Site Status ASGC T1 - will be finished before mid September. Actually the OS migration process.
CernVM-FS Infrastructure for EGI VOs Catalin Condurache - STFC RAL Tier1 EGI Webinar, 5 September 2013.
CERN IT Department CH-1211 Genève 23 Switzerland t SL(C) 5 Migration at CERN CHEP 2009, Prague Ulrich SCHWICKERATH Ricardo SILVA CERN, IT-FIO-FS.
The GridPP DIRAC project DIRAC for non-LHC communities.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE Operations: Evolution of the Role of.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
A Computing Tier 2 Node Eric Fede – LAPP/IN2P3. 2 Eric Fede – 1st Chinese-French Workshop Plan What is a Tier 2 –Context and definition To be a Tier 2.
Software Management Workshop Steve Traylen. Software Management(WG5) The aim of the working group is to look at deficiencies in deployed and upcoming.
Using HLRmon for advanced visualization of resource usage Enrico Fattibene INFN - CNAF ISCG 2010 – Taipei March 11 th, 2010.
1 DIRAC Project Status A.Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille 10 March, DIRAC Developer meeting.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarksEGEE-III INFSO-RI MPI on the grid:
Accounting Update John Gordon. Outline Multicore CPU Accounting Developments Cloud Accounting Storage Accounting Miscellaneous.
UK Status and Plans Catalin Condurache – STFC RAL ALICE Tier-1/Tier-2 Workshop University of Torino, February 2015.
CFI 2004 UW A quick overview with lots of time for Q&A and exploration.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Middleware updates Peter Solagna – EGI.eu OMB – 16/07/ /3/
Scheduling systems Carsten Preuß
CREAM Status and Plans Massimo Sgaravatto – INFN Padova
John Gordon, STFC-RAL GDB 10 October 2007
Moving from CREAM CE to ARC CE
The CREAM CE: When can the LCG-CE be replaced?
Future Test Activities SA3 All Hands Meeting Dublin
Adding Computational Resources to SURAgrid (the document) September 27, 2007 Mary Trauner SURA Consultant.
Presentation transcript:

Your university or experiment logo here Tier1 Deployment Steve Traylen.

Your university or experiment logo here Contents Job Efficiencies CPU Fairshare Installation Method at the Tier1. Recent/Upcoming Services.

Your university or experiment logo here Utilisation

Your university or experiment logo here Efficiencies. New plots are now generated regularly. – A report has been written for the Tier1 Board that describes how these are being calculated. – tilisation_OC_August_2005_V1.0.doc

Your university or experiment logo here Overall, (total cpu/total wall)

Your university or experiment logo here Individual VOs

Your university or experiment logo here Individual Jobs

Your university or experiment logo here Fair Share Policies Currently it is completely based on UNIX groupid. Any spare capacity is allocated to over subscribers in proportion to their allocation. This may change…..

Your university or experiment logo here Fair Share Policies (2) We can create accounts in Maui. –e.g. an LHC account and a non-LHC account. –e.g. a grid account and a non Grid account. These accounts can be given allocations. –But we now have competing scheduling regimes. –Each of the regimes can be weighted but the initial values and subsequent tuning is not obvious.

Your university or experiment logo here Fair Share(3) New plots are now being constructed. They will plot: –allocated % for different VOs. –utilised % for different VOs. –allocated/utilised for different VOs.

Your university or experiment logo here How is tier1 deployed. Still done with. –DHCP/PXE/Kickstart and bunch of scripts. YAIM is just one bit. –Local config packages are built. e.g. tier1-ntp-config, tier1-sendmail-config, tier1-lmsensors- config, tier1-yaim-config,…. –Packages maintained with yum/yumit. Batch workers: –Maintained via parallel SSH. –Down nodes are reinstalled routinely. Complicated service nodes: –Maintained by hand, often the only way. –Backed up to tape.

Your university or experiment logo here How the tier1 is deployed. It is simple especially for new starters. –Only bash and rpm building is required. –Flexible, extending what is there is easy. Problems –Things can get out step, though in reality not a huge problem. Configuration in packages helps a lot. –We changed our netmask everywhere via a package upgrade. Farm wide configuration changes are rare.

Your university or experiment logo here Yumit/Pakiti Yumit renamed as Pakiti –Now support yum/up2date/apt-get. Improvements are being added now. –Currently only packages that could be updated published. –It will also display installed packages. This will flag a WN with a missing/extra RPM. Not a security feature but very useful.

Your university or experiment logo here Recent Deployments. FTS v1.3 is now being deployed. –Install complete by Matt Hodges –Understanding is a lot more thorough than with FTS v1.1.2 when it was thrown up. VOBox primarily for Alice. –Install complete by Catalin Condurache. –Testing/Validation with EIS group underway. –Bugs found – new version released today.

Your university or experiment logo here Upcoming Deployments The two most critical boxes at RAL. –PBS scheduler. Recent crashes have stopped submissions for a day or two at least twice in last three months. –DCache head node. Work is now starting on use of the HA-Linux software package to increase redundancy. – –Other (big) sites in the UK will probably want this.