9th May 2006HEPSYSMAN RAL - Oxford Site Report1 Oxford University Particle Physics Site Report Pete Gronbech Systems Manager.

Slides:



Advertisements
Similar presentations
24th May 2004Hepix Edinburgh - Oxford Site Report1 Oxford University Particle Physics Site Report Pete Gronbech Systems Manager.
Advertisements

Liverpool HEP – Site Report May 2007 John Bland, Robert Fay.
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH Home server AFS using openafs 3 DB servers. Web server AFS Mail Server.
Oxford PP Computing Site Report HEPSYSMAN 28 th April 2003 Pete Gronbech.
23rd April 2002HEPSYSMAN April Oxford University Particle Physics Site Report Pete Gronbech Systems Manager.
Birmingham site report Lawrie Lowe HEP System Managers Meeting, RAL,1 st July 2004.
UCL HEP Computing Status HEPSYSMAN, RAL,
24-Apr-03UCL HEP Computing Status April DESKTOPS LAPTOPS BATCH PROCESSING DEDICATED SYSTEMS GRID MAIL WEB WTS SECURITY SOFTWARE MAINTENANCE BACKUP.
RAL Particle Physics Dept. Site Report. Gareth Smith RAL PPD About 2 staff mainly on windows and general infrastructure About 1.5 staff on departmental.
A couple of slides on RAL PPD Chris Brew CCLRC - RAL - SPBU - PPD.
Physics Network Integration Chris Hunter. Physics network team Chris Hunter : Network Manager David Newton : Network Support Technician Room DWB 663 Phone.
Physics Network Integration Chris Hunter. Physics network team Chris Hunter : Network Manager David Newton : Network Support Technician Room DWB 663 Phone.
Report of Liverpool HEP Computing during 2007 Executive Summary. Substantial and significant improvements in the local computing facilities during the.
Chris Brew RAL PPD Site Report Chris Brew SciTech/PPD.
Birmingham site report Lawrie Lowe: System Manager Yves Coppens: SouthGrid support HEP System Managers’ Meeting, RAL, May 2007.
17th October 2013Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and GridPP Project Manager.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Gareth Smith RAL PPD HEP Sysman. April 2003 RAL Particle Physics Department Site Report.
Edinburgh Site Report 1 July 2004 Steve Thorn Particle Physics Experiments Group.
Southgrid Status Report Pete Gronbech: February 2005 GridPP 12 - Brunel.
RHUL1 Site Report Royal Holloway Sukhbir Johal Simon George Barry Green.
14th April 1999Hepix Oxford Particle Physics Site Report Pete Gronbech Systems Manager.
UCL Site Report Ben Waugh HepSysMan, 22 May 2007.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
27/04/05Sabah Salih Particle Physics Group The School of Physics and Astronomy The University of Manchester
Southgrid Technical Meeting Pete Gronbech: 16 th March 2006 Birmingham.
30-Jun-04UCL HEP Computing Status June UCL HEP Computing Status April DESKTOPS LAPTOPS BATCH PROCESSING DEDICATED SYSTEMS GRID MAIL WEB WTS.
20th October 2003Hepix Vancouver - Oxford Site Report1 Oxford University Particle Physics Site Report Pete Gronbech Systems Manager.
David Hutchcroft on behalf of John Bland Rob Fay Steve Jones And Mike Houlden [ret.] * /.\ /..‘\ /'.‘\ /.''.'\ /.'.'.\ /'.''.'.\ ^^^[_]^^^ * /.\ /..‘\
Paul Scherrer Institut 5232 Villigen PSI HEPIX_AMST / / BJ95 PAUL SCHERRER INSTITUT THE PAUL SCHERRER INSTITUTE Swiss Light Source (SLS) Particle accelerator.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
HEPiX/HEPNT TRIUMF,Vancouver 1 October 18, 2003 NIKHEF Site Report Paul Kuipers
Oxford Update HEPix Pete Gronbech GridPP Project Manager October 2014.
Tier1 Status Report Martin Bly RAL 27,28 April 2005.
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility HEPiX – Fall, 2005.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
INDIACMS-TIFR Tier 2 Grid Status Report I IndiaCMS Meeting, April 05-06, 2007.
28 April 2003Imperial College1 Imperial College Site Report HEP Sysman meeting 28 April 2003.
11th Oct 2005Hepix SLAC - Oxford Site Report1 Oxford University Particle Physics Site Report Pete Gronbech Systems Manager and South Grid Technical Co-ordinator.
RAL PPD Computing A tier 2, a tier 3 and a load of other stuff Rob Harper, June 2011.
Using Virtual Servers for the CERN Windows infrastructure Emmanuel Ormancey, Alberto Pace CERN, Information Technology Department.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH OS X Home server AFS using openafs 3 DB servers Kerberos 4 we will move.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
22nd March 2000HEPSYSMAN Oxford Particle Physics Site Report Pete Gronbech Systems Manager.
2-3 April 2001HEPSYSMAN Oxford Particle Physics Site Report Pete Gronbech Systems Manager.
13th October 2011Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and GridPP Project Manager.
1st July 2004HEPSYSMAN RAL - Oxford Site Report1 Oxford University Particle Physics Site Report Pete Gronbech Systems Manager.
Cloning Windows NT Systems Mainly based on experiences at RAL and Oxford.
HEPix April 2006 NIKHEF site report What’s new at NIKHEF’s infrastructure and Ramping up the LCG tier-1 Wim Heubers / NIKHEF (+SARA)
14th October 2010Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and SouthGrid Technical Co-ordinator.
Brunel University, School of Engineering and Design, Uxbridge, UB8 3PH, UK Henry Nebrensky (not a systems manager) SIRE Group.
HEPSYSMAN May 2007 Oxford & SouthGrid Computing Status (Ian McArthur), Pete Gronbech May 2007 Physics IT Services PP Computing.
HEP Computing Status Sheffield University Matt Robinson Paul Hodgson Andrew Beresford.
Gareth Smith RAL PPD RAL PPD Site Report. Gareth Smith RAL PPD RAL Particle Physics Department Overview About 90 staff (plus ~25 visitors) Desktops mainly.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
11th October 2012Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and GridPP Project Manager.
UKI-SouthGrid Overview and Oxford Status Report Pete Gronbech SouthGrid Technical Coordinator HEPSYSMAN – RAL 10 th June 2010.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
RAL PPD Tier 2 (and stuff) Site Report Rob Harper HEP SysMan 30 th June
BaBar Cluster Had been unstable mainly because of failing disks Very few (
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
Physics Network Integration
Belle II Physics Analysis Center at TIFR
Oxford Site Report HEPSYSMAN
Networks Software.
Presentation transcript:

9th May 2006HEPSYSMAN RAL - Oxford Site Report1 Oxford University Particle Physics Site Report Pete Gronbech Systems Manager

9th May 2006HEPSYSMAN RAL - Oxford Site Report2 Physics Department Computing Services l Physics department restructuring. Reduced staff involved in system management by one. l hubs n A lot of work done to simplify the system and reduce manpower requirements. Haven’t had much effort available for anti-spam. Increased use of the Exchange Servers. Now making use of the Oxford University Computing Service Mail hubs to provide a SPAM score. l Windows Terminal Servers n Still a popular service. More use of remote access to user’s own desktops (XP only) l Web / Database n A lot of work around supporting administration and teaching. l Exchange Servers n Increased size of information store disks from GB. l Windows Front End server – WINFE n Access to windows file system via SCP, SFTP or web browser n Access to exchange server (web and outlook) n Access to address lists (LDAP) for , telephone n VPN service

9th May 2006HEPSYSMAN RAL - Oxford Site Report3 Windows Status l Systems are now almost entirely Windows XP on clients and Windows Server 2003 for services. l Windows XP pro is default OS for new desktops and laptops. l More machines brought into centrally managed domain. l More automation of updates, vulnerability scans etc. l More laptops to support Grant yearWindows Desktops Installed Minimum Spec Maximum Spec Laptops (individual) Laptops (pool) 02/0349P4/2.0P4/ /0450P4/2.6P4/ /0545P4/3.0P4/ Software Licenses: Continued NAG deal (libraries only) New deal for Intel compilers run through OSC group Labview, Mathematica, Maple, IDL System management tools, for imaging, backup, anti-spyware MS OS’s and Office covered by Campus select agreement

9th May 2006HEPSYSMAN RAL - Oxford Site Report4 Network l Gigabit connection to campus operational since July Several technical problems with the link delayed this by over half a year. l Gigabit firewall installed. Purchased commercial unit to minimise manpower required for development and maintenance. Juniper ISG 1000 running netscreen. l Firewall also supports NAT and VPN services which is allowing us to consolidate and simplify the network services. l Moving to the firewall NAT has solved a number of problems we were having previously, including unreliability of videoconferencing connections. l Physics-wide wireless network. Installed in DWB public rooms, Martin Wood and Theory. Will install same in AOPP. New firewall provides routing and security for this network.

9th May 2006HEPSYSMAN RAL - Oxford Site Report5 Network Access Campus Backbone Router Super Janet 4 2.4Gb/s with Super Janet 4 OUCS Firewall depts Physics Firewall Physics Backbone Router 1Gb/s 10Gb/s 1Gb/s 10Gb/s Backbone Edge Router depts 100Mb/s 1Gb/s depts 100Mb/s Backbone Edge Router 10Gb/s

9th May 2006HEPSYSMAN RAL - Oxford Site Report6 Network Security l Constantly under threat from vulnerability scans, worms and viruses. We are attacking the problem in several ways n Boundary Firewall’s ( but these don’t solve the problem entirely as people bring infections in on laptops.) – new firewall n Keeping operating systems patched and properly configured - new windows update server n Antivirus on all systems – More use of Sophos but some problems n Spyware detection – anti-spyware software running on all centrally managed systems n Segmentation of the network into trusted and un-trusted sections – new firewall l Strategy n Centrally manage as many machines as possible to ensure they are uptodate and secure – most windows machines moved into domain n Use Network Address Translation (NAT) service to separate centrally managed and `un-trusted` systems into different networks – new firewall plus new Virtual LANs n Continue to lock-down systems by invoking network policies. The client firewall in Windows XP –SP2 is very useful for excluding network based attacks – centralised client firewall policies

9th May 2006HEPSYSMAN RAL - Oxford Site Report7 Particle Physics Linux l Aim to provide general purpose Linux based system for code development and testing and other Linux based applications. l New Unix Admin (Rosario Esposito) has joined us, so we now have more effort to put into improving this system. l New main server installed (ppslgen) running Scientific Linux (SL3) l File server upgraded to SL4 and 6TB disk array added. l Two dual processor worker nodes reclaimed from Atlas Barrel assembly and connected as SL3 worker nodes. l RH7.3 worker nodes migrated to SL3 l Some performance problems with gnome (Solved by using fixed fonts) and SL3 from exceed. Evaluating alternative (NX) for exceed which doesn’t exhibit this problem (also has better integrated ssl). l Tested VMware player, on Windows Desktop, running a SL image, as a batch worker. It is possible but performance an issue.

9th May 2006HEPSYSMAN RAL - Oxford Site Report8 pplx1morpheus pplxfs1pplxgen pplx2 1Gb/s ppcresst1ppcresst2 ppatlas1atlassbc ppminos1ppminos2 pptb02 pplx3 CDF minos DAQ Atlas DAQ cresst DAQ General Purpose Systems RH 7.3 Fermi RH 7.3 gridtbwn01 pptb01 OLD EDG Grid Development/test cluster tblcfgsece RH 7.3 Fermi PBS Batch Farm 4*Dual 2.4GHz systems RH 7.3 Autumn *Dual 2.4GHz systems RH 7.3 Autumn 2003 matrix LCG LCG2 Oxford Tier 2 - LCG2

2nd December TB ppslgen pplxwn06 pplxwn07 pplxwn08 pplxwn09 pplxwn10 pplxwn11 pplxwn05 pplxgen pplxwn01 pplxwn02 pplxwn03 pplxwn04 pplx2 pplx3 pplxfs1 2 * 2.8GHz P4 1 * 2.4GHz P4 2 * 2.4GHz P4 8 * 700MHz P3 Scientific Linux 3 PP Linux Batch Farm (October05) Red Hat * 2.4GHz P4 2 * 2.2GHz P4 2 * 450MHz P3 2 * 800MHz P3 Migration to Scientific Linux 4TB 1.1TB 2 * 1GHz P3 File Server Addition of New 6TB SATA RAID Array

6TB ppslgen pplxwn06 pplxwn07 pplxwn08 pplxwn09 pplxwn10 pplxwn11 pplxwn05 pplxwn01 pplxwn02 pplxwn03 pplxwn04 pplx2 pplx3 pplxfs1 2 * 2.0GHz P4 2 * 2.8GHz P4 1 * 2.4GHz P4 2 * 2.4GHz P4 8 * 700MHz P3 Scientific Linux 3 PP Linux Batch Farm (Dec 05) Red Hat * 2.2GHz P4 2 * 450MHz P3 2 * 800MHz P3 Migration to Scientific Linux 4TB 1.1TB 2 * 1GHz P3 File Server Addition of New 6TB SATA RAID Array pplxwn12 pplxwn13 pplxwn14 pplxwn15 pplxwn16 2 * 2.0GHz P4 2 * 2.8GHz P4 2 * 2.4GHz P4 pplxgen pplxwn17 2 * 2.8GHz P4

9th May Particle Physics General Purpose Linux Servers : pplxgen & ppslgen ppslgen is a dual 2.4GHz Xeon system with 4GB RAM and runs SL It replaces…. pplxgen is a Dual 2.2GHz Pentium 4 Xeon based system with 4GB ram. It is running SL3.0.5 It was brought on line at the end of August Provides interactive login facilities for code development and test jobs. Long jobs should be sent to the batch queues. New Dual (Dual Core) system to be added shortly.

9th May 2006HEPSYSMAN RAL - Oxford Site Report12 PP batch farm now running SL3 with Torque (the replacement for Open PBS) can be seen below pplxgen This service became fully operational in Feb Additional 4 worker nodes were installed in October These are 1U servers and are mounted at the top of the rack. Currently a total of 39 cpu’s available to ppslgen/torque. Particle Physics General Purpose Batch Farm

9th May 2006HEPSYSMAN RAL - Oxford Site Report13 The original 1.1TB of SCSI disks have been supplemented by a new Eonstor SATA RAID array added in April * 250GB disks gives approx 4TB, and another in Sept 05 with 16*400GB disks about 6TB. Particle Physics: Data Storage The Linux File Server: pplxfs1 Dual 1GHz PIII, 1GB RAM

9th May 2006HEPSYSMAN RAL - Oxford Site Report14 CDF Linux Systems Purchased as part of a JIF grant for the CDF group. IBM x370 8way 700MHz Xeon server with 8GB RAM and 1TB Fibre Channel Disks 11 Dell 2.4GHz P4 Xeon servers with 9TB SCSI Disks Runs CDF software developed at Fermilab and Oxford to process data from the CDF experiment. CDF kit is now 3 years old, the servers that were at RAL have been shared among the four universities. Oxford share IBM x370 8 way SMP 700MHz Xeon 5 disk shelves (~3.3TB) 6 Dual 1GHz PIII servers

9th May 2006HEPSYSMAN RAL - Oxford Site Report15 Two racks each containing 20 Dell dual 2.8GHz Xeon’s with SCSI system disks. 1.6TB SCSI disk array in each rack. Systems are loaded with LCG software version The systems have been heavily used by the LHCb, ATLAS and Biomed Data Challenges. Network Monitoring Box installed and working 5 SouthGrid Testzone machines added for pre release testing. Oxford Tier 2 centre for LHC

9th May 2006HEPSYSMAN RAL - Oxford Site Report16 Oxford RALPPDBirmingham Bristol Cambridge SouthGrid Utilization 2005

9th May 2006HEPSYSMAN RAL - Oxford Site Report17 Oxford Computer Room l Modern processors take a lot of power and generate a lot of heat. l We’ve had many problems with air conditioning units and a power trip. l Need new, properly designed and constructed computer room to deal with increasing requirements. l Local work on the design and the Design Office has checked the cooling and air flow. l Plan is to use one of the old target rooms on level 1, for physics and the new Oxford Supercomputer (800 nodes). l Requirements call for power and cooling between of 0.5 and 1MW l SRIF funding has been secured but this now means its all in the hands of the University’s estates. Now unlikely to be ready before ….

HEPSYSMAN RAL - Oxford Site Report18 Level 1 Computer Room?? Last Year you saw the space we could use. There are now 5 racks of computers located here. Tier 2 Rack 2, Clarendon, Oxford grid development, and Ex RAL CDF IBM 8 way Bad News: No Air Conditioning Room also used as a store Conversion of Room is taking time…But SRIF Funding secured to build joint room for Physics and OSC

9th May 2006HEPSYSMAN RAL - Oxford Site Report19 Future Plans/Upgrades l Networking –Increase the gigabit infrastructure –Segmenting Network to manage untrusted machines l Windows –Server Consolidation –Migration of all windows to single physics authentication domain l General PP –Increased disk storage –Better interactive machines (to supplement pplxgen) Dual dual core Opteron? –Scientific Linux on all systems migrate to SL4 in Autumn to match CERN and LCG/glite software l Tier 2 –Similar spend (£70k)2005 +(£70K)2006 delayed until room ready –Already suffering as a result of cooking our kit, 7 psu failures, 4 UPS battery failures and 2 Hard Disk failures in last 4 months.