1st July 2004HEPSYSMAN RAL - Oxford Site Report1 Oxford University Particle Physics Site Report Pete Gronbech Systems Manager.

Slides:



Advertisements
Similar presentations
24th May 2004Hepix Edinburgh - Oxford Site Report1 Oxford University Particle Physics Site Report Pete Gronbech Systems Manager.
Advertisements

Liverpool HEP – Site Report May 2007 John Bland, Robert Fay.
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH Home server AFS using openafs 3 DB servers. Web server AFS Mail Server.
Oxford PP Computing Site Report HEPSYSMAN 28 th April 2003 Pete Gronbech.
23rd April 2002HEPSYSMAN April Oxford University Particle Physics Site Report Pete Gronbech Systems Manager.
Birmingham site report Lawrie Lowe HEP System Managers Meeting, RAL,1 st July 2004.
UCL HEP Computing Status HEPSYSMAN, RAL,
24-Apr-03UCL HEP Computing Status April DESKTOPS LAPTOPS BATCH PROCESSING DEDICATED SYSTEMS GRID MAIL WEB WTS SECURITY SOFTWARE MAINTENANCE BACKUP.
RAL Particle Physics Dept. Site Report. Gareth Smith RAL PPD About 2 staff mainly on windows and general infrastructure About 1.5 staff on departmental.
A couple of slides on RAL PPD Chris Brew CCLRC - RAL - SPBU - PPD.
Physics Network Integration Chris Hunter. Physics network team Chris Hunter : Network Manager David Newton : Network Support Technician Room DWB 663 Phone.
9th May 2006HEPSYSMAN RAL - Oxford Site Report1 Oxford University Particle Physics Site Report Pete Gronbech Systems Manager.
Report of Liverpool HEP Computing during 2007 Executive Summary. Substantial and significant improvements in the local computing facilities during the.
Chris Brew RAL PPD Site Report Chris Brew SciTech/PPD.
Birmingham site report Lawrie Lowe: System Manager Yves Coppens: SouthGrid support HEP System Managers’ Meeting, RAL, May 2007.
17th October 2013Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and GridPP Project Manager.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Gareth Smith RAL PPD HEP Sysman. April 2003 RAL Particle Physics Department Site Report.
Edinburgh Site Report 1 July 2004 Steve Thorn Particle Physics Experiments Group.
Southgrid Status Report Pete Gronbech: February 2005 GridPP 12 - Brunel.
RHUL1 Site Report Royal Holloway Sukhbir Johal Simon George Barry Green.
Stuart Cunningham - Computer Platforms COMPUTER PLATFORMS Network Operating Systems Week 9.
14th April 1999Hepix Oxford Particle Physics Site Report Pete Gronbech Systems Manager.
UCL Site Report Ben Waugh HepSysMan, 22 May 2007.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
Site report: CERN Helge Meinhard (at) cern ch HEPiX spring CASPUR.
27/04/05Sabah Salih Particle Physics Group The School of Physics and Astronomy The University of Manchester
DECS Community IT DIVISION OF ENGINEERING COMPUTING SERVICES Michigan State University College of Engineering.
April 2001HEPix/HEPNT1 RAL Site Report John Gordon CLRC, UK.
RAL PPD Site Update and other odds and ends Chris Brew.
30-Jun-04UCL HEP Computing Status June UCL HEP Computing Status April DESKTOPS LAPTOPS BATCH PROCESSING DEDICATED SYSTEMS GRID MAIL WEB WTS.
20th October 2003Hepix Vancouver - Oxford Site Report1 Oxford University Particle Physics Site Report Pete Gronbech Systems Manager.
David Hutchcroft on behalf of John Bland Rob Fay Steve Jones And Mike Houlden [ret.] * /.\ /..‘\ /'.‘\ /.''.'\ /.'.'.\ /'.''.'.\ ^^^[_]^^^ * /.\ /..‘\
Group Computing Strategy Introduction and BaBar Roger Barlow June 28 th 2005.
Paul Scherrer Institut 5232 Villigen PSI HEPIX_AMST / / BJ95 PAUL SCHERRER INSTITUT THE PAUL SCHERRER INSTITUTE Swiss Light Source (SLS) Particle accelerator.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
23 Oct 2002HEPiX FNALJohn Gordon CLRC-RAL Site Report John Gordon CLRC eScience Centre.
HEPiX/HEPNT TRIUMF,Vancouver 1 October 18, 2003 NIKHEF Site Report Paul Kuipers
Tier1 Status Report Martin Bly RAL 27,28 April 2005.
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
INDIACMS-TIFR Tier 2 Grid Status Report I IndiaCMS Meeting, April 05-06, 2007.
28 April 2003Imperial College1 Imperial College Site Report HEP Sysman meeting 28 April 2003.
11th Oct 2005Hepix SLAC - Oxford Site Report1 Oxford University Particle Physics Site Report Pete Gronbech Systems Manager and South Grid Technical Co-ordinator.
RAL PPD Computing A tier 2, a tier 3 and a load of other stuff Rob Harper, June 2011.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH OS X Home server AFS using openafs 3 DB servers Kerberos 4 we will move.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
22nd March 2000HEPSYSMAN Oxford Particle Physics Site Report Pete Gronbech Systems Manager.
2-3 April 2001HEPSYSMAN Oxford Particle Physics Site Report Pete Gronbech Systems Manager.
1 PRAGUE site report. 2 Overview Supported HEP experiments and staff Hardware on Prague farms Statistics about running LHC experiment’s DC Experience.
13th October 2011Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and GridPP Project Manager.
14th October 2010Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and SouthGrid Technical Co-ordinator.
Brunel University, School of Engineering and Design, Uxbridge, UB8 3PH, UK Henry Nebrensky (not a systems manager) SIRE Group.
HEPSYSMAN May 2007 Oxford & SouthGrid Computing Status (Ian McArthur), Pete Gronbech May 2007 Physics IT Services PP Computing.
HEP Computing Status Sheffield University Matt Robinson Paul Hodgson Andrew Beresford.
Gareth Smith RAL PPD RAL PPD Site Report. Gareth Smith RAL PPD RAL Particle Physics Department Overview About 90 staff (plus ~25 visitors) Desktops mainly.
| nectar.org.au NECTAR TRAINING Module 5 The Research Cloud Lifecycle.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
11th October 2012Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and GridPP Project Manager.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
Oct. 6, 1999PHENIX Comp. Mtg.1 CC-J: Progress, Prospects and PBS Shin’ya Sawada (KEK) For CCJ-WG.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
Presentation transcript:

1st July 2004HEPSYSMAN RAL - Oxford Site Report1 Oxford University Particle Physics Site Report Pete Gronbech Systems Manager

1st July 2004HEPSYSMAN RAL - Oxford Site Report2

1st July 2004HEPSYSMAN RAL - Oxford Site Report3 Particle Physics Strategy The Server / Desktop Divide Win 2K PC Linux System Desktops Servers General Purpose Unix Server Group DAQ Systems Mail Server Web Server Windows File Server Win 2K PC Win XP PC Approx 200 Windows 2000 Desktop PC’s with Exceed used to access central Linux systems

1st July 2004HEPSYSMAN RAL - Oxford Site Report4 Central Physics Computing Services l hubs n In last year 7.3M messages were relayed, 73% rejected and 5% were viruses. n Anti-virus and anti-spam measures increasingly important in hubs. Some spam inevitably leaks through and clients need to deal with this in a more intelligent way. l Windows Terminal Servers n Use is still increasing 250 users in last three months out of 750 staff/students. Now Win2k and n Introduced an 8 CPU server (TermservMP). Much more powerful system but still awaiting updated versions of some applications which will run properly on OS. l Web / Database n New web server (Windows 2003) in service. n New web applications for lecture lists, Computer inventory, admissions and finals l Exchange Servers n Running two new servers using Exchange 2003 running on Windows server Much better Web interface, support for mobile devices (oma) and for tunnelling through firewalls. l Desktops n Windows XP pro is default OS for new desktops and laptops.

1st July 2004HEPSYSMAN RAL - Oxford Site Report5 Linux l Central Unix systems are Linux based n Red Hat Linux 7.3 is the standard n Treat Linux as just another Unix and hence a server OS to be managed centrally. n Wish to avoid badly managed desktop PC’s running Linux. l Linux based file server (April 2002) l General purpose Linux server installed August 2002 l Batch farm installed

1st July 2004HEPSYSMAN RAL - Oxford Site Report6 pplx1morpheus pplxfs1pplxgen pplx2 1Gb/s ppcresst1ppcresst2 ppatlas1atlassbc ppminos1ppminos2 gridtbwn01 pptb01 pptb02 Grid Development pplx3 CDF minos DAQ Atlas DAQ cresst DAQ General Purpose Systems tblcfgsece RH 7.3 Fermi RH 7.3 Fermi PBS Batch Farm 4*Dual 2.4GHz systems RH 7.3 Autumn *Dual 2.4GHz systems RH 7.3 Autumn 2003 matrix LCG LCG2 Oxford Tier 2 - LCG2

1st July 2004HEPSYSMAN RAL - Oxford Site Report7 The Linux File Server: pplxfs1 8*146GB SCSI disks Dual 1GHz PIII, 1GB RAM

1st July 2004HEPSYSMAN RAL - Oxford Site Report8 New Eonstor IDE RAID array added in April * 250GB disks gives approx 4TB for around £6k. This is our second foray into IDE storage. So far so good.

1st July 2004HEPSYSMAN RAL - Oxford Site Report9 General Purpose Linux Server : pplxgen pplxgen is a Dual 2.2GHz Pentium 4 Xeon based system with 2GB ram. It is running Red Hat 7.3 It was brought on line at the end of August Provides interactive login facilities for code development and test jobs. Long jobs should be sent to the batch queues. Memory to be upgraded to 4GB next week.

1st July 2004HEPSYSMAN RAL - Oxford Site Report10 PP batch farm running Red Hat 7.3 with Open PBS can be seen below pplxgen This service became fully operational in Feb Additional 4 worker nodes were installed in October These are 1U servers and are mounted at the top of the rack. Miscellaneous other nodes bring a total of 21 cpu’s available to PBS.

1st July 2004HEPSYSMAN RAL - Oxford Site Report11

1st July 2004HEPSYSMAN RAL - Oxford Site Report12 CDF Linux Systems Morpheus is an IBM x370 8 way SMP 700MHz Xeon with 8GB RAM and 1TB Fibre Channel disks Installed August 2001 Purchased as part of a JIF grant for the CDF group Runs Fermi Red Hat Uses CDF software developed at Fermilab and Oxford to process data from the CDF experiment.

1st July 2004HEPSYSMAN RAL - Oxford Site Report13 Approx 7.5 TB for SCSI RAID 5 disks are attached to the master node. Each shelf holds 14 * 146GB disks. These are shared via NFS with the worker nodes. OpenPBS batch queuing software is used. Second round of CDF JIF tender: Dell Cluster - MATRIX 10 Dual 2.4GHz P4 Xeon servers running Fermi Linux and SCALI cluster software. Installed December 2002

1st July 2004HEPSYSMAN RAL - Oxford Site Report14 Plenty of space in the second rack for expansion of the cluster. Additional Disk Shelf with 14*146GB plus an extra node was installed in Autumn 2003.

1st July 2004HEPSYSMAN RAL - Oxford Site Report15 Oxford Tier 2 centre for LHC Two racks each containing 20 Dell dual 2.8GHz Xeon’s with SCSI system disks. 1.6TB SCSI disk array in each rack. Systems will be loaded with LCG2 software. SCSI disks and Broadcom Gigabit Ethernet causes some problems with installation. Slow progress being made.

1st July 2004HEPSYSMAN RAL - Oxford Site Report16 Problems of Space, Power and Cooling. Second rack currently temporarily located in theoretical physics computer room. A proposal for a new purpose built computer room on Level 1 (underground) in progress. False floor, large Air conditioning units and power for approx racks to be provided. 1200W/sq m max air cooling, a rack full of 1U servers can create 10KW of heat. Water cooling??

1st July 2004HEPSYSMAN RAL - Oxford Site Report17 OLD Grid development systems. EDG Test bed setup, currently

1st July 2004HEPSYSMAN RAL - Oxford Site Report18 Tape Backup is provided by a Qualstar TLS4480 tape robot with 80 slots and Dual Sony AIT3 drives. Each tape can hold 100GB of data. Installed Jan Netvault 7.1 Software from BakBone is used, running on morpheus, for backup of both cdf and particle physics systems. Main userdisks backed up every weekday night data disks not generally backed up BUT weekly backups to OUCS HFS service provide some security.

Network Access Campus Backbone Router Super Janet 4 2.4Gb/s with Super Janet 4 OUCS Firewall depts Physics Firewall Physics Backbone Router 100Mb/s 1Gb/s 100Mb/s 1Gb/s Backbone Edge Router depts 100Mb/s depts 100Mb/s Backbone Edge Router 1Gb/s

Physics Backbone Upgrade to Gigabit Autumn 2002 desktop Server switch Physics Firewall Physics Backbone Router 1Gb/s 100Mb/s Particle Physics desktop 100Mb/s 1Gb/s 100Mb/s Clarendon Lab 1Gb/s Linux Server Win 2k Server Astro 1Gb/s Theory 1Gb/s Atmos 1Gb/s

1st July 2004HEPSYSMAN RAL - Oxford Site Report21 Network l Gigabit network installed for the physics backbone. l Most PP servers are now interconnected via gigabit. l Many switches have been upgraded to provide 100 mpbs to almost every port with gigabit uplinks to the core network. l Connection to campus remains at 100 mbps, campus upgrade to 10Gbps core not expected till end of l Virtual Private Network (VPN) server getting increased usage, overcomes problems getting some protocols through firewalls. Allows authorised users to get into the Physics network from remote sites, but it has its own security risks…..

1st July 2004HEPSYSMAN RAL - Oxford Site Report22 Network Security l Constantly under threat from worms and viruses. Boundary Firewall’s don’t solve the problem entirely as people bring infections in on laptops. l New firewall based on stateful inspection. Policy is now `default closed`. Some teething problems as we learnt what protocols were required but there has been a very significant improvement in security. l Main firewall passes average 5.8GB/hour (link saturates at peak). Rejects 26,000 connection per hour (7 per second). Mischievous connects rejected 1500/hour, one every 2.5 secs. During blaster worm this reached 80/sec. l Additional firewalls installed to protect the Atlas construction area and to protect us from attacks via dialup or VPN. l Need better control over how laptops access our network. Migrating to a new Network Address Translation system so all portables connect through a managed `gateway`. l Have made it easier to keep Anti-Virus software (Sophos) uptodate via simply connecting to a web page. Important that everyone managing their own machines takes advantage of this. Very useful for both laptops and home systems l Keeping OS’s patched is a major challenge. Easier when machines are all inside one management domain but is still very time consuming. Must compare to perhaps 1-few man months of IT support staff effort to clean out a successful worm from the network.

1st July 2004HEPSYSMAN RAL - Oxford Site Report23 Goals for 2004 (Computing) l Continue to improve Network security n Need better tools for OS patch management n Need users to help with their private laptops –Use automatic updates (e.g. Windows Update) –Update Antivirus software regularly n Segment the network by levels of trust n All the above without adding an enormous management overhead ! l Reduce number of OS’s n Remove last NT4 machines and exchange 5.5 n Digital Unix and VMS very nearly gone. n Getting closer to standardising on RH 7.3 especially as the EDG software is now heading that way. l Still finding it very hard to support laptops but now have a standard clone and recommend IBM laptops. l What version of Linux to run ? Currently all 7.3 but what next? l Looking into Single Sign On for PP systems