23 Oct 2002HEPiX FNALJohn Gordon CLRC-RAL Site Report John Gordon CLRC eScience Centre.

Slides:



Advertisements
Similar presentations
S.L.LloydGrid Steering Committee 8 March 2002 Slide 1 Status of GridPP Overview Financial Summary Recruitment Status EU DataGrid UK Grid Status GridPP.
Advertisements

Tony Doyle - University of Glasgow GridPP EDG - UK Contributions Architecture Testbed-1 Network Monitoring Certificates & Security Storage Element R-GMA.
12th September 2002Tim Adye1 RAL Tier A Tim Adye Rutherford Appleton Laboratory BaBar Collaboration Meeting Imperial College, London 12 th September 2002.
13th November 2002Tim Adye1 RAL Tier A Status Tim Adye Rutherford Appleton Laboratory BaBar UK Collaboration Meeting University of Bristol 13 th November.
Andrew McNab - Manchester HEP - 22 April 2002 EU DataGrid Testbed EU DataGrid Software releases Testbed 1 Job Lifecycle Authorisation at your site More.
GridPP News NeSC opening “Media” dissemination Tier 1/A hardware Web pages Collaboration meetings Nick Brook University of Bristol.
Birmingham site report Lawrie Lowe: System Manager Yves Coppens: SouthGrid support HEP System Managers’ Meeting, RAL, May 2007.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
 Changes to sources of funding for computing in the UK.  Past and present computing resources.  Future plans for computing developments. UK Status &
EU funding for DataGrid under contract IST is gratefully acknowledged GridPP Tier-1A Centre CCLRC provides the GRIDPP collaboration (funded.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
08/06/00 LHCb(UK) Meeting Glenn Patrick LHCb(UK) Computing/Grid: RAL Perspective Glenn Patrick Central UK Computing (what.
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
April 2001HEPix/HEPNT1 RAL Site Report John Gordon CLRC, UK.
David Hutchcroft on behalf of John Bland Rob Fay Steve Jones And Mike Houlden [ret.] * /.\ /..‘\ /'.‘\ /.''.'\ /.'.'.\ /'.''.'.\ ^^^[_]^^^ * /.\ /..‘\
Group Computing Strategy Introduction and BaBar Roger Barlow June 28 th 2005.
Paul Scherrer Institut 5232 Villigen PSI HEPIX_AMST / / BJ95 PAUL SCHERRER INSTITUT THE PAUL SCHERRER INSTITUTE Swiss Light Source (SLS) Particle accelerator.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
Robin Middleton RAL/PPD DG Co-ordination Rome, 23rd June 2001.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
March 2003 CERN 1 EDG and AliEn in Prague Dagmar Adamova INP Rez near Prague.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
28 April 2003Imperial College1 Imperial College Site Report HEP Sysman meeting 28 April 2003.
Sejong STATUS Chang Yeong CHOI CERN, ALICE LHC Computing Grid Tier-2 Workshop in Asia, 1 th December 2006.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
Laboratório de Instrumentação e Física Experimental de Partículas GRID Activities at LIP Jorge Gomes - (LIP Computer Centre)
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH OS X Home server AFS using openafs 3 DB servers Kerberos 4 we will move.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
WP8 Meeting Glenn Patrick1 LHCb Grid Activities in UK Grid WP8 Meeting, 16th November 2000 Glenn Patrick (RAL)
October 2002 INFN Catania 1 The (LHCC) Grid Project Initiative in Prague Dagmar Adamova INP Rez near Prague.
19th September 2003Tim Adye1 RAL Tier A Status Tim Adye Rutherford Appleton Laboratory BaBar UK Collaboration Meeting Royal Holloway 19 th September 2003.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
Tier1 Andrew Sansum GRIDPP 10 June GRIDPP10 June 2004Tier1A2 Production Service for HEP (PPARC) GRIDPP ( ). –“ GridPP will enable testing.
Tier1A Status Andrew Sansum 30 January Overview Systems Staff Projects.
RAL Site report John Gordon ITD October 1999
Cluster Configuration Update Including LSF Status Thorsten Kleinwort for CERN IT/PDP-IS HEPiX I/2001 LAL Orsay Tuesday, December 08, 2015.
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
CDF computing in the GRID framework in Santander
RAL Site Report John Gordon HEPiX/HEPNT Catania 17th April 2002.
CASPUR Site Report Andrei Maslennikov Lead - Systems Rome, April 2006.
UK Grid Meeting Glenn Patrick1 LHCb Grid Activities in UK Grid Prototype and Globus Technical Meeting QMW, 22nd November 2000 Glenn Patrick (RAL)
2-Sep-02Steve Traylen, RAL WP6 Test Bed Report1 RAL and UK WP6 Test Bed Report Steve Traylen, WP6
Partner Logo A Tier1 Centre at RAL and more John Gordon eScience Centre CLRC-RAL HEPiX/HEPNT - Catania 19th April 2002.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
Status of India CMS Grid Computing Facility (T2-IN-TIFR) Rajesh Babu Muda TIFR, Mumbai On behalf of IndiaCMS T2 Team July 28, 20111Status of India CMS.
11th April 2003Tim Adye1 RAL Tier A Status Tim Adye Rutherford Appleton Laboratory BaBar UK Collaboration Meeting Liverpool 11 th April 2003.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
BaBar and the GRID Tim Adye CLRC PP GRID Team Meeting 3rd May 2000.
W.A.Wojcik/CCIN2P3, Nov 1, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
15-Feb-02Steve Traylen, RAL WP6 Test Bed Report1 RAL/UK WP6 Test Bed Report Steve Traylen, WP6 PPGRID/RAL, UK
A UK Computing Facility John Gordon RAL October ‘99HEPiX Fall ‘99 Data Size Event Rate 10 9 events/year Storage Requirements (real & simulated data)
Oct. 6, 1999PHENIX Comp. Mtg.1 CC-J: Progress, Prospects and PBS Shin’ya Sawada (KEK) For CCJ-WG.
W.A.Wojcik/CCIN2P3, HEPiX at SLAC, Oct CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
11th September 2002Tim Adye1 BaBar Experience Tim Adye Rutherford Appleton Laboratory PPNCG Meeting Brighton 11 th September 2002.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
CCIN2P3 Site Report - BNL, Oct 18, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center.
Grid Computing at NIKHEF Shipping High-Energy Physics data, be it simulated or measured, required strong national and trans-Atlantic.
SAM at CCIN2P3 configuration issues
UK GridPP Tier-1/A Centre at CLRC
The INFN TIER1 Regional Centre
Kanga Tim Adye Rutherford Appleton Laboratory Computing Plenary
Presentation transcript:

23 Oct 2002HEPiX FNALJohn Gordon CLRC-RAL Site Report John Gordon CLRC eScience Centre

General PP Facilities New UK Supercomputer BaBar TierA Centre Networking

Computing Farm This years new hardware consists of 4 racks holding 156 dual cpu PCs, a total of GHz Pentium III Tualatin cpus. Each box has 1GB of memory, a 40GB internal disk and 100Mb ethernet. Inside The Tape Robot The tape robot was upgraded last year and now uses 60GB STK 9940 tapes. It currently holds 45TB but could hold 330TB when full. UK GridPP Tier1/A Centre at CLRC 40TByte disk-based Disk Farm The new mass storage unit can store 40Tb or raw data after a RAID 5 overhead. The PCs are clustered on network switches with up to 8x1000Mbit ethernet out of each rack. Prototype Tier 1 centre for CERN LHC and FNAL experiments Tier A centre for SLAC BaBar experiment Testbed for EU DataGrid project

Free and out of robot Free 36TB HEP Data

HPCx UK SuperComputer for next 6 years –Collaboration of CLRC Daresbury Laboratory, Edinburgh EPCC. IBM Sited at CLRC-DL Double in performance every 2 years ie 2 upgrades Capability computing –Target to get 50% of the jobs using 50% of the machine Hardware –40x32 IBM pSeries 690 Regatta-H nodes (Power4 CPUs) – GHz cpus – estimated peak performance 6.6TeraFLOPS –IBM Colony switch connects blocks of 8 cpus (ie looks like 160x8, not 40x32) –1280 GB of Memory –2x32 already in place as a migration aid. –Service testing mid November, service December.

HPCx Software –Capability computing on around 1000 high performance CPUs –Terascale Applications team Parallelising applications for 1000’s of CPUs –Different architecture compared to T3E etc –HPCx is a cluster of 32 Processor machines compared to MPP style of T3E –Some MPI operations now very slow (eg barriers, all-to-all communications)

RAL Tier A RAL is TierA Centre for BaBar –Like CC-IN2P3 but concentrating on different data. –Shared resource with LHC and other experiments –Use

Hardware 104 “noma”-like machines allocated to BaBar –156+old farm shared with other experiments –6 BaBar Suns (4-6 CPUs each) 20 TB disk for BaBar –Also using ~10 TB of pool disk for data transfers –All disk servers on Gigabit ethernet –Pretty good server performance … as well as existing RAL facilities –622 Mbits/s network to SLAC and elsewhere –AFS cell –100TB Tape robot –Many years’ experience running BaBar software

Problems Disk problems tracked down to a bad batch of drives –All drives are now being replaced by the manufacturer our disks should be done in ~1 month –By using spare servers, replacement shouldn’t interrupt service Initially suffered from lack of support staff and out- of-hours support (for US hours) –Two new system managers now in post –Two more being recruited (one just for BaBar) –Additional staff have been able to help with problems at weekends –Discussing more formal arrangements

RAL Batch CPU Use

RAL Batch Users (running at least one non-trivial job each week) A total of 113 new BaBar users registered since December

Data at RAL All data in Kanga format is at RAL –19 TB currently on disk Series-8 + series-10 + reskimmed series-10 AllEvents + streams data + signal+generic MC New data copied from SLAC within 1-2 days RAL is now the primary Kanga analysis site –New data is archived to tape at SLAC and then deleted from disk

Changes since July Two new RedHat 6 front-end machines –Dedicated to BaBar use –Login to babar.gridpp.rl.ac.uk Trial RedHat 7.2 service –One front-end and (currently) 5 batch workers –Once we are happy with the configuration, many/all of the rest of the batch workers will be rapidly upgraded ssh AFS token passing installed on front-ends –So, your local (eg. SLAC) token is available when you log in Trial Grid Gatekeeper available (EDG 1.2) –Allows job submission from the Grid Improved new user registration procedures

Plans Upgrade full farm to RedHat 7.2 –Leave RedHat 6 front-end for use with older releases Upgrade Suns to Solaris 8 and integrate into PBS queues Install data dedicated import-export machines –Fast (Gigabit) network connection –Special firewall rules to allow scp, bbftp, bbcp, etc. AFS authentication improvements –PBS token passing and renewal –integrated login (AFS token on login, like SLAC)

Plans Objectivity support –Works now for private federations, but no data import Support Grid “generic accounts”, so special RAL user registration is no longer necessary Procure next batch of hardware –Delivery probably early 2003

Network Tier1 internal networking will be a hybrid of –100Mb to nodes of cpu farms with 1Gb up from switches –1Gb to disk servers –1Gb to tape servers UK academic network SuperJANET4 –2.5Gbit backbone upgrading to 10Gb in 2002 RAL 622Mb into SJ4 upgraded to 2.5Gb June 02 SJ4 has 2.5Gb interconnect to Geant 2.5Gb links to ESnet and Abilene just for research users UK involved in networking development –internal with Cisco on QoS –external with DataTAG Lamda CERN -> Starlight Private connections