Winnie Lacesso Bristol Storage June 2009. 2 DPM LCG Storage lcgse01 = DPM built in 2005 by Yves Coppens & Pete Gronbech SuperMicro X5DPAGG (Streamline.

Slides:



Advertisements
Similar presentations
Southgrid Status Pete Gronbech: 21 st March 2007 GridPP 18 Glasgow.
Advertisements

SouthGrid Status Pete Gronbech: 12 th March 2008 GridPP 20 Dublin.
Jon Wakelin, Physics & ACRC Bristol. 2 ACRC Server Rooms –PTR – 48 APC water cooled racks (Hot aisle, cold aisle) –MVB – 12 APC water cooled racks (Hot.
Winnie Lacesso Bristol Site Report May Scope User Support / Servers / Config Security / Network UKI-SOUTHGRID-BRIS-HEP Upcoming: major infrastructure.
Edinburgh (ECDF) Update Wahid Bhimji On behalf of the ECDF Team HepSysMan,10 th June 2010 June-10 Hepsysman1Wahid Bhimji - ECDF  Edinburgh Setup  Hardware.
Birmingham site report Lawrie Lowe: System Manager Yves Coppens: SouthGrid support HEP System Managers’ Meeting, RAL, May 2007.
Southwest Tier 2 Center Status Report U.S. ATLAS Tier 2 Workshop - Harvard Mark Sosebee for the SWT2 Center August 17, 2006.
Computer Cluster at UTFSM Yuri Ivanov, Jorge Valencia.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
Northgrid Status Alessandra Forti Gridpp25 Ambleside 25 August 2010.
Computing Resources Joachim Wagner Overview CNGL Cluster MT Group Cluster School Cluster Desktop PCs.
Site Report HEPHY-UIBK Austrian federated Tier 2 meeting
IFIN-HH LHCB GRID Activities Eduard Pauna Radu Stoica.
Site Report US CMS T2 Workshop Samir Cury on behalf of T2_BR_UERJ Team.
SouthGrid Status Pete Gronbech: 4 th September 2008 GridPP 21 Swansea.
Winnie Lacesso Bristol Site Report June Staff & Users Departmental Physics / Networks: JP Melot, Neil Laws (Microsoft); Rhys Morris (Astrophysics.
Southgrid Status Report Pete Gronbech: February 2005 GridPP 12 - Brunel.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
A comparison of distributed data storage middleware for HPC, GRID and Cloud Mikhail Goldshtein 1, Andrey Sozykin 1, Grigory Masich 2 and Valeria Gribova.
Southgrid Technical Meeting Pete Gronbech: 16 th March 2006 Birmingham.
Oxford STEP09 Report Ewan MacMahon/ Pete Gronbech HEPSYSMAN RAL 2nd July 2009.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
30-Jun-04UCL HEP Computing Status June UCL HEP Computing Status April DESKTOPS LAPTOPS BATCH PROCESSING DEDICATED SYSTEMS GRID MAIL WEB WTS.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
StoRM Some basics and a comparison with DPM Wahid Bhimji University of Edinburgh GridPP Storage Workshop 31-Mar-101Wahid Bhimji – StoRM.
SouthGrid Status Pete Gronbech: 2 nd April 2009 GridPP22 UCL.
Oxford Update HEPix Pete Gronbech GridPP Project Manager October 2014.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
D0SAR - September 2005 Andre Sznajder 1 Rio GRID Initiatives : T2-HEPGRID Andre Sznajder UERJ(Brazil)
Tape logging- SAM perspective Doug Benjamin (for the CDF Offline data handling group)
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
Wahid, Sam, Alastair. Now installed on production storage Edinburgh: srm.glite.ecdf.ed.ac.uk  Local and global redir work (port open) e.g. root://srm.glite.ecdf.ed.ac.uk//atlas/dq2/mc12_8TeV/NTUP_SMWZ/e1242_a159_a165_r3549_p1067/mc1.
São Paulo Regional Analysis Center SPRACE Status Report 22/Aug/2006 SPRACE Status Report 22/Aug/2006.
Southgrid Technical Meeting Pete Gronbech: 26 th August 2005 Oxford.
UKI-SouthGrid Update Hepix Pete Gronbech SouthGrid Technical Coordinator April 2012.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Optimisation of Grid Enabled Storage at Small Sites Jamie K. Ferguson University of Glasgow – Jamie K. Ferguson – University.
Tier2 Centre in Prague Jiří Chudoba FZU AV ČR - Institute of Physics of the Academy of Sciences of the Czech Republic.
Southgrid Technical Meeting Pete Gronbech: 24 th October 2006 Cambridge.
KIT – The cooperation of Forschungszentrum Karlsruhe GmbH und Universität Karlsruhe (TH) Hadoop on HEPiX storage test bed at FZK Artem Trunov.
Southgrid Technical Meeting Pete Gronbech: May 2005 Birmingham.
Tier1 Andrew Sansum GRIDPP 10 June GRIDPP10 June 2004Tier1A2 Production Service for HEP (PPARC) GRIDPP ( ). –“ GridPP will enable testing.
CASPUR Site Report Andrei Maslennikov Lead - Systems Amsterdam, May 2003.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
HEP Computing Status Sheffield University Matt Robinson Paul Hodgson Andrew Beresford.
Computational Research in the Battelle Center for Mathmatical medicine.
UKI-SouthGrid Overview and Oxford Status Report Pete Gronbech SouthGrid Technical Coordinator HEPSYSMAN – RAL 10 th June 2010.
Florida Tier2 Site Report USCMS Tier2 Workshop Livingston, LA March 3, 2009 Presented by Yu Fu for the University of Florida Tier2 Team (Paul Avery, Bourilkov.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Progress report from University of Cyprus.
SA1 operational policy training, Athens 20-21/01/05 Presentation of the HG Node “Isabella” and operational experience Antonis Zissimos Member of ICCS administration.
Tested, seen, heard… Andrei Maslennikov Rome, April 2006.
RAL PPD Tier 2 (and stuff) Site Report Rob Harper HEP SysMan 30 th June
3/16/2005CSS/SCS/Farms and Clustered System 1 CDF Farms Transition Old Production farm (fbs/dfarm/mysql) SAM-farm (Condor/CAF/SAM)
BaBar Cluster Had been unstable mainly because of failing disks Very few (
CASTOR Status at RAL CASTOR External Operations Face To Face Meeting Bonny Strong 10 June 2008.
Evangelos Markatos and Charalampos Gkikas FORTH-ICS Athens, th Mar Institute of Computer Science - FORTH Christos.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
Tier2 Centre in Prague Jiří Chudoba FZU AV ČR - Institute of Physics of the Academy of Sciences of the Czech Republic.
Cluster Status & Plans —— Gang Qin
ICEPP, University of Tokyo
SuperB – INFN-Bari Giacinto DONVITO.
Pete Gronbech GridPP Project Manager April 2016
The Beijing Tier 2: status and plans
Moroccan Grid Infrastructure MaGrid
Cluster Active Archive
Computing Board Report CHIPP Plenary Meeting
Oxford Site Report HEPSYSMAN
STORM & GPFS on Tier-2 Milan
UTFSM computer cluster
Presentation transcript:

Winnie Lacesso Bristol Storage June 2009

2 DPM LCG Storage lcgse01 = DPM built in 2005 by Yves Coppens & Pete Gronbech SuperMicro X5DPAGG (Streamline Computing), 2 Intel Xeon 2.8GHz, 2GB RAM, 32-bit SL3.x, Adaptec 39320A Ultra320 dual-channel May 2006:Transtec T6100 = Infortrend/EonStor A08U-G2421 8x400GB RAID5 = 2.2TB usable; 673GB = any-VO, 1.527TB = CMS May 2007: Transtec PV610S = Infortrend A16U-G x750GB as 2xRAID6 = 8.4TB usable, all CMS-only all ext3 filesystems; both RAID arrays nearly full Feb-May 2008: intermittent SCSI problems with 16-bay June 2008: rebuild lgse01 as SL4 32-bit; July-Aug: SCSI problems increase, always w/16-bay, causing errors in dpm filesystems :( Aug: replace Adaptec SCSI ctlr w/LSI: No help. Add +2GB RAM. Sept/Oct/Nov - trying to debug, RAID array rejected 5 disks in 3 months; Vendor finally admits to replace hardware. Arrives in Dec. New hardware replaced January - excellent working since then.

3 HPC-LCG Storage HPC has used DPM = lcgse01 so far HPC uses gpfs so Jon Wakelin looked into StoRM which can (supposedly) leverage gpfs for bulk access (instead of going thru server = bottleneck) lcgse02 = Viglen 1U, X7DBU mobo, 2 x Intel E5405 = 8 x 2.0GHz, 16GB RAM, 2 x 250GB RAID1 disks‏, dual PSU gridftp01 = identical but only 8GB RAM SL bit, gpfs (currently) - kernel versions are constrained by gpfs (currently ) StoRM FrontEnd + Backend on one machine (common config) StoRM supports gsiftp, rfio & file protocol Passing all OPS, LHCb, CMS SAM tests since forever :)

4 GPFS & HPC storage Storageless Physics gpfs cluster = {lcgse02,gridftp01} plus 3 test nodes Storage gpfs cluster = 4 x DDN I/O servers (filers) & 44TB usable Jon got them multiclustered over public network so StoRM can write But after Jon left we found out rfio does not work - must be a config problem with ACLs within gpfs, but we can't find it yet HPC WN gpfs cluster needs to be multiclustered with Storage gpfs cluster, so LCG jobs on WN can ask lcgse02 for file:/ location of their data and access it over gpfs. HPC maintenance outage in May - multiclustering failed with openssl errors - no help from IBM gpfs experts New Storage Admin Bob Cregan will debug it!

5 StoRM SE, GPFS New hardware for HPC CE & StoRM SE, also gridftp server & new MON (syslog, Nagios, etc): X7DBU Xeon E5405 with 2GB RAM/core HPC CE working well except gpfs timeouts – patchy OPS SAM fails Problems with StoRM - gpfs multiclustering not yet working, rfio permission problems (ACLs??) - thought Jon left it in working order but guess not... New Storage Admin (Bob Cregan) will help get gpfs multiclustering working Good performance on new hardware!