9/22/10 OSG Storage Forum 1 CMS Florida T2 Storage Status Bockjoo Kim for the CMS Florida T2.

Slides:



Advertisements
Similar presentations
Storage Workshop Summary Wahid Bhimji University Of Edinburgh On behalf all of the participants…
Advertisements

Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
CHEPREO Tier-3 Center Achievements. FIU Tier-3 Center Tier-3 Centers in the CMS computing model –Primarily employed in support of local CMS physics community.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
NWfs A ubiquitous, scalable content management system with grid enabled cross site data replication and active storage. R. Scott Studham.
Site Report US CMS T2 Workshop Samir Cury on behalf of T2_BR_UERJ Team.
Installing software on personal computer
Data Storage Willis Kim 14 May Types of storages Direct Attached Storage – storage hardware that connects to a single server Direct Attached Storage.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Lustre at Dell Overview Jeffrey B. Layton, Ph.D. Dell HPC Solutions |
1 INDIACMS-TIFR TIER-2 Grid Status Report IndiaCMS Meeting, Sep 27-28, 2007 Delhi University, India.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
Florida Tier2 Site Report USCMS Tier2 Workshop Fermilab, Batavia, IL March 8, 2010 Presented by Yu Fu For the Florida CMS Tier2 Team: Paul Avery, Dimitri.
OSG Public Storage and iRODS
A comparison of distributed data storage middleware for HPC, GRID and Cloud Mikhail Goldshtein 1, Andrey Sozykin 1, Grigory Masich 2 and Valeria Gribova.
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Multi-Tiered Storage with Xrootd at ATLAS Western Tier 2 Andrew Hanushevsky Wei Yang SLAC National Accelerator Laboratory 1CHEP2012, New York
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
Wide Area Network Access to CMS Data Using the Lustre Filesystem J. L. Rodriguez †, P. Avery*, T. Brody †, D. Bourilkov *, Y.Fu *, B. Kim *, C. Prescott.
03/03/09USCMS T2 Workshop1 Future of storage: Lustre Dimitri Bourilkov, Yu Fu, Bockjoo Kim, Craig Prescott, Jorge L. Rodiguez, Yujun Wu.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Grid Lab About the need of 3 Tier storage 5/22/121CHEP 2012, The need of 3 Tier storage Dmitri Ozerov Patrick Fuhrmann CHEP 2012, NYC, May 22, 2012 Grid.
OSG Storage Architectures Tuesday Afternoon Brian Bockelman, OSG Staff University of Nebraska-Lincoln.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
UMD TIER-3 EXPERIENCES Malina Kirn October 23, 2008 UMD T3 experiences 1.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Purdue Tier-2 Site Report US CMS Tier-2 Workshop 2010 Preston Smith Purdue University
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Status SC3 SARA/Nikhef 20 juli Status & results SC3 throughput phase SARA/Nikhef Mark van de Sanden.
Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
US-CMS T2 Centers US-CMS Tier 2 Report Patricia McBride Fermilab GDB Meeting August 31, 2007 Triumf - Vancouver.
US ATLAS Western Tier 2 Status Report Wei Yang Nov. 30, 2007 US ATLAS Tier 2 and Tier 3 workshop at SLAC.
Tier-2 storage A hardware view. HEP Storage dCache –needs feed and care although setup is now easier. DPM –easier to deploy xrootd (as system) is also.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Florida Tier2 Site Report USCMS Tier2 Workshop Livingston, LA March 3, 2009 Presented by Yu Fu for the University of Florida Tier2 Team (Paul Avery, Bourilkov.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
Final Implementation of a High Performance Computing Cluster at Florida Tech P. FORD, X. FAVE, K. GNANVO, R. HOCH, M. HOHLMANN, D. MITRA Physics and Space.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
A. Mohapatra, T. Sarangi, HEPiX-Lincoln, NE1 University of Wisconsin-Madison CMS Tier-2 Site Report D. Bradley, S. Dasu, A. Mohapatra, T. Sarangi, C. Vuosalo.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
StoRM + Lustre Proposal YAN Tian On behalf of Distributed Computing Group
Martina Franca (TA), 07 November Installazione, configurazione, testing e troubleshooting di Storage Element.
Unit 2 VIRTUALISATION. Unit 2 - Syllabus Basics of Virtualization Types of Virtualization Implementation Levels of Virtualization Virtualization Structures.
New directions in storage | ISGC 2015, Taipei | Patrick Fuhrmann | 19 March 2015 | 1 Presenter: Patrick Fuhrmann dCache.org Patrick Fuhrmann, Paul Millar,
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
Storage at SMU OSG Storage 9/22/2010 Justin Ross Southern Methodist University.
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
Atlas Tier 3 Overview Doug Benjamin Duke University.
STORAGE EXPERIENCES AT MWT2 (US ATLAS MIDWEST TIER2 CENTER) Aaron van Meerten University of Chicago Sarah Williams Indiana University OSG Storage Forum,
GPCF* Update Present status as a series of questions / answers related to decisions made / yet to be made * General Physics Computing Facility (GPCF) is.
BeStMan/DFS support in VDT OSG Site Administrators workshop Indianapolis August Tanya Levshina Fermilab.
Status of BESIII Distributed Computing
dCache “Intro” a layperson perspective Frank Würthwein UCSD
Experiences with http/WebDAV protocols for data access in high throughput computing
CMS analysis job and data transfer test results
Ákos Frohner EGEE'08 September 2008
The INFN Tier-1 Storage Implementation
Composition and Operation of a Tier-3 cluster on the Open Science Grid
Presentation transcript:

9/22/10 OSG Storage Forum 1 CMS Florida T2 Storage Status Bockjoo Kim for the CMS Florida T2

Storage Hardware Status User space data RAID: gatoraid1, gatoraid2, storing CMS software, $DATA, $APP, etc. 3ware controller based RAID5, mounted via NFS. 7.3TB Resilient dCache: 2 x 250 (500) GB SATA drives on 126 worker nodes. 63TB Non-resilient RAID dCache: FibreChannel RAID5 (pool03, pool04, pool05, pool06) + 3ware-based SATA RAID5 (pool02), with 10GbE or bonded multiple 1GbE network. 163TB Lustre storage: Our main storage system now. Serves Bestman. Areca- controller-based RAID5 (dedicated) + RAID Inc Falcon III FibreChannel RAID5 (shared), with InfiniBand+3ware SATA RAID5 (pool01) 215TB+20TB+6.3TB dCache GridFTP servers: Running on 20 worker nodes: 20x1Gbps Dedicated GridFTP servers: Serving the Lustre storage. 3 dedicated GridFTP servers. 3x10Gbps, dual quad-core processors, 32 GB memory. Server Nodes: 2*SRM servers (for dCache and Bestman), 2*PNFS servers (one backup), dCacheAdmin server, dCap server, dCacheAdminDoor server. 2 OSG Storage Forum

Storage Software Status A.dCache : SRM node: 64-bit Centos 5.5 2GB memory. Serves srm and gPlazma Admin node: 32-bit Centos 4.2 2GB memory. Serves lm, dCache(PoolManager), dir, httpd, utility, infoProvider, statistics, and replica. Chimera node: 64-bit Centos GB memory. Serves Chimera Dcap node: 32-bit Centos 4.2 2GB memory. Serves dcap and gsidcap Admin door node: 32-bit Centos 4.2 2GB memory. Serves admin interface Pool/gridftp nodes: 64-bit Centos 5 Resilient and RAID pools. Serves pool B.Bestman Lustre: SRM node: 64-bit Centos 5.5 4GB memory. Serves Bestman srm + plugin for the gridftp selection mechanism. GridFTP node: 64-bit Centos GB memory. Serves gridftp (VDT) MDS node: 64-bit Centos GB memory. Serves MDS and MDT, infiniband OSS node: 64-bit Centos GB memory. Serves OSS and OST, infiniband. 3 OSG Storage Forum

Storage Hardwares 4 OSG Storage Forum RAID dCache Resilient dCache Areca+WD Lustre Adaptec+Hitachi Lustre

5 OSG Storage Forum Motivation for Lustre Motivation – Best in comparison with various available parallel filesystems according to UF HPC’s tests. – Widely used, proven performance, reliability and scalability. – Relatively long history. Mature and stable. – Existing local experience and expertise at UF Tier2 and UF HPC. – Excellent support from the Lustre community, SUN/Oracle and UF HPC experts, frequent updates, prompt patch for new kernels. Hardware selection: – With proper carefully chosen hardware, RAID’s can be inexpensive yet with excellent performance and reasonably high reliability. – Areca ARC-1680ix-16 PCIe SAS RAID controller with 4GB memory, each controller supports 16 drives. – 2TB enterprise-class SATA drives, most cost effective at the time. – iStarUSA drive chassis connected to I/O server through extension PCIe cable: one I/O server can drive multiple chassis, more cost effective. – Cost: <$300/TB raw space, ~$400/TB net usable space with configuration of 4+1 RAID5’s and a global hot spare drive for every 3 RAID5’s.

6 OSG Storage Forum Lustre at Florida T2

7 OSG Storage Forum Florida T2 OSS/OST We have a total of 5 OSSes and 6 OSTs per OSS

8 OSG Storage Forum Florida SE Capacity Summary ResourceRawUsable (after redundancy, overhead and/or resilient) dCache non-resilient pools (RAID’s)173.2TB137.1TB dCache resilient pools (worker nodes)93.0TB35TB Dedicated Tier2 production Lustre320TB215TB Test Lustre8TB6.3TB Shared HPC Lustre (Tier2’s share)N/A30TB Total594.2TB423.4TB

9 OSG Storage Forum Experience with Lustre Remote Access Test with nearby institutions, FIU and FIT. Used Lustre as dCache pools (need localflock to make it work) It is not good for small large number of files (bad for $APP) It is not without problems: LBUG, expensive, metadata recovery time maybe long, etc

10 OSG Storage Forum Performance: Data Moving Florida -> Caltech 480MB/s averageCaltech -> Florida 350MB/s average Four Day Data Transfer Test

11 OSG Storage Forum Performance: dCache vs Lustre LustredCache File Read Performance

12 OSG Storage Forum Performance: Lustre I/O Lustre I/O seen from Network Monitoring Out In 1.4GB/s max

13 OSG Storage Forum Plans For Upgrade Total raw ~600 TB, net usable 425 TB. Trying to meet the 2010 USCMS T2 milestone: 570 TB Planning to deploy 256 TB (raw) new RAID’s (similar to current OSS/OST configuration) in the production Lustre system in FY10. Budget planned: ~$80k. We tested a system based on Hitachi + Adaptec and Positive result. Upgrade will be based on this system. Satisfied with Lustre+Bestman+GridFTP and will stay with it. Will put future SE’s into Lustre. Present non-resilient dCache pools (RAID’s) will gradually fade out and migrate to Lustre. Resilient dCache pools on worker nodes will be kept.

14 OSG Storage Forum Researches Supported Within CMS, we support 4 groups and the local CMS physicists – Muon Detector Group Arranged write access to the dCache and the fair share. Local users also requested large portion of muon group datasets. – SUSY Physicis Group In 2009, we provided the space for the private SUSY production. Large number of datasets that are produced by the local SUSY group are still in the analysis 2 DBS. – Particle Flow / JetMet Group We were not able to help much with this group because we did not have enough space at the time because most space was occupied by the Muon group. Not much interaction, only with the validation datasets were hosted for this group. Recently one of our postdocs is interacting with us to coordinate the necessary dataset hosting. – Higgs Thanks to addition of the Lustre storage. OSG Support – We only support SCEC via a development lustre system at the moment due to the space limitation and manpower.

15 OSG Storage Forum XROOTD Servers We have recently deployed two xrootd servers They are customized to CMS by Brian One for dCache and the other for Lustre dCache : 50MB/s locally (dccp rate is 75MB/s)

16 OSG Storage Forum Extenci ExTENCI: Extending Science Through Enhanced National CyberInfrastructure; – Funded by NSF/DOE for two years: $2M total – Work together between Open Science Grid and TeraGrid – The project just started Lustre will be one of the major projects for ExTENCI Project goal is to evaluate a Global wide-area file system for performance and robustness Florida Focus will be mainly on the secure Lustre WAN access and the project result will potentially allow secure access to remote (Tier-2 and other) resources in US-CMS with less administrative overhead

17 OSG Storage Forum Summary 423TB of dCache+Lustre in production. Planning to increase the capacity to 570TB soon based on the Adaptec+Hitachi Established a production Lustre+BestMan+GridFTP storage system. A review document within USCMS being prepared. Satisfied with Lustre, will put most of current dCache non-resilient pools into the Lustre system. Supports 4 associated analysis groups within CMS, an OSG group, and the local CMS users. XROOTD deployed Extenci participation

Additional Slides 18OSG Storage Forum

19 OSG Storage Forum Motivation of RAID dCache – Our past experience shows resilient pools on worker nodes relatively often went down or lost data due to CPU load and memory usage as well as hard drive etc hardware issues. – Therefore we want to separate CE and SE in hardware so job load on CE worker nodes will not affect storage. – High-performance-oriented RAID hardware is supposed to perform better. – Dedicated, specially designed RAID hardware is expected to be more robust and reliable. – Hardware RAID will not need full replicated copies of data like in resilient dCache and Hadoop, more efficient in disk usage. – Fewer equipments are involved in dedicated large RAID’s, easier to manage and maintain.

20 OSG Storage Forum Diagram of a Lustre SE