First Look at the New NFSv4.1 Based dCache Art Kreymer, Stephan Lammel, Margaret Votava, and Michael Wang for the CD-REX Department CD Scientific Computing.

Slides:



Advertisements
Similar presentations
Steve Traylen Particle Physics Department Experiences of DCache at RAL UK HEP Sysman, 11/11/04 Steve Traylen
Advertisements

Xrootd and clouds Doug Benjamin Duke University. Introduction Cloud computing is here to stay – likely more than just Hype (Gartner Research Hype Cycle.
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool.
NWfs A ubiquitous, scalable content management system with grid enabled cross site data replication and active storage. R. Scott Studham.
Frangipani: A Scalable Distributed File System C. A. Thekkath, T. Mann, and E. K. Lee Systems Research Center Digital Equipment Corporation.
Jun 29, 20101/25 Storage Evaluation on FG, FC, and GPCF Jun 29, 2010 Gabriele Garzoglio Computing Division, Fermilab Overview Introduction Lustre Evaluation:
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Windows Server MIS 424 Professor Sandvig. Overview Role of servers Performance Requirements Server Hardware Software Windows Server IIS.
Report : Zhen Ming Wu 2008 IEEE 9th Grid Computing Conference.
Interposed Request Routing for Scalable Network Storage Darrell Anderson, Jeff Chase, and Amin Vahdat Department of Computer Science Duke University.
CSC 456 Operating Systems Seminar Presentation (11/13/2012) Leon Weingard, Liang Xin The Google File System.
File System Benchmarking
Investigation of Storage Systems for use in Grid Applications 1/20 Investigation of Storage Systems for use in Grid Applications ISGC 2012 Feb 27, 2012.
Projects. High Performance Computing Projects Design and implement an HPC cluster with one master node and two compute nodes. (Hint: use Rocks HPC Cluster.
The Center for Autonomic Computing is supported by the National Science Foundation under Grant No NSF CAC Seminannual Meeting, October 5 & 6,
Ceph Storage in OpenStack Part 2 openstack-ch,
Mar 24, 20111/17 Investigation of storage options for scientific computing on Grid and Cloud facilities Mar 24, 2011 Keith Chadwick for Gabriele Garzoglio.
StoRM Some basics and a comparison with DPM Wahid Bhimji University of Edinburgh GridPP Storage Workshop 31-Mar-101Wahid Bhimji – StoRM.
Lustre File System Evaluation at FNAL CHEP'09, Prague March 23, 2009 Stephen Wolbers for Alex Kulyavtsev, Matt Crawford, Stu Fuess, Don Holmgren, Dmitry.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
Sensitivity of Cluster File System Access to I/O Server Selection A. Apon, P. Wolinski, and G. Amerson University of Arkansas.
Amy Apon, Pawel Wolinski, Dennis Reed Greg Amerson, Prathima Gorjala University of Arkansas Commercial Applications of High Performance Computing Massive.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Investigation of Storage Systems for use in Grid Applications 1/20 Investigation of Storage Systems for use in Grid Applications ISGC 2012 Feb 27, 2012.
Large Scale Parallel File System and Cluster Management ICT, CAS.
CHEP04 Performance Analysis of Cluster File System on Linux Yaodong CHENG IHEP, CAS
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
1EMC CONFIDENTIAL—INTERNAL USE ONLY FAST VP and Exchange Server 2010 Don Turner Consultant Systems Integration Engineer Microsoft TPM.
OSG Abhishek Rana Frank Würthwein UCSD.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
Padova, 5 October StoRM Service view Riccardo Zappi INFN-CNAF Bologna.
BlueArc IOZone Root Benchmark How well do VM clients perform vs. Bare Metal clients? Bare Metal Reads are (~10%) faster than VM Reads. Bare Metal Writes.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
KIT – University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association STEINBUCH CENTRE FOR COMPUTING - SCC
Benchmarking Storage Systems How to characterize the system Storage Network Clients Specific benchmarks iozone mdtest h5perf Hdf5-aggregation (tiff2nexus)
Testing CernVM-FS scalability at RAL Tier1 Ian Collier RAL Tier1 Fabric Team WLCG GDB - September
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
CRISP WP18, High-speed data recording Krzysztof Wrona, European XFEL PSI, 18 March 2013.
New directions in storage | ISGC 2015, Taipei | Patrick Fuhrmann | 19 March 2015 | 1 Presenter: Patrick Fuhrmann dCache.org Patrick Fuhrmann, Paul Millar,
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
An Introduction to GPFS
CD FY10 Budget and Tactical Plan Review FY10 Tactical Plans for Scientific Computing Facilities / General Physics Computing Facility (GPCF) Stu Fuess 06-Oct-2009.
GPCF* Update Present status as a series of questions / answers related to decisions made / yet to be made * General Physics Computing Facility (GPCF) is.
The Distributed Application Debugger (DAD)
Experience of Lustre at QMUL
Scalable sync-and-share service with dCache
WP18, High-speed data recording Krzysztof Wrona, European XFEL
StoRM: a SRM solution for disk based storage systems
Diskpool and cloud storage benchmarks used in IT-DSS
dCache “Intro” a layperson perspective Frank Würthwein UCSD
Experiences with http/WebDAV protocols for data access in high throughput computing
Experience of Lustre at a Tier-2 site
CMS analysis job and data transfer test results
Legacy NFS provides sharing, but does not scale
Large Scale Test of a storage solution based on an Industry Standard
Mapping the Data Warehouse to a Multiprocessor Architecture
Objective Understand the concepts of modern operating systems by investigating the most popular operating system in the current and future market Provide.
Overview Context Test Bed Lustre Evaluation Standard benchmarks
Lecture 15 Reading: Bacon 7.6, 7.7
Objective Understand the concepts of modern operating systems by investigating the most popular operating system in the current and future market Provide.
Presentation transcript:

First Look at the New NFSv4.1 Based dCache Art Kreymer, Stephan Lammel, Margaret Votava, and Michael Wang for the CD-REX Department CD Scientific Computing Facilities Leaders Meeting 13 December 2011

First look at dCache with NFSv Michael Wang Introduction –Investigate alternatives to the BlueArc-based IF central disk servers BA has performed fairly well for the, currently, relatively modest requirements of the IF experiments (~0.5 PB) Will it continue to satisfy the growing requirements of the IF experiments in the years to come in a reasonable, cost-effective way? –Started surveying storage options available: NFSv4.1: all the rage among the major storage vendors (Panasas, IBM, EMC, NetApp, even BlueArc) Despite all the hype, no stable server implementation readily available for evaluation –Stumbled upon presentation on the web by DESY dCache team: Described a stable NFSv4.1 implementation in a new Chimera-based version of dCache All the nice features of the old dCache PLUS all files in exported filesystem tree now directly accessible (POSIX compliant) without special protocols (like DCAP) ! i.e. dCache filesystem can now appear & behave like a regular nfs accessible area on a worker node

First look at dCache with NFSv Michael Wang Introduction –Approached our local dCache experts: REX and DMS meeting where DMS gave overview of the new dCache. DMS department set up a test dCache system (version ) for us to evaluate (many thanks to Dmitry Litvintsev, Yujun Wu, Terry Jones, Stan Naymola and Gene Oleynik from DMS for their support). –Brief overview of talk: Description of test setup Present some initial test results Focus is on technical I/O performance:  no discussion on other nice features of NFSv4.1 (e.g. ACLs)  no cost comparisons and studies (relative to BA)

First look at dCache with NFSv Michael Wang Test setup –Client side: SLF6 Virtual Machines on Fermicloud (many thanks to Steve Timm and Farooq Lowe of Fermigrid Dept.) Linux kernel (a renamed 3.0 kernel) –Server side: dCache with one head node, two pool nodes Each pool node has 2 RAID6 partitions with 4x250GB SATA drives each

First look at dCache with NFSv Michael Wang Throughput test results IOzone in cluster mode with sequential write and read test. Increased number of clients beyond 10 (multiple clients per VM) but aggregate data transferred Is fixed to 40GB. One 4GB file transferred per client. One client per VM.

First look at dCache with NFSv Michael Wang Monitoring pool node disk activity during IOzone test Pool node 1 Partition A Pool node 1 Partition B Pool node 2 Partition A Pool node 2 Partition B Strip chart recording: (x-axis, y-axis) = (time, MB/sec) Disk write rateDisk read rate

First look at dCache with NFSv Michael Wang Metadata test results Mdtest with multiple MPI tasks, each creating/”stat”-ing/removing 100 directories zero-length files.

First look at dCache with NFSv Michael Wang Conclusion –Presented some preliminary test results on the new NFSv4.1 Chimera-based version of dCache –Results look promising, throughput scales well with number of pool nodes –Metadata performance may be adequate for now but may be a cause for concern in the future (need to consult and discuss with the developers) –Will do more “real-world” tests, e.g. with Art Kreymer’s BlueArc performance monitoring scripts –More details can be found in a write-up in CD DocDB (CS-doc- 4583): –Details on setting up VM clients with pNFS-enabled Linux kernels available on Fermi Redmine IF-storage project Wiki: –Many thanks to DMS and Fermigrid Depts. for their unwavering support!

First look at dCache with NFSv Michael Wang End

First look at dCache with NFSv Michael Wang Monitoring pool node disk activity Pool node 1 Partition A Pool node 1 Partition B Pool node 2 Partition A Pool node 2 Partition B