Tom Byrne, 12 th November 2014 Ceph – status update and xrootd testing Alastair Dewhurst, Tom Byrne 1.

Slides:



Advertisements
Similar presentations
DPM Basics and its status and plans Wahid Bhimji University of Edinburgh GridPP Storage Workshop – Apr 2010 Apr-101Wahid Bhimji – DPM.
Advertisements

ATLAS Tier-3 in Geneva Szymon Gadomski, Uni GE at CSCS, November 2009 S. Gadomski, ”ATLAS T3 in Geneva", CSCS meeting, Nov 091 the Geneva ATLAS Tier-3.
Proof/Proof-on-Demand at Atlas T3 Shuwei Ye (BNL) Aug 20, 2014.
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
Ceph scalable, unified storage files, blocks & objects Tommi Virtanen / DreamHostOpenStack Conference
DPM Italian sites and EPEL testbed in Italy Alessandro De Salvo (INFN, Roma1), Alessandra Doria (INFN, Napoli), Elisabetta Vilucchi (INFN, Laboratori Nazionali.
Staging to CAF + User groups + fairshare Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE Offline week,
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
Hadoop Distributed File System by Swathi Vangala.
Wahid Bhimji University of Edinburgh P. Clark, M. Doidge, M. P. Hellmich, S. Skipsey and I. Vukotic 1.
Brainstorming Thin-provisioned Tier 3s.
Filesytems and file access Wahid Bhimji University of Edinburgh, Sam Skipsey, Chris Walker …. Apr-101Wahid Bhimji – Files access.
CSC 456 Operating Systems Seminar Presentation (11/13/2012) Leon Weingard, Liang Xin The Google File System.
MapReduce: Simplified Data Processing on Large Clusters Jeffrey Dean and Sanjay Ghemawat.
StoRM Some basics and a comparison with DPM Wahid Bhimji University of Edinburgh GridPP Storage Workshop 31-Mar-101Wahid Bhimji – StoRM.
Your university or experiment logo here NextGen Storage Shaun de Witt (STFC) With Contributions from: James Adams, Rob Appleyard, Ian Collier, Brian Davies,
Xrootd, XrootdFS and BeStMan Wei Yang US ATALS Tier 3 meeting, ANL 1.
Storage Wahid Bhimji DPM Collaboration : Tasks. Xrootd: Status; Using for Tier2 reading from “Tier3”; Server data mining.
Wenjing Wu Andrej Filipčič David Cameron Eric Lancon Claire Adam Bourdarios & others.
CERN IT Department CH-1211 Geneva 23 Switzerland t Storageware Flavia Donno CERN WLCG Collaboration Workshop CERN, November 2008.
Experience with the Thumper Wei Yang Stanford Linear Accelerator Center May 27-28, 2008 US ATLAS Tier 2/3 workshop University of Michigan, Ann Arbor.
Δ Storage Middleware GridPP10 What’s new since GridPP9? CERN, June 2004.
PROOF Cluster Management in ALICE Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE CAF / PROOF Workshop,
Support in setting up a non-grid Atlas Tier 3 Doug Benjamin Duke University.
Wahid, Sam, Alastair. Now installed on production storage Edinburgh: srm.glite.ecdf.ed.ac.uk  Local and global redir work (port open) e.g. root://srm.glite.ecdf.ed.ac.uk//atlas/dq2/mc12_8TeV/NTUP_SMWZ/e1242_a159_a165_r3549_p1067/mc1.
Introduction to HDFS Prasanth Kothuri, CERN 2 What’s HDFS HDFS is a distributed file system that is fault tolerant, scalable and extremely easy to expand.
Grid Computing at Yahoo! Sameer Paranjpye Mahadev Konar Yahoo!
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
RAL Site Report Castor Face-to-Face meeting September 2014 Rob Appleyard, Shaun de Witt, Juan Sierra.
Status & Plan of the Xrootd Federation Wei Yang 13/19/12 US ATLAS Computing Facility Meeting at 2012 OSG AHM, University of Nebraska, Lincoln.
INFSO-RI Enabling Grids for E-sciencE gLite Data Management and Interoperability Peter Kunszt (JRA1 DM Cluster) 2 nd EGEE Conference,
Introduction to HDFS Prasanth Kothuri, CERN 2 What’s HDFS HDFS is a distributed file system that is fault tolerant, scalable and extremely easy to expand.
Storage Federations and FAX (the ATLAS Federation) Wahid Bhimji University of Edinburgh.
C. Aiftimiei, E. Ferro / January LCFGng server installation Cristina Aiftimiei, Enrico Ferro INFN-LNL.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
ATLAS XRootd Demonstrator Doug Benjamin Duke University On behalf of ATLAS.
DPM Python tools Ivan Calvet IT/SDC-ID DPM Workshop 10 th October 2014.
CERN IT Department CH-1211 Genève 23 Switzerland t Streams Service Review Distributed Database Workshop CERN, 27 th November 2009 Eva Dafonte.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS XROOTD news New release New features.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
PROOF tests at BNL Sergey Panitkin, Robert Petkus, Ofer Rind BNL May 28, 2008 Ann Arbor, MI.
Awesome distributed storage system
Evaluating the performance of Seagate Kinetic Drives Technology and its integration with the CERN EOS storage system Ivana Pejeva openlab Summer Student.
Lesson 6-Using Utilities to Accomplish Complex Tasks.
SRM v2.2 Production Deployment SRM v2.2 production deployment at CERN now underway. – One ‘endpoint’ per LHC experiment, plus a public one (as for CASTOR2).
Testing Infrastructure Wahid Bhimji Sam Skipsey Intro: what to test Existing testing frameworks A proposal.
Data Analysis w ith PROOF, PQ2, Condor Data Analysis w ith PROOF, PQ2, Condor Neng Xu, Wen Guan, Sau Lan Wu University of Wisconsin-Madison 30-October-09.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
Andrea Manzi CERN EGI Conference on Challenges and Solutions for Big Data Processing on cloud 24/09/2014 Storage Management Overview 1 24/09/2014.
First Experiences with Ceph on the WLCG Grid Rob Appleyard Shaun de Witt, James Adams, Brian Davies.
An Analysis of Data Access Methods within WLCG Shaun de Witt, Andrew Lahiff (STFC)
New Features of Xrootd SE Wei Yang US ATLAS Tier 2/Tier 3 meeting, University of Texas, Arlington,
Security recommendations DPM Jean-Philippe Baud CERN/IT.
DPM in FAX (ATLAS Federation) Wahid Bhimji University of Edinburgh As well as others in the UK, IT and Elsewhere.
CERN IT Department CH-1211 Genève 23 Switzerland t DPM status and plans David Smith CERN, IT-DM-SGT Pre-GDB, Grid Storage Services 11 November.
CERN IT-Storage Strategy Outlook Alberto Pace, Luca Mascetti, Julien Leduc
Federating Data in the ALICE Experiment
a brief summary for users
Tom Byrne, Bruno Canning
Global Data Access – View from the Tier 2
Diskpool and cloud storage benchmarks used in IT-DSS
GFAL 2.0 Devresse Adrien CERN lcgutil team
RAL Tom Byrne George Vasilakakos, Bruno Canning, Alastair Dewhurst, Ian Johnson, Alison Packer.
CTA: CERN Tape Archive Overview and architecture
Data Management cluster summary
bitcurator-access-webtools Quick Start Guide
INFNGRID Workshop – Bari, Italy, October 2004
Ceph at the Tier-1 Tom Byrne.
Presentation transcript:

Tom Byrne, 12 th November 2014 Ceph – status update and xrootd testing Alastair Dewhurst, Tom Byrne 1

Tom Byrne, 12 th November 2014 Introduction On 15 th October gave overview talk on plans for Ceph at RAL Tier 1. Will aim to provide updates on progress made focusing on the xrootd deployment and testing. Current Ceph cluster with 7 nodes using 2013 generation hardware. 2

Tom Byrne, 12 th November 2014 S3 gateway At last meeting we had S3 gateway on virtual machine: Hope to have firewall holes + x.509 authentication working by next week. S3 gateway ‘does it’s own thing’ with files which means it is difficult to use with other plugins. Will investigate writing own WebDAV gateway. 3

Tom Byrne, 12 th November 2014 CERN plugins CERN have four plugins based on XRootD for CEPH: radosfs (impl. file & directories in rados) xrootd-rados-oss (interfacing radosfs as OSS plug-in) xrootd-diamond-ofs (adding checksumming & TPC) xrootd-auth-change-id (adding NFS server style authentication to xrootd) Our work has been on the xrootd-diamond-ofs Setup instructions can be found: 4

Tom Byrne, 12 th November 2014 Xrootd deployment Used RPMs provided on wiki to setup XrootD gateway Had to setup a Cache tier because it currently doesn’t work directly with erasure coded pools This is because the file is opened and then appended to, CERN are working on patching it to work with EC. There are two pools: Data and Meta-Data 5

Tom Byrne, 12 th November 2014 Cache Tier Cache Tier is using mostly default settings 3 replicas of the data Will create a ‘cold’ erasure coded copy instantly LRU algorithm to clean up data. We would prefer not to use a Cache Tier and have direct access to Erasure coded pool It would be possible to have a ~10% Cache Tier in front of the storage. We believe Erasure coded pool should work well as we are not appending to files. 6

Tom Byrne, 12 th November 2014 Diamond data Plugin splits file into chunks which are stored with a GUID in Ceph: Makes it hard to manage files and write other plugins. 7 ~]# rados -p diamond-data ls | grep 774b1a83-14d0-4fb9-a6c0-10e36c32febf | sort 774b1a83-14d0-4fb9-a6c0-10e36c32febf 774b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// a 774b1a83-14d0-4fb9-a6c0-10e36c32febf// b 774b1a83-14d0-4fb9-a6c0-10e36c32febf// c 774b1a83-14d0-4fb9-a6c0-10e36c32febf// d 774b1a83-14d0-4fb9-a6c0-10e36c32febf// e 774b1a83-14d0-4fb9-a6c0-10e36c32febf// f ~]# rados -p diamond-data ls | grep 774b1a83-14d0-4fb9-a6c0-10e36c32febf | sort 774b1a83-14d0-4fb9-a6c0-10e36c32febf 774b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// b1a83-14d0-4fb9-a6c0-10e36c32febf// a 774b1a83-14d0-4fb9-a6c0-10e36c32febf// b 774b1a83-14d0-4fb9-a6c0-10e36c32febf// c 774b1a83-14d0-4fb9-a6c0-10e36c32febf// d 774b1a83-14d0-4fb9-a6c0-10e36c32febf// e 774b1a83-14d0-4fb9-a6c0-10e36c32febf// f

Tom Byrne, 12 th November 2014 Diamond meta-data 8

Tom Byrne, 12 th November 2014 Testing Have tried commands from: UI (using xrootd v3.3.6) Node (using xrootd v4.0.4) Can copy files in and out: 9 ~]# xrdcp./ivukotic\:group.test.hc.NTUP_SMWZ.root root://gdss541//root/ivukotic:group.test.hc.NTUP_SMWZ.root.1root://gdss541//root/ivukotic:group.test.hc.NTUP_SMWZ.root.1 [760.2MB/760.2MB][100%][==================================================][95.03MB/s] ~]# xrdcp./ivukotic\:group.test.hc.NTUP_SMWZ.root root://gdss541//root/ivukotic:group.test.hc.NTUP_SMWZ.root.1root://gdss541//root/ivukotic:group.test.hc.NTUP_SMWZ.root.1 [760.2MB/760.2MB][100%][==================================================][95.03MB/s] ~]# xrdcp root://gdss541//root/ivukotic:group.test.hc.NTUP_SMWZ.root /ivukotic\:group.test.hc.NTUP_SMWZ.root [760.2MB/760.2MB][100%][==================================================][58.48MB/s] ~]# xrdcp root://gdss541//root/ivukotic:group.test.hc.NTUP_SMWZ.root /ivukotic\:group.test.hc.NTUP_SMWZ.root [760.2MB/760.2MB][100%][==================================================][58.48MB/s]

Tom Byrne, 12 th November 2014 “Filesystem” Can create directories with UNIX style permissions. Setup is “Fragile” – frequently need to restart xrootd. Dies when doing “ls –l” 10 xrdfs gdss541 mkdir "/atlas/?owner=10763&group=1307" ~]# xrdfs gdss541 ls /atlas/ /atlas/ivukotic:group.test.hc.NTUP_SMWZ.root /atlas/test ~]# xrdfs gdss541 ls /atlas/ /atlas/ivukotic:group.test.hc.NTUP_SMWZ.root /atlas/test

Tom Byrne, 12 th November 2014 Direct Read Code from Wahid: git clone Wanted to try 4 tests: Read 10% of the file and use 30MB cache Read 100% of the file and use 30MB cache Read 10% of the file and use 100MB cache – CRASHED! Read 100% of the file and use 100MB cache – CRASHED! 11 30MB Cache 1st2nd3rdAverage 100%CPU Time /s Disk IO MB/s %CPU Time /s Disk IO MB/s

Tom Byrne, 12 th November 2014 Future plans 3 threads of development: Get simplified xrootd to work. Look into GridFTP gateway – Spoken to Brian Bockelman who has made equivalent for HDFS. Look into Webdav gateway – Instructions to get started on Ceph wiki and will speak to DPM developers. Need to start looking at xattr We have procured mac mini for future Calamari builds. 12

Tom Byrne, 12 th November 2014 Summary We got S3 gateway to work, but it wasn’t quite what we wanted. Testing Diamond plugin with help from CERN. Do not need all the features. Question: Why do all the plugins create their own data formats? If we go with an object store we will have to write our own plugins but this does not appear to be an impossible task. 13