Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/i t DBES Solutions for WAN data access: xrootd and NFSv4.1 Andrea Sciabà.

Slides:



Advertisements
Similar presentations
Distributed Xrootd Derek Weitzel & Brian Bockelman.
Advertisements

Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
CERN - IT Department CH-1211 Genève 23 Switzerland t SVN Pilot: CVS Replacement Manuel Guijarro Jonatan Hugo Hugosson Artur Wiecek David.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES News on monitoring for CMS distributed computing operations Andrea.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
ALICE data access WLCG data WG revival 4 October 2013.
CERN IT Department CH-1211 Geneva 23 Switzerland t The Experiment Dashboard ISGC th April 2008 Pablo Saiz, Julia Andreeva, Benjamin.
CERN IT Department CH-1211 Genève 23 Switzerland t Plans and Architectural Options for Physics Data Analysis at CERN D. Duellmann, A. Pace.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES PhEDEx Monitoring Nicolò Magini CERN IT-ES-VOS For the PhEDEx.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
Xrootd, XrootdFS and BeStMan Wei Yang US ATALS Tier 3 meeting, ANL 1.
CERN IT Department CH-1211 Genève 23 Switzerland t Tier0 Status - 1 Tier0 Status Tony Cass LCG-LHCC Referees Meeting 18 th November 2008.
Evolution, by tackling new challenges| CHEP 2015, Japan | Patrick Fuhrmann | 16 April 2015 | 1 Patrick Fuhrmann On behave of the project team Evolution,
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Grid Lab About the need of 3 Tier storage 5/22/121CHEP 2012, The need of 3 Tier storage Dmitri Ozerov Patrick Fuhrmann CHEP 2012, NYC, May 22, 2012 Grid.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Redirector xrootd proxy mgr Redirector xrootd proxy mgr Xrd proxy data server N2N Xrd proxy data server N2N Global Redirector Client Backend Xrootd storage.
Status & Plan of the Xrootd Federation Wei Yang 13/19/12 US ATLAS Computing Facility Meeting at 2012 OSG AHM, University of Nebraska, Lincoln.
CERN IT Department CH-1211 Geneva 23 Switzerland GT WG on Storage Federations First introduction Fabrizio Furano
ROOT and Federated Data Stores What Features We Would Like Fons Rademakers CERN CC-IN2P3, Nov, 2011, Lyon, France.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
ATLAS XRootd Demonstrator Doug Benjamin Duke University On behalf of ATLAS.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS XROOTD news New release New features.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Andrea Sciabà Hammercloud and Nagios Dan Van Der Ster Nicolò Magini.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
CERN IT Department CH-1211 Genève 23 Switzerland t CERN IT Monitoring and Data Analytics Pedro Andrade (IT-GT) Openlab Workshop on Data Analytics.
Federated Data Stores Volume, Velocity & Variety Future of Big Data Management Workshop Imperial College London June 27-28, 2013 Andrew Hanushevsky, SLAC.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Overview of DMLite Ricardo Rocha ( on behalf of the LCGDM team.
Efi.uchicago.edu ci.uchicago.edu Ramping up FAX and WAN direct access Rob Gardner on behalf of the atlas-adc-federated-xrootd working group Computation.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
Andrea Manzi CERN EGI Conference on Challenges and Solutions for Big Data Processing on cloud 24/09/2014 Storage Management Overview 1 24/09/2014.
EMI INFSO-RI Patrick Fuhrmann EMI Data area leader At the EGI Technical Forum 2011, in Lyon EMI-Data The second year.
An Analysis of Data Access Methods within WLCG Shaun de Witt, Andrew Lahiff (STFC)
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Evolution of WLCG Data & Storage Management Outcome of Amsterdam.
New Features of Xrootd SE Wei Yang US ATLAS Tier 2/Tier 3 meeting, University of Texas, Arlington,
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Protocols in DPM Ricardo Rocha.
HEPiX IPv6 Working Group David Kelsey (STFC-RAL) GridPP33 Ambleside 22 Aug 2014.
Federating Data in the ALICE Experiment
a brief summary for users
Jean-Philippe Baud, IT-GD, CERN November 2007
WLCG IPv6 deployment strategy
DPM at ATLAS sites and testbeds in Italy
Global Data Access – View from the Tier 2
Status of the SRM 2.2 MoU extension
Diskpool and cloud storage benchmarks used in IT-DSS
StoRM Architecture and Daemons
Introduction to Data Management in EGI
Data Federation with Xrootd Wei Yang US ATLAS Computing Facility meeting Southern Methodist University, Oct 11-12, 2011.
Brookhaven National Laboratory Storage service Group Hironori Ito
Ákos Frohner EGEE'08 September 2008
Large Scale Test of a storage solution based on an Industry Standard
Support for ”interactive batch”
Summary of the dCache workshop
Presentation transcript:

Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Solutions for WAN data access: xrootd and NFSv4.1 Andrea Sciabà Workshp CCR INFN GRID 2011 La Biodola, May 16-20

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 2 Outline Xrootd as a WAN data access protocol –CMS –ATLAS The NFS 4.1 protocol and WLCG –Status and plans

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 3 What is xrootd? A file access and data transfer protocol –Typically (but not necessarily) used for ROOT files An actual storage solution for scalable and low latency data access Supports clustering of xrootd servers –Also in the WAN Consists of xrootd-cmsd server pairs –xrootd: handles data and redirections –cmsd: clusters the xrootd daemons together

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 4 CMS global xrootd: purpose and goals Purpose –Provide a service which allows jobs to seamlessly access data located on remote sites with an acceptable loss of efficiency Goals –Reliability: seamless failover –Transparency: global name space –Usability: native to ROOT –Global: applicable to all CMS data –Automatic caching not a primary goal Adopted technology is xrootd –Global redirector mechanism –xrootd data access protocol

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 5 A global xrootd federation

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 6 xrootd architecture System overlaid on existing SEs –Xrootd, dCache, HDFS, POSIX-based –Needs an xrootd proxy (or door) at each site

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 7 Use cases Fallback: seamless failover even between sites Interactive use: debug single files Overflow: allow jobs to go to sites without the data (Caching): xrootd- enabled T3 can locally cache remote data

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 8 Current status A prototype was already successfully demonstrated –One global redirector –CMSSW improved for high latency data access First production deployment in the US –1 Tier-1, ~6 Tier-2; few users so far –Adds a “regional” redirector layer –Adds a monitoring system based on Nagios, MonALISA, alerts –Fallback use case done on May 1 st –Overflow use case due on June 1 st –Interactive use due on August 1 st

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 9 Xrootd monitoring Using Nagios –Status of servers and redirectors –Check availability of random files Using MonALISA –File-open and read rates –Transfer rates –CPU, memory, disk, network etc. for xrd servers (optional) Future improvements –Per user / per file statistics to detect system misuse

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 10 Issues Redirection to broken servers possible Truncated files visible to users Can exhaust network bandwidth at large scales (if sites do not throttle bandwidth of network services) Important monitoring information still missing –Opened files and rates per user (now only per host) –Namespace query rates Caching is an open issue

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 11 ATLAS xrootd Goals similar to the CMS system, but more focus on caching and replication –Access data at Tier-3 sites via xrootd to reduce load on Tier-1/2 sites –Tier-3 sites can fetch data from Tier-1, Tier-2 or other Tier-3 –Automatic replication of popular datasets –Easy to fetch ROOT files from remote sites to a desktop at CERN

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 12 Tasks specific to ATLAS Define and implement a global name space –Global physical dataset path (PDP) Couple xrootd to LFC Enable local file caching at Tier-3 sites –Use the xrootd transfer service (FRM) plugin to copy files between different xrootd servers or even from non-xrootd SEs Support ATLAS data security policies

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 13 Architecture

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 14 Caching via xrootd

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 15 Status Demonstrator finished, now in preproduction phase Deployed in the US at 4 Tier-2, 2 Tier-3, 1 test site ATLAS DDM client tools adapted to get files via their PDP Testing Tier-3 xrootd transfer service using –Native xrdcp copy –Automatic file transfers between Tier-3 sites –Trigger from user root analysis jobs

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 16 Plans Move to production phase Set up a global redirector at BNL Add more sites in US and in Europe Test Tier-3 site caching Add another global redirector in Europe Use X.509 for authentication/authorization

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 17 NFS 4.1 and pNFS: introduction Why is pNFS interesting? What is the status and timeline for pNFS deployment ? Availability of components Protocol verification and performance evaluation Issues to be considered Security Wide area transfers Some results from the dCache NFS 4.1/pNFS evaluation at Grid-Lab Status of DPM NFS 4.1

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 18 Why is pNFS interesting Benefits of parallel I/O Delivers very high application performance Allows for massive scalability without diminished performance Benefits of NFS (or mostly any standard) Interoperability among vendor solutions Eliminates risks of deploying proprietary technology

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 19 Additional benefits Simplicity Regular mount point and real POSIX I/O Can be used by unmodified applications Data client provided by the OS vendor Smart caching (block caching) done by OS vendors

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 20 Availability for production use

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 21 pNFS client support in SL5/6 Full NFS 4.1/pNFS client available in Redhat already back-ported pNFS into but only available for business partners right now Official back port into RH6.x (kernel rh ) expected with RH6.2 (maybe earlier) Based on RH6.x it appear in SL6.x

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 22 pNFS EMI server support dCache Full implementation available starting from Plain and Kerberos authentication DPM Read and write, ready to test Plain and Kerberos authentication StoRM Will become available with GPFS pNFS availability

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 23 Non-EMI implementations GFS2 Available with (now) Uses pNFS for reading and only NFS 4.1 for writing NetApp Available with ONTAP® 8.1 (Nowi) DESY waiting for a test system Others IBM, BlueArc, EMC … are all members of the relevant CITI projects and have server software ready but seem to be reluctant to provide products just yet

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 24 Security issues All tests (up to now) done without strong authentication NFS4.1/pNFS clients are only available with Kerberos authentication How about X509 certificates? Theoretically possible but kernel developers only interested in Kerberos Existing solution: creating a Kerberos token from a X509 certificate

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 25 Wide Area issues On wide area performance: No tests done yet Planned between DESY and CERN after CERN test system will become available

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 26 Possible wide area NFS 4.1 deployment NFSv4 referrals permit a server to redirect a client to another location Central configuration of referral hosts through FedFS

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 27 DESY Grid Lab Since mid of last year, DESY provides a Tier 2-like testbed with dCache/pNFS server and pNFS-enabled SL5 worker nodes Available to verify a client/framework against pNFS (NFS 4.1) DESY folks together with ATLAS, CMS and with help of ROOT have been running all kind of evaluations Results have been presented at CHEP’10 and at 2010 Spring HEPIX

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 28 The DESY pNFS “Tier 2”

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 29 Performance limitations NFSv4.1 dCap Limited by disk bandwidth Limited WN 1Gb network Limited 20 Gb network Removing server disk congestion effect by keeping all data in file system cache of the pool Number of threads Mbytes/sec Total throughput does not depend on the protocol

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 30 xRoot / NFS 4.1 in dCache (I) Non-Optimized Files Read entire file TreeTCache OFF Worst Case for ROOT Xrootd/dCache Xrootd/SLAC NFS/pNFS Optimized for ROOT Read entire file TreeTCache ON Best Case for ROOT Reading an entire ROOT file using a ROOT script If setting is bad for ROOT, SLAC and dCache xroot implementations behave the same For full file read, NFS behaves as well as SLAC/xRoot

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 31 xRoot / NFS 4.1 in dCache (II) Trying to find a case where NFS 4.1 is really bad (and found one) Xrootd/SLAC Xrootd/dCache NFS/pNFS dCap (old vers) Optimized for ROOT Read two branches TreeTCache ON Vector read effect. The ROOT driver is not doing vector read for plain file systems but for dCap/xRoot

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 32 DPM NFS 4.1 – status First full feature prototype in use internally –Integrated in a DPM testbed at CERN –Mounted on a client machine running a kernel –Standard NFS functional tests running daily Features –Namespace management (mkdir, ls, rm, stat, …) –Permissions (with Kerberos authentication) –Space usage reports (du, df, …) Ongoing –Extend tests to real world use cases from experiments –Performance improvements in the DPM to cope with NFS access patterns (heavy on metadata queries)

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 33 Conclusions (I) Both ATLAS and CMS are now starting to put in production a globally distributed data storage based on xrootd Main goals are seamless access, fallback, diskless Tier-3’s, caching dCache and DPM seriously committed to provide NFS 4.1 support dCache.org is member of the CITI group which is coordinating the NFS 4.1 efforts Compatibility regularly verified

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 34 Conclusions (II) NFS 4.1 performance Already exceeding expectations in dCache Measurements strongly depend on the way the file (ROOT) was written the access profile (ROOT script) Still the protocol seems to be well understood by dCache Performance tests in DPM to be performed using real applications

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 35 Acknowledgements Thanks to B. Bockelman (CMS), P. Fuhrmann (dCache), D. Benjamin (ATLAS) for their input