Distributed Xrootd Derek Weitzel & Brian Bockelman.

Slides:



Advertisements
Similar presentations
Windows Server ® 2008 File Services Infrastructure Planning and Design Published: June 2010 Updated: November 2011.
Advertisements

Database Architectures and the Web
Business Continuity and DR, A Practical Implementation Mich Talebzadeh, Consultant, Deutsche Bank
Web Caching Schemes1 A Survey of Web Caching Schemes for the Internet Jia Wang.
Chapter 9 : Distributed Database.
June 21, PROOF - Parallel ROOT Facility Maarten Ballintijn, Rene Brun, Fons Rademakers, Gunter Roland Bring the KB to the PB.
Grids and Grid Technologies for Wide-Area Distributed Computing Mark Baker, Rajkumar Buyya and Domenico Laforenza.
World Wide Web Caching: Trends and Technology Greg Barish and Katia Obraczka USC Information Science Institute IEEE Communications Magazine, May 2000 Presented.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
CVMFS: Software Access Anywhere Dan Bradley Any data, Any time, Anywhere Project.
RAID-x: A New Distributed Disk Array for I/O-Centric Cluster Computing Kai Hwang, Hai Jin, and Roy Ho.
Communication Part IV Multicast Communication* *Referred to slides by Manhyung Han at Kyung Hee University and Hitesh Ballani at Cornell University.
Microsoft ® Application Virtualization 4.6 Infrastructure Planning and Design Published: September 2008 Updated: February 2010.
Client/Server Grid applications to manage complex workflows Filippo Spiga* on behalf of CRAB development team * INFN Milano Bicocca (IT)
Distributed Data Stores – Facebook Presented by Ben Gooding University of Arkansas – April 21, 2015.
A Scalable Framework for the Collaborative Annotation of Live Data Streams Thesis Proposal Tao Huang
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
Implementing Database Snapshot & Database Mirroring in SQL Server 2005 Presented by Tarek Ghazali IT Technical Specialist Microsoft SQL Server MVP Microsoft.
CSC 456 Operating Systems Seminar Presentation (11/13/2012) Leon Weingard, Liang Xin The Google File System.
DONE-10: Adminserver Survival Tips Brian Bowman Product Manager, Data Management Group.
Database Design – Lecture 16
The Next Generation Root File Server Andrew Hanushevsky Stanford Linear Accelerator Center 27-September-2004
Distributed File Systems
Overcast: Reliable Multicasting with an Overlay Network CS294 Paul Burstein 9/15/2003.
Lecture 3: Sun: 16/4/1435 Distributed Computing Technologies and Middleware Lecturer/ Kawther Abas CS- 492 : Distributed system.
Hadoop Hardware Infrastructure considerations ©2013 OpalSoft Big Data.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
IMDGs An essential part of your architecture. About me
Designing Authentication for a Microsoft Windows 2000 Network Designing Authentication in a Microsoft Windows 2000 Network Designing Kerberos Authentication.
PhysX CoE: LHC Data-intensive workflows and data- management Wahid Bhimji, Pete Clarke, Andrew Washbrook – Edinburgh And other CoE WP4 people…
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
Loosely Coupled Parallelism: Clusters. Context We have studied older archictures for loosely coupled parallelism, such as mesh’s, hypercubes etc, which.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Plethora: A Wide-Area Read-Write Storage Repository Design Goals, Objectives, and Applications Suresh Jagannathan, Christoph Hoffmann, Ananth Grama Computer.
Distributed Information Systems. Motivation ● To understand the problems that Web services try to solve it is helpful to understand how distributed information.
Microsoft Management Seminar Series SMS 2003 Change Management.
Performance Tests of DPM Sites for CMS AAA Federica Fanzago on behalf of the AAA team.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Implementation and performance analysis of.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
HADOOP DISTRIBUTED FILE SYSTEM HDFS Reliability Based on “The Hadoop Distributed File System” K. Shvachko et al., MSST 2010 Michael Tsitrin 26/05/13.
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
CMS Issues. Background – RAL Infrastructure TM Nsd Xrd- mgr TM Nsd Xrd- mgr TM Rhd stagerd TGW Rhd stagerd TGW Cupv Vmgr Vdqm nsd Cupv Vmgr Vdqm nsd Cupv.
PanDA Status Report Kaushik De Univ. of Texas at Arlington ANSE Meeting, Nashville May 13, 2014.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
Federated Data Stores Volume, Velocity & Variety Future of Big Data Management Workshop Imperial College London June 27-28, 2013 Andrew Hanushevsky, SLAC.
09/13/04 CDA 6506 Network Architecture and Client/Server Computing Peer-to-Peer Computing and Content Distribution Networks by Zornitza Genova Prodanoff.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
J. Templon Nikhef Amsterdam Physics Data Processing Group Storage federations Why & How : site perspective Jeff Templon Pre-GBD october CDN vision.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Distributed File System. Outline Basic Concepts Current project Hadoop Distributed File System Future work Reference.
Replicazione e QoS nella gestione di database grid-oriented Barbara Martelli INFN - CNAF.
Improving Performance using the LINUX IO Scheduler Shaun de Witt STFC ISGC2016.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
An Analysis of Data Access Methods within WLCG Shaun de Witt, Andrew Lahiff (STFC)
Any Data, Anytime, Anywhere Dan Bradley representing the AAA Team At OSG All Hands Meeting March 2013, Indianapolis.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Solutions for WAN data access: xrootd and NFSv4.1 Andrea Sciabà.
Efi.uchicago.edu ci.uchicago.edu Sharing Network Resources Ilija Vukotic Computation and Enrico Fermi Institutes University of Chicago Federated Storage.
Presenter: Yue Zhu, Linghan Zhang A Novel Approach to Improving the Efficiency of Storing and Accessing Small Files on Hadoop: a Case Study by PowerPoint.
CMS data access Artem Trunov. CMS site roles Tier0 –Initial reconstruction –Archive RAW + REC from first reconstruction –Analysis, detector studies, etc.
Global Data Access – View from the Tier 2
Spark Presentation.
Thoughts on Computing Upgrade Activities
Ákos Frohner EGEE'08 September 2008
Gregory Kesden, CSE-291 (Storage Systems) Fall 2017
Gregory Kesden, CSE-291 (Cloud Computing) Fall 2016
Hadoop Technopoints.
Peta-Cache: Electronics Discussion I Gunther Haller
Presentation transcript:

Distributed Xrootd Derek Weitzel & Brian Bockelman

Outline Motivation Data Access and Xrootd Concepts WLCG Demonstrator Project Sample T3 sites Interacting with Users

Motivation Data scalability for the next LHC run in 2013 Tiered architecture = many copies Explicit data placement implies we must be smarter than our users and know future usage patterns This is a common problem, internet companies have found solutions a decade ago. Common solution = Caching

Data Access We want to work on improving end-user data access. Goals: Reliability: The end-user should never see an I/O error or failure propagated up to their application unless no USCMS site can serve the file. Failures should be caught as early as possible and I/O retried or rerouted to a different site (possibly degrading the service slightly). Transparency: All actions of the underlying system should be automatic for the user – catalog lookups, redirections, reconnections. There should not be a different workflow for accessing the data ``close by" versus halfway around the world. This implies the system serves user requests almost instantly; opening files should be a ``lightweight" operation. Usability: All CMS application frameworks (CMSSW, FWLite, bare ROOT) must natively integrate with any proposed solution. The proposed solution must not degrade the event processing rate significantly.ROOT Global: A CMS user should be able to get at any CMS file through the Xrootd service. To achieve these goals, we are investigating xrootd-based technology.

Xrootd Concepts Getting access to data: – User contacts a redirector. – Redirector searches through all the nodes which subscribe to it to see who has the data. – Selects the “Best Node” to serve the file. – Client is redirected to the correct data node. Originally, this was done within a site – the “redirector” was the headnode and the “data server” was a big disk server.

Xrootd Concepts

In fact, we can make the redirector a global node and the “data server” an entire site, as demonstrated in the previous slide. This is a one-layer tree: you can actually have an arbitrary number of layers. – Xrootd will first query “down the tree” to look for close-by replicas; if none are found, it goes “up the tree” to find others in the cluster. In order to provide high reliability, the concepts of multiple parent nodes was added – meaning Xrootd is really providing a resilient network to find data. Can be an arbitrary graph, not necessarily a tree.

WLCG Demonstrator Project Deliverables Create a global-scale data access architecture architecture targeted for user analysis jobs. – Based upon xrootd, using both remote streaming and local caches. – Discover and understand the weak points of the xrootd software which would prevent the system from moving into production for CMS. Improve CMSSW I/O to decrease sensitivity to latency (reduce # of reads by a factor of 10). Target T3s and physicist workstations as initial consumers of this architecture. Understand effects of the new architecture on the network and storage system.

CMSSW Deliverables Fault tolerant file opens – If a file is not found locally, try another method (such as reading from global xrootd). Resistant to latency – Reduce number of reads (good for the soul) – Grouping reads into 1 transaction with storage (“vector reads”) – Issuing reads before needing data (prefetching)

Sample T3 Configs - Diskless Deployed at T3_US_Omaha!

Sample T3 Config – Remote Fallback Requires capabilities coming in CMSSW_3_9_0

Sample T3 Configs –Full Xrootd Deploy Deploying at T3_US_UCR!

Sample T3 Configs Note that you don’t need to switch your site to xrootd except in the last picture. – However, the last case is the most tightly integrated and has the best caching. – So, you can benefit and participate with hardly any work on your side

Interacting with Users Any CMSSW user tool already works with this. Doesn’t require grid software (Mac friendly!) – But does require grid certificate! (Demo with Fireworks, network permitting)

Documentation sXrootdArchitecture