XRootD Release 4.5 And Beyond

Slides:



Advertisements
Similar presentations
Wei Lu 1, Kate Keahey 2, Tim Freeman 2, Frank Siebenlist 2 1 Indiana University, 2 Argonne National Lab
Advertisements

Distributed Xrootd Derek Weitzel & Brian Bockelman.
EGEE is a project funded by the European Union under contract IST Using SRM: DPM and dCache G.Donvito,V.Spinoso INFN Bari
1 Chapter 11: File-System Interface  File Concept  Access Methods  Directory Structure  File System Mounting  File Sharing  Protection  Chapter.
Adding scalability to legacy PHP web applications Overview Mario A. Valdez-Ramirez.
Web Caching Schemes1 A Survey of Web Caching Schemes for the Internet Jia Wang.
Extensible Scalable Monitoring for Clusters of Computers Eric Anderson U.C. Berkeley Summer 1997 NOW Retreat.
Other File Systems: AFS, Napster. 2 Recap NFS: –Server exposes one or more directories Client accesses them by mounting the directories –Stateless server.
ArcGIS for Server Reference Implementations An ArcGIS Server’s architecture tour.
PRASHANTHI NARAYAN NETTEM.
NETOP ONDEMAND What’s new in version 2.1? DECEMBER 09 NETOP ONDEMAND1.
Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
CVMFS: Software Access Anywhere Dan Bradley Any data, Any time, Anywhere Project.
TAM STE Series 2008 © 2008 IBM Corporation WebSEAL SSO, Session 108/2008 TAM STE Series WebSEAL SSO, Session 1 Presented by: Andrew Quap.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
CSC 456 Operating Systems Seminar Presentation (11/13/2012) Leon Weingard, Liang Xin The Google File System.
10 May 2007 HTTP - - User data via HTTP(S) Andrew McNab University of Manchester.
Xrootd Authentication & Authorization Andrew Hanushevsky Stanford Linear Accelerator Center 6-June-06.
The Next Generation Root File Server Andrew Hanushevsky Stanford Linear Accelerator Center 27-September-2004
Computing on the Cloud Jason Detchevery March 4 th 2009.
Scalable Web Server on Heterogeneous Cluster CHEN Ge.
Scalla/xrootd Introduction Andrew Hanushevsky, SLAC SLAC National Accelerator Laboratory Stanford University 6-April-09 ATLAS Western Tier 2 User’s Forum.
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Communicating Security Assertions over the GridFTP Control Channel Rajkumar Kettimuthu 1,2, Liu Wantao 3,4, Frank Siebenlist 1,2 and Ian Foster 1,2,3 1.
Xrootd Update Andrew Hanushevsky Stanford Linear Accelerator Center 15-Feb-05
Chapter 10 Chapter 10: Managing the Distributed File System, Disk Quotas, and Software Installation.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
ROOT and Federated Data Stores What Features We Would Like Fons Rademakers CERN CC-IN2P3, Nov, 2011, Lyon, France.
Extending ISA/IAG beyond the limit. AGAT Security suite - introduction AGAT Security suite is a set of unique components that allow extending ISA / IAG.
Andrew McNab - HTTP/HTTPS extensions HTTP/HTTPS as Grid data transport 6 March 2003 Andrew McNab, University of Manchester
Federated Data Stores Volume, Velocity & Variety Future of Big Data Management Workshop Imperial College London June 27-28, 2013 Andrew Hanushevsky, SLAC.
1 Xrootd-SRM Andy Hanushevsky, SLAC Alex Romosan, LBNL August, 2006.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
An Introduction to GPFS
New Features of Xrootd SE Wei Yang US ATLAS Tier 2/Tier 3 meeting, University of Texas, Arlington,
dCache Paul Millar, on behalf of the dCache Team
Web and Proxy Server.
ArcGIS for Server Security: Advanced
a brief summary for users
Jean-Philippe Baud, IT-GD, CERN November 2007
WLCG IPv6 deployment strategy
Scalable sync-and-share service with dCache
Apache web server Quick overview.
Network Cost Services for Network-Aware FI Applications
Third Party Transfers & Attribute URI ideas
N-Tier Architecture.
70-293: MCSE Guide to Planning a Microsoft Windows Server 2003 Network, Enhanced Chapter 6: Planning, Configuring, And Troubleshooting WINS.
draft-ietf-simple-message-sessions-00 Ben Campbell
Chapter 11: File-System Interface
Data Federation with Xrootd Wei Yang US ATLAS Computing Facility meeting Southern Methodist University, Oct 11-12, 2011.
Brookhaven National Laboratory Storage service Group Hironori Ito
Whether you decide to use hidden frames or XMLHttp, there are several things you'll need to consider when building an Ajax application. Expanding the role.
Ákos Frohner EGEE'08 September 2008
Gregory Kesden, CSE-291 (Storage Systems) Fall 2017
Gregory Kesden, CSE-291 (Cloud Computing) Fall 2016
Internet Networking recitation #12
Chapter 16: Distributed System Structures
Dynamic DNS support for EGI Federated cloud
A Web-Based Data Grid Chip Watson, Ian Bird, Jie Chen,
Support for ”interactive batch”
Directory Structure A collection of nodes containing information about all files Directory Files F 1 F 2 F 3 F 4 F n Both the directory structure and the.
Secondary Storage Management Brian Bershad
Grid Coordination by Using the Grid Coordination Protocol
Specialized Cloud Architectures
Any Data, Anytime, Anywhere
Secondary Storage Management Hank Levy
Database System Architectures
Preventing Privilege Escalation
Lecture 4: File-System Interface
Presentation transcript:

XRootD Release 4.5 And Beyond XRootD Workshop Tokyo Stanford University/SLAC November 10, 2016 Andrew Hanushevsky, SLAC http://xrootd.org

November Release 4.5 Highlights Request signing Dual stack networking option Client reporting of release at login time Separate negative cache timeout in cmsd Allow file names with spaces Allow host names starting with digit Automatic URL prefixing Zip archive support

Request Signing Protects XRootD servers from bad actors Client Server Request Response Delete File x Client can cryptographically sign requests Sever verifies request came from the same authenticated client Bad actor problem avoided

Enabling Request Signing Request signing gives you peace of mind When allowing R/W access to the server Especially on the WAN Server configuration option sec.level {all|local|remote] [relaxed] level all applies level to local and remote client local and remote provide split options relaxed provides a migration path It requires signing only for 4.5 and up clients

Request Signing Levels none The default compatible Only destructive operations Is compatible with R/O access for old clients standard | intense | pedantic Each requires more operations to be signed

Request Verification by Level Operation Compatible Standard Intense Pedantic admin verified auth --- bind chmod close decrypt dirlist endsess getfile locate login mkdir mv open read open Write ping prepare protocol putfile query query special read readv rm rmdir set set special sigver stat statx sync truncate verifyw write

Dual Stack Networking Option Pre-4.3 clients may report as IPv6 only This is a big headache for IPv4-only servers New server-side option xrd.network assumev4 Server will assume client has IPv4 Only applied to 4.4 or older clients Server can’t detect a client’s release level until…

Client Release Reporting Client will report it’s release level Happens at login time This allows future server-side bypasses When a particular release has a bug Upgrading server can bypass old client bugs Make client migration much easier

Separate Negative Cache Timeout The cmsd caches file location Implicitly caches missing files as well Default is 8 hours Incorrectly missing files will be missing 8 hours Unless data server updates the cahe New server configuration option cms.fxhold noloc ntime[h|m|s]  ntime expiration for cached missing files only

Allow File Names With Spaces XRootD allows spaces in file names Every operation except rename Protocol extension covers rename now All 4.5 plus clients use the protocol extension

Host Names Starting With Digits XRootD originally adhered to RFC 952 Hostnames may contain letters, digits, dashes But may not start with a digit Now it adheres to RFC 1123 Supplants RFC 952 allows 1st char as a digit Required for auto-generated hostnames Typically a problem for VM’s and containers

Automatic URL Prefixing Required by fully firewalled sites Outside World Forwarding Proxy Server Inside Client Much like HTTP proxy Open(“root://x//file”) Automatically converted to Open(“root://proxy//root://x//file”) Done as a configurable client plug-in Usable by 4.0 and above clients Will address multi-tenant sites later

Zip Archive Support Fully implemented as a client feature Allows extraction of file from archive No need to transmit the whole archive Covered by Elvin’s talk

January Release 4.6 Async I/O Proxy Handling Disk Caching Proxy Possible other addition Extreme (multi-source) copy Perhaps others to be determined

Async I/O Proxy Handling Proxy now handles async I/O requests Previously converted async to sync I/O Motivation Improve streaming performance of xrdcp Improve Disk Caching Proxy performance Certain background operations (e.g. pre-fetch) May require tuning to get best performance See xrootd.async directive

Disk Caching Proxy File or block level caching

Typical Disk Caching Proxy Uses Remote XRootD Clusters Speed up Remote Access Self-Managed SE’s Local Clients Caching Proxy Speed up Random Access XRootD Server HDFS Caching Proxy Speed up HD Access SSD XRootD Cluster FS

More On Disk Caching Proxy High potential to solve vexing problems Reduced remote latency for user analysis Just in time data Avoids pre-placement delays Optimized disk space utilization Cached data access via xroot and http Currently being tested at scale MWT2, SLAC, University of Notre Dame

Disk Caching Proxy Caveats Needs installation of tcmalloc or jemalloc Avoids memory growth problems in glibc It’s very easy to overload the proxy Lager sites should consider caching clusters Two or more proxies clustered together Fully supported upon release

Disk Caching Proxy Deployment Target sites without ATLAS managed disk Opportunistic sites OSG diskless sites Pacific Research Platform sites (NSF funded program) NASA, NREN, U Washington, UC School System Includes University California LHC sites (3 ATLAS T3’s) Interconnected via CalREN, ESNET, & Pacific Wave Sites and networking may expand

In The Pipeline for 4.7 or 4.8 Space quotas Scalable Service Interface

Space Quotas Experiments want write access to XRootD Current development Non-hierarchical logical path based quotas Quota is soft Roughly within a specified resolution Periodic recalibration Will be a plug-in so can be replaced This is much harder than imagined!

Scalable Service Interface Framework for distributed services Builds on robust & recoverable XRootD Uses a remote object execution model Current deployed for LSST qserv Distributed unshared mySQL servers Successfully being used with 100’s of nodes API is still being refined Will be released when finalized

Future Enhancements (not yet set) Multi-source copy Multi-Source load balancing client Eliminating cmsd write lookup delays Tracking file ownership Eliminating 64-node limit (exploratory) HTTP 2 plug-in

Multi-Source Copy Implement the –sources xrdcp option The source can be a redirector or metalink I/O automatically balanced across sources Advanced algorithm to avoid ending tail XRootD Source 1 Source 2 Source 3 Source 4 xrdcp –sources 4 source target

Multi-Source Load Balancing Client Similar to xrdcp but with a big twist The source can be a redirector or metalink A new source is added only if current one slow Can bounce around multiple sources Determines by real-time performance metrics XRootD Source 1 Source 2 Source 3 Source 4 TFile.open(source);

The Missing File Problem Linux Server Machine A xroot Server Data Files xrdcp root://R//foo /tmp Linux Server Machine R xroot Server Application 1 Who has /foo? xroot Client Redirector open(“/foo”); Linux 2 Nope! Client Machine 5 /foo Data Files xroot Server File deemed not to exist if there is no response after 5 seconds! OK for read access not so much for file creation. Linux Server Machine B

Eliminating cmsd Write Lookup Delay The cmsd uses a no response model No response -> file does not exist Extremely scalable for analysis use case Usually always looking for existing files Not so good for creating files A small change in protocol can fix this Required for efficient handling of experiments desire for writable clusters

Tracking File Ownership UID/GID tracking of ownership Available for certain authentication methods GSI with a gridmap file Kerberos (in domain only) Simple Shared Secret Unix Must start XRootD as root Security considerations abound May allow of uid/gid based quotas

New Third Party Transfer New 3rd Party Transfer Plan to use forwardable credentials X.509 (i.e. Grid Certificates) Allows almost universal 3rd party access Only one of three parties needs to support it The File Residency Manager already does this Will coexist with current mechanism

Eliminating 64-Node Limit I xrootd cmsd xrootd cmsd Manager (Root Node) xrootd cmsd xrootd cmsd 641 = 64 xrootd cmsd 642 = 4096 Supervisors (Interior Nodes) xrootd cmsd 643 = 262144 xrootd cmsd 644 = 16777216 Data Server (Leaf Nodes)

Eliminating 64-Node Limit II A B64 tree architecture is generally ideal Fast scaling and highly parallel file search But it’s cumbersome for large clusters Need to deploy sufficient supervisor nodes Exploring different type of trees B128 B256 B512 etc Parallelism is the biggest stumbling block However, it would simplify configuration

That’s All! What’s Your Wish List?