Xrootd Roadmap Atlas Tier 3 Meeting University of Chicago September 12-13, 2011 Andrew Hanushevsky, SLAC

Slides:



Advertisements
Similar presentations
Wei Lu 1, Kate Keahey 2, Tim Freeman 2, Frank Siebenlist 2 1 Indiana University, 2 Argonne National Lab
Advertisements

Potential Data Access Architectures using xrootd OSG All Hands Meeting Harvard University March 7-11, 2011 Andrew Hanushevsky, SLAC
Role Based VO Authorization Services Ian Fisk Gabriele Carcassi July 20, 2005.
GUMS status Gabriele Carcassi PPDG Common Project 12/9/2004.
Xrootd Update OSG All Hands Meeting University of Nebraska March 19-23, 2012 Andrew Hanushevsky, SLAC
Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
1 © 2006 Cisco Systems, Inc. All rights reserved. Session Number Presentation_ID Using the Cisco Technical Support & Documentation Website for Security.
Web Servers Web server software is a product that works with the operating system The server computer can run more than one software product such as .
Scalla Back Through The Future Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 8-April-10
XRootD Roadmap To Start The Second Decade Root Workshop Saas-Fee March 11-14, 2013 Andrew Hanushevsky, SLAC
The Next Generation Root File Server Andrew Hanushevsky Stanford Linear Accelerator Center 27-September-2004
Scalla/xrootd Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 19-August-2009 Atlas Tier 2/3 Meeting
Xrootd Demonstrator Infrastructure OSG All Hands Meeting Harvard University March 7-11, 2011 Andrew Hanushevsky, SLAC
Scalla/xrootd Andrew Hanushevsky, SLAC SLAC National Accelerator Laboratory Stanford University 19-May-09 ANL Tier3(g,w) Meeting.
Scalla/xrootd Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 29-October-09 ATLAS Tier 3 Meeting at ANL
May 8, 20071/15 VO Services Project – Status Report Gabriele Garzoglio VO Services Project – Status Report Overview and Plans May 8, 2007 Computing Division,
XMPP Concrete Implementation Updates: 1. Why XMPP 2 »XMPP protocol provides capabilities that allows realization of the NHIN Direct. Simple – Built on.
PanDA Multi-User Pilot Jobs Maxim Potekhin Brookhaven National Laboratory Open Science Grid WLCG GDB Meeting CERN March 11, 2009.
Scalla/xrootd Introduction Andrew Hanushevsky, SLAC SLAC National Accelerator Laboratory Stanford University 6-April-09 ATLAS Western Tier 2 User’s Forum.
Bookkeeping Tutorial. Bookkeeping & Monitoring Tutorial2 Bookkeeping content  Contains records of all “jobs” and all “files” that are created by production.
The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within.
Xrootd Monitoring Atlas Software Week CERN November 27 – December 3, 2010 Andrew Hanushevsky, SLAC.
New perfSonar Dashboard Andy Lake, Tom Wlodek. What is the dashboard? I assume that everybody is familiar with the “old dashboard”:
Support in setting up a non-grid Atlas Tier 3 Doug Benjamin Duke University.
Communicating Security Assertions over the GridFTP Control Channel Rajkumar Kettimuthu 1,2, Liu Wantao 3,4, Frank Siebenlist 1,2 and Ian Foster 1,2,3 1.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Xrootd Update Andrew Hanushevsky Stanford Linear Accelerator Center 15-Feb-05
CERN Using the SAM framework for the CMS specific tests Andrea Sciabà System Analysis WG Meeting 15 November, 2007.
OSG Tier 3 support Marco Mambelli - OSG Tier 3 Dan Fraser - OSG Tier 3 liaison Tanya Levshina - OSG.
Accelerating Debugging In A Highly Distributed Environment CHEP 2015 OIST Okinawa, Japan April 28, 2015 Andrew Hanushevsky, SLAC
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
Performance and Scalability of xrootd Andrew Hanushevsky (SLAC), Wilko Kroeger (SLAC), Bill Weeks (SLAC), Fabrizio Furano (INFN/Padova), Gerardo Ganis.
Xrootd Update Alice Tier 1/2 Workshop Karlsruhe Institute of Technology (KIT) January 24-26, 2012 Andrew Hanushevsky, SLAC
MultiJob pilot on Titan. ATLAS workloads on Titan Danila Oleynik (UTA), Sergey Panitkin (BNL) US ATLAS HPC. Technical meeting 18 September 2015.
Configuring and Troubleshooting Identity and Access Solutions with Windows Server® 2008 Active Directory®
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
ATLAS XRootd Demonstrator Doug Benjamin Duke University On behalf of ATLAS.
XRootD & ROOT Considered Root Workshop Saas-Fee September 15-18, 2015 Andrew Hanushevsky, SLAC
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Bookkeeping Tutorial. 2 Bookkeeping content  Contains records of all “jobs” and all “files” that are produced by production jobs  Job:  In fact technically.
Scalla Authorization xrootd /cmsd Andrew Hanushevsky SLAC National Accelerator Laboratory CERN Seminar 10-November-08
Slide 1/29 Informed Prefetching in ROOT Leandro Franco 23 June 2006 ROOT Team Meeting CERN.
National Energy Research Scientific Computing Center (NERSC) CHOS - CHROOT OS Shane Canon NERSC Center Division, LBNL SC 2004 November 2004.
ALCF Argonne Leadership Computing Facility GridFTP Roadmap Bill Allcock (on behalf of the GridFTP team) Argonne National Laboratory.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Xrootd Proxy Service Andrew Hanushevsky Heinz Stockinger Stanford Linear Accelerator Center SAG September-04
Scalla As a Full-Fledged LHC Grid SE Wei Yang, SLAC Andrew Hanushevsky, SLAC Alex Sims, LBNL Fabrizio Furano, CERN SLAC National Accelerator Laboratory.
Daniele Spiga PerugiaCMS Italia 14 Feb ’07 Napoli1 CRAB status and next evolution Daniele Spiga University & INFN Perugia On behalf of CRAB Team.
Federated Data Stores Volume, Velocity & Variety Future of Big Data Management Workshop Imperial College London June 27-28, 2013 Andrew Hanushevsky, SLAC.
SRM-2 Road Map and CASTOR Certification Shaun de Witt 3/3/08.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
Free, online, technical courses Take a free online course. Microsoft Virtual Academy.
A System for Monitoring and Management of Computational Grids Warren Smith Computer Sciences Corporation NASA Ames Research Center.
New Features of Xrootd SE Wei Yang US ATLAS Tier 2/Tier 3 meeting, University of Texas, Arlington,
OSG STORAGE OVERVIEW Tanya Levshina. Talk Outline  OSG Storage architecture  OSG Storage software  VDT cache  BeStMan  dCache  DFS:  SRM Clients.
Monitoring Dynamic IOC Installations Using the alive Record Dohn Arms Beamline Controls & Data Acquisition Group Advanced Photon Source.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Solutions for WAN data access: xrootd and NFSv4.1 Andrea Sciabà.
CERN IT Department CH-1211 Genève 23 Switzerland t DPM status and plans David Smith CERN, IT-DM-SGT Pre-GDB, Grid Storage Services 11 November.
ALL THINGS IIS TERRI DONAHUE
SQL Database Management
WLCG Transfers Dashboard
Jean-Philippe Baud, IT-GD, CERN November 2007
SLAC National Accelerator Laboratory
Brookhaven National Laboratory Storage service Group Hironori Ito
Discussions on group meeting
Data Management cluster summary
IIS.
STATEL an easy way to transfer data
Presentation transcript:

xrootd Roadmap Atlas Tier 3 Meeting University of Chicago September 12-13, 2011 Andrew Hanushevsky, SLAC

September 12-13, 20112ATLAS Tier 3 Meeting Outline Recent additions FS-Independent Extended Attribute Framework Shared-Everything File System Support Meta-Manager throttling On the horizon Near future Way out there Conclusion Acknowledgements

September 12-13, 20113ATLAS Tier 3 Meeting Recent Additions FS-Independent Extended Attribute Framework Used to save file-specific information Migration time, residency requirements, checksums Shared-Everything File System Support Optimize file discovery in distributed file systems xrootd dCache, DPM, GPFS, HDFS, Lustre, proxy xrootd Meta-Manager throttling Configurable per-site query limits

September 12-13, 20114ATLAS Tier 3 Meeting On the horizon Security New source build procedures EPEL Guidelines Integrated checksums Extended monitoring Alternate Name2Name Plug-in Dropping RH4 support Planned for 3.2 release

September 12-13, 20115ATLAS Tier 3 Meeting Security Part 1 Enabling x509 authentication & Authorization Motivated by server-server transfers via FRM Required by ATLAS security Requires additional site configuration Obtaining site x509 certificates These are certificate and private key pem files Placing these in /etc/grid-security/xrd By default, xrdcert.pem and xrdkey.pem

September 12-13, 20116ATLAS Tier 3 Meeting Security Part 2 Periodically creating voms proxy certificates This is to get a valid voms ATLAS extension Probably done via cron job These are used by the FRM to authenticate site Installing the ATLAS x509 mapping plug-in Likely distributed via the OSG rpm We haven’t fleshed out the details

September 12-13, 20117ATLAS Tier 3 Meeting Security Part 3 Periodically creating voms proxy certificates This is to get a valid voms ATLAS extension Probably done via cron job Though are considering long-lived proxy certificate These are used by the FRM to authenticate site Installing the ATLAS x509 mapping plug-in Likely distributed via the OSG rpm We still have to discuss this with OSG

September 12-13, 20118ATLAS Tier 3 Meeting Security Part 4 Configuring xrootd to force x509 authentication xrootd.seclib libpath/XrdSec.so sec.protocol libpath gsi \ -authzfun:libXrdAuthzAtlas.so \ -gmapopt:10 -gmapto:0 ofs.authorize acc.authdb path/dbfname The dbfname contains the line: g atlas / rl

September 12-13, 20119ATLAS Tier 3 Meeting Security Part 5 What all this does Requires client to provide ATLAS certificate The xrootd.seclib and sec.protocol directives libXrdAuthzAtlas.so maps valid cert to group atlas The authzfun parameter in sec.protocol The authdb file says group atlas has r/o access ‘g’ for group ‘/’ for everything you export ‘rl’ for read and search access

September 12-13, ATLAS Tier 3 Meeting Security Part 6 Simplifying the side-effects Normally, this requires everyone to have a cert! This is very intrusive for most T3 sites We can restrict this to only the proxy server This means you need to run a proxy server Many if not most sites will need to run one due to firewalls Only outside clients will need to have a valid cert It is possible to do this without a proxy The configuration becomes a bit more complicated

September 12-13, ATLAS Tier 3 Meeting Certificate Issues! Voms certs only issued to individuals This makes site and host certs problematic We really have no real solution to this now Virtual solution Get user cert that corresponds to xrootd user Have that cert validated by voms Use it as site service certificate For host identify & individual host access We don’t really know if this will work, SIGH

September 12-13, ATLAS Tier 3 Meeting Security Status We have a working autz mapping function Based on s/w from Matevz Tadel & Brian Bockelman Needs some clean-up and better packaging Distribution needs to be decided Likely via OSG just like gridftp add-ons Certificate plan needs to be put into place How to obtain one & creating voms proxy certificates Where to place all of these Will likely need additional sec.protocol options

September 12-13, ATLAS Tier 3 Meeting Build using cmake Currently, we support two source build methods autotools & configure.classic The 3.1 release will use cmake This will displace autotools Support for configure.classic yet unclear The root team wants it maintained for MacOS We are in negotiations This means you will need to install cmake Only if you want to build from the source

September 12-13, ATLAS Tier 3 Meeting EPEL Guidelines New guidelines prohibit installing ‘.a’ files Most ‘.a’ files will be replaced by ‘.so’ files We are trying to consolidate libraries This will limit the number of installed shared libraries Impact is minimal except for plug-in writers Will likely need to change your link step This will occur when we switch to cmake Planned for 3.1 release

September 12-13, ATLAS Tier 3 Meeting Integrated Checksums Currently, checksum calculated outboard Program is specified via configuration file The xrdadler32 command also checksums Records checksum in extended attributes for future New xrootd will do inboard checksumming Will record checksum in extended attributes Many configuration options available Should speed up SRM queries Planned for 3.1

September 12-13, ATLAS Tier 3 Meeting Extended Monitoring Redirect information will be provided Selectable via configuration option Will provide information on who went where Currently, only available via debug log output (yech) Per client I/O monitoring will be flushable Currently, I/O statistics flushed when buffer full Will be able to specify a flush window Based on code provided by Matevz Tadel, CMS Planned for 3.1 (redirect info likely in 3.1.x)

September 12-13, ATLAS Tier 3 Meeting Alternate N2N Plug-in Currently, name-to-name plug-in limited It is not passed contextual information Will be able to specify alternate plug-in Alternate will get contextual information I.E. url cgi information This makes it binary compatible with past plug-ins Planned for 3.1

September 12-13, ATLAS Tier 3 Meeting Things in the near future Extended POSC xrootd New xrootd client Specialized meta-manager Integrated alerts New async I/O model

September 12-13, ATLAS Tier 3 Meeting Extended POSC Currently, adding “?ofs.posc=1” enables POSC Persist On Successful Close This can be extended to support checksums E.G. “?ofs.posc=%adler32:csval” File persists on successful close AND Supplied checksum matches Privilege & error ending states not yet defined Planned for 1Q12

September 12-13, ATLAS Tier 3 Meeting New Client Current client uses basic thread model Limits scaling and is resource intensive New client will use robust thread model Scalable and fully asynchronous Will be the platform for future features E.G. plug-in caches, local redirects, etc Planned for 1Q12 This will likely be an alpha release

September 12-13, ATLAS Tier 3 Meeting Specialized Meta-Manager Current MM is a regular manager with mm role This limits what the meta-manager can do Extending it unduly impacts the manager’s code The specialized MM is a separate daemon Will allow many more subscribers Can better optimize handling federated managers Planned 2Q12

September 12-13, ATLAS Tier 3 Meeting Integrated Alerts Currently, alerts based on using monitoring Monitoring provides broad usage information Alerts are therefore macro-scale We want to send a separate alert stream Based on unusual internal events E.G. unexpected latency, server recovery actions, etc Planned 3Q12 Part of message and logging restructuring

September 12-13, ATLAS Tier 3 Meeting New Async I/O Model xrootd Currently xrootd uses OS supplied async I/O This is not particularly useful In Linux it is simulated as it was never kernel level In other OS’s it uses a lot of CPU resources xrootd In fact, xrootd normally bypasses it for most requests The next version will use a thread model Based on looking ahead on the request stream This should be more applicable to most requests Planned 4Q12

September 12-13, ATLAS Tier 3 Meeting Way Out There Xinetd proxy support IPV6

September 12-13, ATLAS Tier 3 Meeting Xinetd Based Proxy xrootd Currently, proxy support provided by xrootd With some tinkering it’s possible to use xinetd We don’t know yet if this is useful Internal security will apply to external clients Also, seems limited to very small static clusters Requires re-config every time the cluster changes But it may appeal to some sites No implementation plan yet Does anyone really want this?

September 12-13, ATLAS Tier 3 Meeting IPV6 Currently, all new code supports IPV6 But existing code needs to change We are not sure how critical this really is And it has side-effects E.G. all IP address in messages would change No implementation plan yet How critical is this in practice?

September 12-13, ATLAS Tier 3 Meeting Conclusion xrootd xrootd is under active development Always looking for new ideas Feel free to suggest them Be a contributor You too can contribute to the code base xrootd Consider joining the xrootd collaboration Currently CERN, SLAC, and Duke are members See more at

September 12-13, ATLAS Tier 3 Meeting Acknowledgements Current Software Contributors ATLAS: Doug Benjamin, Patrick McGuigan, Danila Oleynik, Artem Petrosyan CERN: Fabrizio Furano, Lukasz Janyst, Andreas Peters, David Smith CMS: Brian Bockelman (unl), Matevz Tadel (ucsd) Fermi/GLAST: Tony Johnson FZK: Artem Trunov BeStMan LBNL: Alex Sim, Junmin Gu, Vijaya Natarajan (BeStMan team) Root: Gerri Ganis, Beterand Bellenet, Fons Rademakers OSG: Tim Cartwright, Tanya Levshina SLAC: Andrew Hanushevsky, Wilko Kroeger, Daniel Wang, Wei Yang Operational Collaborators ANL, BNL, CERN, FZK, IN2P3, SLAC, UCSD, UTA, UoC, UNL, UVIC, UWisc US Department of Energy Contract DE-AC02-76SF00515 with Stanford University