Xrootd Update Alice Tier 1/2 Workshop Karlsruhe Institute of Technology (KIT) January 24-26, 2012 Andrew Hanushevsky, SLAC

Slides:



Advertisements
Similar presentations
Potential Data Access Architectures using xrootd OSG All Hands Meeting Harvard University March 7-11, 2011 Andrew Hanushevsky, SLAC
Advertisements

Xrootd Roadmap Atlas Tier 3 Meeting University of Chicago September 12-13, 2011 Andrew Hanushevsky, SLAC
Xrootd Update OSG All Hands Meeting University of Nebraska March 19-23, 2012 Andrew Hanushevsky, SLAC
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
Duke and ANL ASC Tier 3 (stand alone Tier 3’s) Doug Benjamin Duke University.
Chapter 9 Chapter 9: Managing Groups, Folders, Files, and Object Security.
Extensible Scalable Monitoring for Clusters of Computers Eric Anderson U.C. Berkeley Summer 1997 NOW Retreat.
Hands-On Microsoft Windows Server 2003 Networking Chapter 7 Windows Internet Naming Service.
70-270, MCSE/MCSA Guide to Installing and Managing Microsoft Windows XP Professional and Windows Server 2003 Chapter Nine Managing File System Access.
Check Disk. Disk Defragmenter Using Disk Defragmenter Effectively Run Disk Defragmenter when the computer will receive the least usage. Educate users.
Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
Windows Server 2008 Chapter 11 Last Update
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
1 The SpaceWire Internet Tunnel and the Advantages It Provides For Spacecraft Integration Stuart Mills, Steve Parkes Space Technology Centre University.
Scalla Back Through The Future Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 8-April-10
XRootD Roadmap To Start The Second Decade Root Workshop Saas-Fee March 11-14, 2013 Andrew Hanushevsky, SLAC
Benjamin Lavalley, Sr. Product Marketing Manager Kaseya 2 Upgrade Review.
Chapter Oracle Server An Oracle Server consists of an Oracle database (stored data, control and log files.) The Server will support SQL to define.
The Next Generation Root File Server Andrew Hanushevsky Stanford Linear Accelerator Center 27-September-2004
Scalla/xrootd Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 19-August-2009 Atlas Tier 2/3 Meeting
5 Chapter Five Web Servers. 5 Chapter Objectives Learn about the Microsoft Personal Web Server Software Learn how to improve Web site performance Learn.
1 Week #9 File Services DFS Overview Configuring DFS Namespaces Configuring DFS Replication Windows Server 2008 Storage Management Overview Managing Storage.
Xrootd Demonstrator Infrastructure OSG All Hands Meeting Harvard University March 7-11, 2011 Andrew Hanushevsky, SLAC
Ideas to Improve SharePoint Usage 4. What are these 4 Ideas? 1. 7 Steps to check SharePoint Health 2. Avoid common Deployment Mistakes 3. Analyze SharePoint.
Scalla/xrootd Andrew Hanushevsky, SLAC SLAC National Accelerator Laboratory Stanford University 19-May-09 ANL Tier3(g,w) Meeting.
Scalla/xrootd Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 29-October-09 ATLAS Tier 3 Meeting at ANL
Scalla/xrootd Introduction Andrew Hanushevsky, SLAC SLAC National Accelerator Laboratory Stanford University 6-April-09 ATLAS Western Tier 2 User’s Forum.
FAX UPDATE 26 TH AUGUST Running issues FAX failover Moving to new AMQ server Informing on endpoint status Monitoring developments Monitoring validation.
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Xrootd Monitoring Atlas Software Week CERN November 27 – December 3, 2010 Andrew Hanushevsky, SLAC.
What’s new in Kentico CMS 5.0 Michal Neuwirth Product Manager Kentico Software.
Overview Managing a DHCP Database Monitoring DHCP
Xrootd Update Andrew Hanushevsky Stanford Linear Accelerator Center 15-Feb-05
Status & Plan of the Xrootd Federation Wei Yang 13/19/12 US ATLAS Computing Facility Meeting at 2012 OSG AHM, University of Nebraska, Lincoln.
Chapter 10 Chapter 10: Managing the Distributed File System, Disk Quotas, and Software Installation.
ROOT and Federated Data Stores What Features We Would Like Fons Rademakers CERN CC-IN2P3, Nov, 2011, Lyon, France.
OSG Tier 3 support Marco Mambelli - OSG Tier 3 Dan Fraser - OSG Tier 3 liaison Tanya Levshina - OSG.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
Accelerating Debugging In A Highly Distributed Environment CHEP 2015 OIST Okinawa, Japan April 28, 2015 Andrew Hanushevsky, SLAC
Performance and Scalability of xrootd Andrew Hanushevsky (SLAC), Wilko Kroeger (SLAC), Bill Weeks (SLAC), Fabrizio Furano (INFN/Padova), Gerardo Ganis.
SKYPIAX, how to add Skype capabilities to FreeSWITCH (and Asterisk) CHICAGO, USA, September 2009.
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
ATLAS XRootd Demonstrator Doug Benjamin Duke University On behalf of ATLAS.
XRootD & ROOT Considered Root Workshop Saas-Fee September 15-18, 2015 Andrew Hanushevsky, SLAC
Scalla Authorization xrootd /cmsd Andrew Hanushevsky SLAC National Accelerator Laboratory CERN Seminar 10-November-08
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
SRM Space Tokens Scalla/xrootd Andrew Hanushevsky Stanford Linear Accelerator Center Stanford University 27-May-08
Scalla As a Full-Fledged LHC Grid SE Wei Yang, SLAC Andrew Hanushevsky, SLAC Alex Sims, LBNL Fabrizio Furano, CERN SLAC National Accelerator Laboratory.
Federated Data Stores Volume, Velocity & Variety Future of Big Data Management Workshop Imperial College London June 27-28, 2013 Andrew Hanushevsky, SLAC.
1 Xrootd-SRM Andy Hanushevsky, SLAC Alex Romosan, LBNL August, 2006.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Dynamic staging to a CAF cluster Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE CAF / PROOF Workshop,
Virtual Machine Movement and Hyper-V Replica
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
New Features of Xrootd SE Wei Yang US ATLAS Tier 2/Tier 3 meeting, University of Texas, Arlington,
Monitoring Dynamic IOC Installations Using the alive Record Dohn Arms Beamline Controls & Data Acquisition Group Advanced Photon Source.
AAF tips and tricks Arsen Hayrapetyan Yerevan Physics Institute, Armenia.
Federating Data in the ALICE Experiment
SQL Database Management
Jean-Philippe Baud, IT-GD, CERN November 2007
Maintaining Windows Server 2008 File Services
SLAC National Accelerator Laboratory
Data Federation with Xrootd Wei Yang US ATLAS Computing Facility meeting Southern Methodist University, Oct 11-12, 2011.
Data Management cluster summary
IIS.
Specialized Cloud Architectures
Summary of the dCache workshop
Presentation transcript:

xrootd Update Alice Tier 1/2 Workshop Karlsruhe Institute of Technology (KIT) January 24-26, 2012 Andrew Hanushevsky, SLAC

January 24-26, 20122Alice Tier 1-2 Workshop Outline Recent additions Recent changes On the horizon Near future Things that you might not know The xrootd collaboration Conclusion Acknowledgements

January 24-26, 20123Alice Tier 1-2 Workshop Recent Additions Extended Attribute Framework Integrated checksums Shared-Everything File System Support Static ENOENT redirection Caching proxy server Federated site shares Monitoring Extended Read-fork-read client

January 24-26, 20124Alice Tier 1-2 Workshop Agnostic Extended Attributes FS-Independent Extended Attribute Framework Used to save file-specific information File system must support extended attributes Normally not a problem except for EXT3 May need to manually enable support in /etc/fstab /dev/hda3 /home ext3 defaults,user_xattr 1 2 If disabled and needed, you will get a log error message Functions needing xattr information then turned off

January 24-26, 20125Alice Tier 1-2 Workshop Current Xattr Information XrdCks.xxxxxx (e.g. md5) checksum XrdFrm.yyyFile Residency Manager info XrdFrm.CpyTime file was copied out XrdFrm.MemMemory mapped access attributes XrdFrm.PinDisk pinning attributes XrdFrm.PfnFile’s physical name Only for symlinked spaces

January 24-26, 20126Alice Tier 1-2 Workshop Future Xattr Information Original file creator The inexorable start of quotas, sigh Extended Access Information Keep last three accesses Number of parallel readers, % read, access time Can be used for better migration/purging

January 24-26, 20127Alice Tier 1-2 Workshop Viewing Xattr Information getfattr -d [-e en] [-m pattern] pathname See Linux man page for details Most information is binary so must use proper en Use pattern to restrict what variable is displayed In Linux, variables are always prefixed by “user.” frm_admin {chksum | find}... Displays info in human understandable terms See

January 24-26, 20128Alice Tier 1-2 Workshop Will I Need To Migrate? Yes, if you use MPS (Migration, Purge, and Staging) or oss.cache or oss.space directives No, otherwise or run in compatibility mode oss.runmodeold directive Will be supported until 2014 Use “frm_admin convert” for conversion See

January 24-26, 20129Alice Tier 1-2 Workshop Integrated Checksums Can configure xrootd to handle checksums xrootd.chksum [max num] {adler32|crc32|md5} Add new ones via ofs.ckslib directive and plug-in Optimize vis ofs.cksrdsz directive Checksum saved in extended attributes Displayed on query Automatically recomputed when file changes Can be disabled External program no longer needed But this is still supported

January 24-26, Alice Tier 1-2 Workshop Shared FS Support The problem The solution Use cms.dfs directive to eliminate duplicate hits Many tuning options available Lookup at redirector, limits, caching, etc See Redirector Data Server Client Shared FS open(x)stat(x)have(x) dCache GPFS HDFS Lustre etc

January 24-26, Alice Tier 1-2 Workshop Static ENOENT Redirection Allows you to redirect client on ENOENT Works for redirectors and data servers Any operation involving a file xrootd.redirect ? path host:port Up to four different “?” redirects allowed Provides for selective redirects based on path Allows more specialized missing file handling Alternative to the FRM staging add-on

January 24-26, Alice Tier 1-2 Workshop ENOENT Redirects What we envision... xyzzy is a caching proxy server or proxy cluster Provides high performance WAN access Client accesses data via WAN in this case Redirector Data Server Client open(x) have(x) xyzzy:1234 ? goto xyzzy:1234 open(x) if redirector_host xrootd.redirect ? / xyzzy:1234 fi

January 24-26, Alice Tier 1-2 Workshop But You Can Do Much More! The xyzzy host can do anything really Have a specialized reselection Then redirect client elsewhere Requires a plug-in implementation ENOENT redirection is a powerful tool Can implement any site-specific recovery actions The logic is centralized and can be uniform

January 24-26, Alice Tier 1-2 Workshop Caching Proxy Server The proxy server plug-in has in-memory cache Must be enabled via pss.memcache directive Many tuning options; some of which are... Cache and page size Read ahead size Maximum block size to cache Root file access optimization See

January 24-26, Alice Tier 1-2 Workshop Federated Site Shares The problem The solution Use cms.sched directive to establish site share Use the gshr and gsdflt options See Meta Manager Site A Site B Site C Client Each site gets an equal number of requests all things being equal

January 24-26, Alice Tier 1-2 Workshop Monitoring Readv requests fully monitored By default, only a readv request summary provided Can request a full unwind of a readv request Per client I/O monitoring now flushable Before, I/O statistics flushed when buffer full Can specify a flush window Based on code provided by Matevz Tadel, CMS Authentication info can now be monitored

January 24-26, Alice Tier 1-2 Workshop Read-Fork-Read Client The current client allows forking This allows sharing data between processes Read conditions data Fork n times for parallel processing Read event data using pre-read conditions data Extensively used by CMS Substantially reduces memory load Critical for large multi-core worker nodes

January 24-26, Alice Tier 1-2 Workshop Recent Changes Switched to git and cmake for build Using EPEL guidelines

January 24-26, Alice Tier 1-2 Workshop Switched To cmake We have dropped old build procedures CVS, autotools, and configure.classic Using git repository for source code git clone git clone Standardized on cmake Source builds, RPM builds, and installs Version 2.8 recommended but 2.6 should work See README file in the top level source directory

January 24-26, Alice Tier 1-2 Workshop Switch Had Side Effects Xrootd will not be part of the root package Will be able to get it separately via RPM process Use yum to install Repositories: xrootd.org or xrootd.cern.ch See Must include repository in /etc/yum.repos.d Process normally determined by experiment

January 24-26, Alice Tier 1-2 Workshop Conform To EPEL Guidelines New guidelines prohibit installing ‘.a’ files All ‘.a’ files are replaced by ‘.so’ files But we consolidated libraries This limits the number of installed shared libraries Still, there are more than there used to be Impact is minimal except for plug-in writers Will likely need to change your link step But now you can easily re-use supplied classes

January 24-26, Alice Tier 1-2 Workshop On the horizon Extended monitoring Disk caching proxy server Integrated 3 rd party copy New client Dropping RH4 support Planned for 3.2 release

January 24-26, Alice Tier 1-2 Workshop Extended Monitoring Redirect information will be provided Selectable via configuration option Will provide extensive information Who went where for what operations Currently, only available via debug log output (yech) Authentication details Number of attempts, failures, bypasses, etc

January 24-26, Alice Tier 1-2 Workshop Disk Caching Proxy Server The current proxy server will be extended Will allow for memory as well as disk caching Data can stick around on the proxy for re-use We will be fleshing out the details in February Data Server Client Proxy ServerMemoryDisk Local Area Network Wide Area Network

January 24-26, Alice Tier 1-2 Workshop Integrated 3 rd Party Copy Currently, xrd3cp provides 3 rd party copy We plan to include this functionality in the base Actual protocol will differ since pull is a simpler model This does not change xrootd protocol just the ofs plugin Part of the xrdcp rewrite Better handling of streams More understandable options

January 24-26, Alice Tier 1-2 Workshop New Client Current client uses a dedicated thread model Limits scaling and is resource intensive New client will use a thread pool model Scalable and fully asynchronous Will be the platform for future features E.G. plug-in caches, local redirects, etc

January 24-26, Alice Tier 1-2 Workshop Things in the near future Extended POSC Specialized meta-manager Integrated alerts New async I/O model

January 24-26, Alice Tier 1-2 Workshop Extended POSC Currently, adding “?ofs.posc=1” enables POSC Persist On Successful Close This can be extended to support checksums E.G. “?ofs.posc=adler32:csval” File persists on successful close AND Supplied checksum matches Privilege & error ending states not yet defined

January 24-26, Alice Tier 1-2 Workshop Specialized Meta-Manager Current MM is a regular manager with mm role This limits what the meta-manager can do Extending it unduly impacts the manager’s code The specialized MM is a separate daemon Will allow many more subscribers Limit now is 64 will likely increase it to 256. Can better optimize handling federated managers

January 24-26, Alice Tier 1-2 Workshop Integrated Alerts Currently, alerts based on using monitoring Monitoring provides broad usage information Alerts are therefore macro-scale We want to send a separate alert stream Based on unusual internal events E.G. unexpected latency, server recovery actions, etc Part of message and logging restructuring

January 24-26, Alice Tier 1-2 Workshop New Async I/O Model xrootd Currently xrootd uses OS supplied async I/O This is not particularly useful In Linux it is simulated as it was never kernel level In other OS’s it uses a lot of CPU resources xrootd In fact, xrootd normally bypasses it for most requests The next version will use a thread model Based on looking ahead on the request stream This should be more applicable to most requests

January 24-26, Alice Tier 1-2 Workshop IPV6 Currently, all new code supports IPV6 But existing code needs to change We are not sure how critical this really is And it has side-effects E.G. all IP addresses in messages would change No implementation plan yet How critical is this in practice?

January 24-26, Alice Tier 1-2 Workshop Did You Know? You can do load based server selection Avoid overloaded servers Step 1: Get a load collector program We supply one called cms_MonPerf Step 2: Configure its use cms.perf [ int time ] [ pgm prog ] Step 3: Configure how to calculate load cms.sched parameters

January 24-26, Alice Tier 1-2 Workshop The cms.sched Directive I cms.sched parameters cpu pct io pct mem pct pag pct runq pct space pct Percentage on indicated item used to calculate load Total percentages should equal 100!

January 24-26, Alice Tier 1-2 Workshop The cms.sched Directive II cms.sched parameters fuzz pct Percentage difference two load value must have in order to be considered different Default is 0 maxload value Maximum load allowed to select server If it’s the only server that has the file and it can’t be selected the client gets an ENOENT condition This can be used to replicate the file or select an external source of the file

January 24-26, Alice Tier 1-2 Workshop Xrootd Collaboration Mutually interested institutions contributing effort for development and maintenance SLAC (founder) CERN (2010) JINR (fall 2011) UCSD (winter 2011) Newest member!

January 24-26, Alice Tier 1-2 Workshop Conclusion xrootd xrootd is under active development Always looking for new ideas Feel free to suggest them Be a contributor You too can contribute to the code base xrootd Consider joining the xrootd collaboration It costs no money to join See more at

January 24-26, Alice Tier 1-2 Workshop Acknowledgements Current Software Contributors ATLAS: Doug Benjamin, Patrick McGuigan, Danila Oleynik, Artem Petrosyan CERN: Fabrizio Furano, Lukasz Janyst, Andreas Peters, David Smith CMS: Brian Bockelman (unl), Matevz Tadel (ucsd) Fermi/GLAST: Tony Johnson FZK: Artem Trunov BeStMan LBNL: Alex Sim, Junmin Gu, Vijaya Natarajan (BeStMan team) Root: Gerri Ganis, Beterand Bellenet, Fons Rademakers OSG: Tim Cartwright, Tanya Levshina SLAC: Andrew Hanushevsky, Wilko Kroeger, Daniel Wang, Wei Yang Operational Collaborators ANL, BNL, CERN, FZK, IN2P3, SLAC, UCSD, UTA, UoC, UNL, UVIC, UWisc US Department of Energy Contract DE-AC02-76SF00515 with Stanford University