Xrootd Demonstrator Infrastructure OSG All Hands Meeting Harvard University March 7-11, 2011 Andrew Hanushevsky, SLAC

Slides:



Advertisements
Similar presentations
Potential Data Access Architectures using xrootd OSG All Hands Meeting Harvard University March 7-11, 2011 Andrew Hanushevsky, SLAC
Advertisements

Andrew Hanushevsky7-Feb Andrew Hanushevsky Stanford Linear Accelerator Center Produced under contract DE-AC03-76SF00515 between Stanford University.
Xrootd Roadmap Atlas Tier 3 Meeting University of Chicago September 12-13, 2011 Andrew Hanushevsky, SLAC
Distributed Xrootd Derek Weitzel & Brian Bockelman.
Xrootd Update OSG All Hands Meeting University of Nebraska March 19-23, 2012 Andrew Hanushevsky, SLAC
Duke and ANL ASC Tier 3 (stand alone Tier 3’s) Doug Benjamin Duke University.
Notes to the presenter. I would like to thank Jim Waldo, Jon Bostrom, and Dennis Govoni. They helped me put this presentation together for the field.
Web Servers How do our requests for resources on the Internet get handled? Can they be located anywhere? Global?
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
Scalla Back Through The Future Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 8-April-10
XRootD Roadmap To Start The Second Decade Root Workshop Saas-Fee March 11-14, 2013 Andrew Hanushevsky, SLAC
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
The Next Generation Root File Server Andrew Hanushevsky Stanford Linear Accelerator Center 27-September-2004
Scalla/xrootd Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 19-August-2009 Atlas Tier 2/3 Meeting
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
1 School of Computer, National University of Defense Technology A Profile on the Grid Data Engine (GridDaEn) Xiao Nong
Scalla/xrootd Andrew Hanushevsky, SLAC SLAC National Accelerator Laboratory Stanford University 19-May-09 ANL Tier3(g,w) Meeting.
Scalla/xrootd Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 29-October-09 ATLAS Tier 3 Meeting at ANL
Tier 3 Data Management, Tier 3 Rucio Caches Doug Benjamin Duke University.
Multi-Tiered Storage with Xrootd at ATLAS Western Tier 2 Andrew Hanushevsky Wei Yang SLAC National Accelerator Laboratory 1CHEP2012, New York
Xrootd, XrootdFS and BeStMan Wei Yang US ATALS Tier 3 meeting, ANL 1.
Scalla/xrootd Introduction Andrew Hanushevsky, SLAC SLAC National Accelerator Laboratory Stanford University 6-April-09 ATLAS Western Tier 2 User’s Forum.
Xrootd Monitoring Atlas Software Week CERN November 27 – December 3, 2010 Andrew Hanushevsky, SLAC.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Xrootd Update Andrew Hanushevsky Stanford Linear Accelerator Center 15-Feb-05
Status & Plan of the Xrootd Federation Wei Yang 13/19/12 US ATLAS Computing Facility Meeting at 2012 OSG AHM, University of Nebraska, Lincoln.
ROOT and Federated Data Stores What Features We Would Like Fons Rademakers CERN CC-IN2P3, Nov, 2011, Lyon, France.
OSG Tier 3 support Marco Mambelli - OSG Tier 3 Dan Fraser - OSG Tier 3 liaison Tanya Levshina - OSG.
ROOT-CORE Team 1 PROOF xrootd Fons Rademakers Maarten Ballantjin Marek Biskup Derek Feichtinger (ARDA) Gerri Ganis Guenter Kickinger Andreas Peters (ARDA)
02-June-2008Fabrizio Furano - Data access and Storage: new directions1.
Accelerating Debugging In A Highly Distributed Environment CHEP 2015 OIST Okinawa, Japan April 28, 2015 Andrew Hanushevsky, SLAC
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
Performance and Scalability of xrootd Andrew Hanushevsky (SLAC), Wilko Kroeger (SLAC), Bill Weeks (SLAC), Fabrizio Furano (INFN/Padova), Gerardo Ganis.
Xrootd Present & Future The Drama Continues Andrew Hanushevsky Stanford Linear Accelerator Center Stanford University HEPiX 13-October-05
Performance Tests of DPM Sites for CMS AAA Federica Fanzago on behalf of the AAA team.
Xrootd Update Alice Tier 1/2 Workshop Karlsruhe Institute of Technology (KIT) January 24-26, 2012 Andrew Hanushevsky, SLAC
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
ATLAS XRootd Demonstrator Doug Benjamin Duke University On behalf of ATLAS.
Scalla/xrootd Andrew Hanushevsky, SLAC SLAC National Accelerator Laboratory Stanford University 08-June-10 ANL Tier3 Meeting.
Scalla Advancements xrootd /cmsd (f.k.a. olbd) Fabrizio Furano CERN – IT/PSS Andrew Hanushevsky Stanford Linear Accelerator Center US Atlas Tier 2/3 Workshop.
XRootD & ROOT Considered Root Workshop Saas-Fee September 15-18, 2015 Andrew Hanushevsky, SLAC
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Scalla Authorization xrootd /cmsd Andrew Hanushevsky SLAC National Accelerator Laboratory CERN Seminar 10-November-08
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
1 Active Directory Service in Windows 2000 Li Yang SID: November 2000.
Xrootd Proxy Service Andrew Hanushevsky Heinz Stockinger Stanford Linear Accelerator Center SAG September-04
SRM Space Tokens Scalla/xrootd Andrew Hanushevsky Stanford Linear Accelerator Center Stanford University 27-May-08
Scalla As a Full-Fledged LHC Grid SE Wei Yang, SLAC Andrew Hanushevsky, SLAC Alex Sims, LBNL Fabrizio Furano, CERN SLAC National Accelerator Laboratory.
Federated Data Stores Volume, Velocity & Variety Future of Big Data Management Workshop Imperial College London June 27-28, 2013 Andrew Hanushevsky, SLAC.
1 Xrootd-SRM Andy Hanushevsky, SLAC Alex Romosan, LBNL August, 2006.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
An Analysis of Data Access Methods within WLCG Shaun de Witt, Andrew Lahiff (STFC)
Any Data, Anytime, Anywhere Dan Bradley representing the AAA Team At OSG All Hands Meeting March 2013, Indianapolis.
A System for Monitoring and Management of Computational Grids Warren Smith Computer Sciences Corporation NASA Ames Research Center.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
New Features of Xrootd SE Wei Yang US ATLAS Tier 2/Tier 3 meeting, University of Texas, Arlington,
SLACFederated Storage Workshop Summary Andrew Hanushevsky SLAC National Accelerator Laboratory April 10-11, 2014 SLAC.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Solutions for WAN data access: xrootd and NFSv4.1 Andrea Sciabà.
Federating Data in the ALICE Experiment
a brief summary for users
Global Data Access – View from the Tier 2
SLAC National Accelerator Laboratory
Data Federation with Xrootd Wei Yang US ATLAS Computing Facility meeting Southern Methodist University, Oct 11-12, 2011.
Brookhaven National Laboratory Storage service Group Hironori Ito
Presentation transcript:

xrootd Demonstrator Infrastructure OSG All Hands Meeting Harvard University March 7-11, 2011 Andrew Hanushevsky, SLAC

March 7-11, 20112OSG All Hands Meeting Goals xrootd Describe xrootd architecture configurations Show how these can be used by demos Alice (in production), Atlas, and CMS Overview of the File Residency Manager How it addresses file placement Cover recent and future developments Conclusion

March 7-11, 20113OSG All Hands Meeting The Motivation Can we access HEP data as a single repository? Treat it like a Virtual Mass Storage System Is cache-driven grid data distribution feasible? The last missing file issue (Alice production) Adaptive file placement at Tier 3’s (Atlas demo) Analysis at storage-starved sites (CMS demo) xrootd Does xrootd provide the needed infrastructure?

March 7-11, 20114OSG All Hands Meeting A Simple xrootd Cluster /my/file 3: I DO! 1: open(“/my/file”)4: Try open() at A 5: open(“/my/file”) Data Servers Manager (a.k.a. Redirector) Client cmsdxrootdcmsdxrootdcmsdxrootdcmsdxrootd Who has “/my/file”? 2: Who has “/my/file”? ABC

March 7-11, 20115OSG All Hands Meeting The Fundamentals xrootdcmsd An xrootd-cmsd pair is the building block xrootd xrootd provides the client interface Handles data and redirections cmsd xrootd cmsd manages xrootd’s (i.e. forms clusters) Monitors activity and handles file discovery The building block is uniformly stackable Can build a wide variety of configurations Much like you would do with Lego  blocks Extensive plug-ins provide adaptability

March 7-11, 20116OSG All Hands MeetingServersServersServers Federating xrootd Clusters 1: open(“/my/file”)5: Try open() at ANL Distributed Clusters Meta-Manager (a.k.a. Global Redirector) Client AB C /my/file cmsdxrootdcmsdxrootd cmsdxrootd AB C/my/file cmsdxrootdcmsdxrootd cmsdxrootd AB C /my/file cmsdxrootdcmsdxrootd cmsdxrootd ANLSLACUTA cmsdxrootd A B C Who has “/my/file”? 3: Who has “/my/file”? 8: open(“/my/file”) 4: I DO! cmsdxrootd Manager (a.k.a. Local Redirector) Manager (a.k.a. Local Redirector) Manager (a.k.a. Local Redirector) cmsdxrootd 6: open(“/my/file”) 7: Try open() at A cmsdxrootd Who has “/my/file”? 2: Who has “/my/file”? But I’m behind a firewall! Can I still play? Data is uniformly available from three distinct sites /my/file An exponentially parallel search! (i.e. O(2n))

March 7-11, 20117OSG All Hands Meeting xrootd Firewalls & xrootd xrootd xrootd is a very versatile system It can be a server, manager, or supervisor Desires are all specified in a single configuration file xrootd libXrdPss.so plug-in creates an xrootd chameleon xrootd xrootd Allows xrootd to be a client to another xrootd So, all the basic roles can run as proxies Transparently getting around fire-walls Assuming you run the proxy role on a border machine

March 7-11, 20118OSG All Hands Meeting Border Machines A Simple xrootd Proxy Cluster /my/file 6: I DO! 1: open(“/my/file”) 3: open(“/my/file”) Data Servers Manager (a.k.a. Redirector) Client cmsdxrootdcmsdxrootdcmsdxrootd ABC 4: open(“/my/file”) Firewall 7: Try open() at A 8: open(“/my/file”) Proxy Servers cmsdxrootdcmsdxrootdXY 2: Try open() at X Proxy Manager (a.k.a. Proxy Redirector) cmsdxrootd Who has “/my/file”? 5: Who has “/my/file”? cmsdxrootd Proxy Managers Can Federate With a Meta-Manager How does help in a Federated cluster?

March 7-11, 20119OSG All Hands Meeting Demonstrator Specific Features A uniform file access infrastructure Usable even in the presence of firewalls Access to files across administrative domains Each site can enforce its own rules Site participation proportional to scalability Essentially the bit-torrent social model Increased opportunities for HEP analysis A foundation for novel approaches to efficiency

March 7-11, OSG All Hands Meeting Alice & Atlas Approach Real-time placing of files at a site Built on top of the File Residency Manager (FRM) xrootd FRM - xrootd service that controls file residency Locally configured to handle events such as A requested file is missing A file is created or an existing file is modified Disk space is getting full Alice uses an “only when necessary” model Atlas will use a “when analysis demands” model

March 7-11, 2011OSG All Hands Meeting11 Using FRM For File Placement xrootd frm_xfrd Transfer Queue Configuration File Client xrootd Data Server Remote Storage all.export /atlas/atlasproddisk stage frm.xfr.copycmd in /opt/xrootd/bin/xrdcp \ –f –np root://globalredirector/$SRC $DST 1 open(missing_file) 2 Insert xfr request 3 Tell client wait 4 Read xfr request Transfer Agent 5 Launch xfr agent 7 Notify xrootd OK 6 Copy in file Wakeup client 8dq2get globus-url-copy gridFTP scp wget xrdcp etc

March 7-11, OSG All Hands Meeting FRM Even Works With Firewalls xrdcp 2 Read xfr request 3 ssh xfr agent 5 Notify xrootd to run client frm_xfrd Transfer Queue xrootd Data Server xrootdBig Bad Internet Border Machine 4 Copy in file 1 Write xfr request Need to setup ssh identity keys ● The FRM needs one or more border machines ● The server transfer agent simply launches the real agent across the border ● How it’s done frm.xfr.copycmd in noalloc ssh bordermachine /opt/xrootd/bin/xrdcp –f \ root://globalredirector/$LFN root://mynode/$LFN?ofs.posc=1

March 7-11, OSG All Hands Meeting Storage-Starved Sites (CMS) Provide direct access to missing files This is basically a freebie of the system However, latency issues exist Naively, as much as 3x increase in wall-clock time Can be as low as 5% depending on job’s CPU/IO ratio The root team is aggressively working to reduce it On the other hand... May be better than not doing analysis at such sites No analysis is essentially infinite latency

March 7-11, OSG All Hands Meeting Security xrootd xrootd supports needed security models Most notably grid certificates (GSI) Human cost needs to be considered Does read-only access require this level of security? Considering that the data is unusable without a framework Each deployment faces different issues Alice uses light-weight internal security Atlas will use server-to-server certificates CMS will need to deploy the full grid infrastructure

March 7-11, OSG All Hands Meeting Recent Developments FS-Independent Extended Attribute Framework Used to save file-specific information Migration time, residency requirements, checksums Shared-Everything File System Support Optimize file discovery in distributed file systems xrootd dCache, DPM, GPFS, HDFS, Lustre, proxy xrootd Meta-Manager throttling Configurable per-site query limits

March 7-11, OSG All Hands Meeting Future Major Developments Integrated checksums Inboard computation, storage, and reporting Outboard computation already supported Specialized Meta-Manager Allows many more subscriptions than today Internal DNS caching and full IPV6 support Automatic alerts Part of message and logging restructuring

March 7-11, OSG All Hands Meeting Conclusion xrootd xrootd mates well with demo requirements Can federated almost any file system Gives a uniform view of massive amounts of data Assuming per-experiment common logical namespace Secure and firewall friendly Ideal platform for adaptive caching systems Completely open source under a BSD license See more at

March 7-11, OSG All Hands Meeting Acknowledgements Current Software Contributors ATLAS: Doug Benjamin CERN: Fabrizio Furano, Lukasz Janyst, Andreas Peters, David Smith Fermi/GLAST: Tony Johnson FZK: Artem Trunov BeStMan LBNL: Alex Sim, Junmin Gu, Vijaya Natarajan (BeStMan team) Root: Gerri Ganis, Beterand Bellenet, Fons Rademakers OSG: Tim Cartwright, Tanya Levshina SLAC: Andrew Hanushevsky, Wilko Kroeger, Daniel Wang, Wei Yang UNL: Brian Bockelman UoC: Charles Waldman Operational Collaborators ANL, BNL, CERN, FZK, IN2P3, SLAC, UTA, UoC, UNL, UVIC, UWisc US Department of Energy Contract DE-AC02-76SF00515 with Stanford University