IT-SDC : Support for Distributed Computing An HTTP federation prototype for LHCb Fabrizio Furano 1.

Slides:



Advertisements
Similar presentations
Data Management Expert Panel - WP2. WP2 Overview.
Advertisements

1 User Analysis Workgroup Update  All four experiments gave input by mid December  ALICE by document and links  Very independent.
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
Embedded Web Hyung-min Koo. 2 Table of Contents Introduction of Embedded Web Introduction of Embedded Web Advantages of Embedded Web Advantages of Embedded.
Lecture 2 Page 1 CS 236, Spring 2008 Security Principles and Policies CS 236 On-Line MS Program Networks and Systems Security Peter Reiher Spring, 2008.
Copyright 2004 Monash University IMS5401 Web-based Systems Development Topic 2: Elements of the Web (g) Interactivity.
1 CS 502: Computing Methods for Digital Libraries Lecture 22 Web browsers.
Oxford Jan 2005 RAL Computing 1 RAL Computing Implementing the computing model: SAM and the Grid Nick West.
SM3121 Software Technology Mark Green School of Creative Media.
Christopher M. Pascucci Basic Structural Concepts of.NET Browser – Server Interaction.
Bigtable: A Distributed Storage System for Structured Data F. Chang, J. Dean, S. Ghemawat, W.C. Hsieh, D.A. Wallach M. Burrows, T. Chandra, A. Fikes, R.E.
Working with domains and Active Directory
M i SMob i S Mob i Store - Mobile i nternet File Storage Platform Chetna Kaur.
Data management at T3s Hironori Ito Brookhaven National Laboratory.
Overcast: Reliable Multicasting with an Overlay Network CS294 Paul Burstein 9/15/2003.
PhysX CoE: LHC Data-intensive workflows and data- management Wahid Bhimji, Pete Clarke, Andrew Washbrook – Edinburgh And other CoE WP4 people…
Algoval: Evaluation Server Past, Present and Future Simon Lucas Computer Science Dept Essex University 25 January, 2002.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
OSG Storage Architectures Tuesday Afternoon Brian Bockelman, OSG Staff University of Nebraska-Lincoln.
Δ Storage Middleware GridPP10 What’s new since GridPP9? CERN, June 2004.
Freelib: A Self-sustainable Digital Library for Education Community Ashraf Amrou, Kurt Maly, Mohammad Zubair Computer Science Dept., Old Dominion University.
Developer TECH REFRESH 15 Junho 2015 #pttechrefres h Understand your end-users and your app with Application Insights.
CERN IT Department CH-1211 Geneva 23 Switzerland GT WG on Storage Federations First introduction Fabrizio Furano
GLite – An Outsider’s View Stephen Burke RAL. January 31 st 2005gLite overview Introduction A personal view of the current situation –Asked to be provocative!
INTRODUCTION TO WEB APPLICATION Chapter 1. In this chapter, you will learn about:  The evolution of the Internet  The beginning of the World Wide Web,
Parameter Study Principles & Practices. What is Parameter Study? Parameter study is the application of a single algorithm over a set of independent inputs:
A Simple Introduction to Git: a distributed version-control system CS 5010 Program Design Paradigms “Bootcamp” Lesson 0.5 © Mitchell Wand, This.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
Company small business cloud solution Client UNIVERSITY OF BEDFORDSHIRE.
Digital Literacy Concepts and basic vocabulary. Digital Literacy Knowledge, skills, and behaviors used in digital devices (computers, tablets, smartphones)
PanDA Status Report Kaushik De Univ. of Texas at Arlington ANSE Meeting, Nashville May 13, 2014.
CERN IT Department CH-1211 Geneva 23 Switzerland GT HTTP solutions for data access, transfer, federation Fabrizio Furano (presenter) on.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS XROOTD news New release New features.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Overview of DMLite Ricardo Rocha ( on behalf of the LCGDM team.
IT-SDC : Support for Distributed Computing Dynamic Federations: scalable, high performance Grid/Cloud storage federations Fabrizio Furano - Oliver Keeble.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
1 DIRAC Data Management Components A.Tsaregorodtsev, CPPM, Marseille DIRAC review panel meeting, 15 November 2005, CERN.
Andrea Manzi CERN EGI Conference on Challenges and Solutions for Big Data Processing on cloud 24/09/2014 Storage Management Overview 1 24/09/2014.
IT-SDC : Support for Distributed Computing Dynafed FTS3 Human Brain Project use cases Fabrizio Furano Alejandro Alvarez.
EGEE is a project funded by the European Union under contract IST Issues from current Experience SA1 Feedback to JRA1 A. Pacheco PIC Barcelona.
1 UNIT 13 The World Wide Web. Introduction 2 Agenda The World Wide Web Search Engines Video Streaming 3.
1 EMI INFSO-RI Dynamic Federations Seamless aggregation of standard-protocol-based storage endpoints Fabrizio Furano Patrick Fuhrmann Paul Millar.
DDN Web Object Scalar for Big Data Management Shaun de Witt, Roger Downing (STFC) Glenn Wright (DDN)
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
© CGI Group Inc. User Guide Subversion client TortoiseSVN.
IT-SDC : Support for Distributed Computing Dynamic Federation of Grid and Cloud Storage Fabrizio Furano, Oliver Keeble, Laurence Field Speaker: Fabrizio.
EMI is partially funded by the European Commission under Grant Agreement RI Future Proof Storage with DPM Oliver Keeble (on behalf of the CERN IT-GT-DMS.
1 EMI INFSO-RI Dynamic Federations Seamless aggregation of standard-protocol-based storage endpoints Fabrizio Furano Patrick Fuhrmann Paul Millar.
Dynamic Federation of Grid and Cloud Storage Fabrizio Furano, Oliver Keeble, Laurence Field Speaker: Fabrizio Furano.
Federating Data in the ALICE Experiment
Dynamic Storage Federation based on open protocols
L’analisi in LHCb Angelo Carbone INFN Bologna
Ricardo Rocha ( on behalf of the DPM team )
ATLAS Grid Information System
Ideas for a HEP computing model
Vincenzo Spinoso EGI.eu/INFN
Future of WAN Access in ATLAS
dCache “Intro” a layperson perspective Frank Würthwein UCSD
Dynafed, DPM and EGI DPM workshop 2016 Speaker: Fabrizio Furano
GT Dynamic Federations
Introduction to Data Management in EGI
Taming the protocol zoo
GFAL 2.0 Devresse Adrien CERN lcgutil team
DCache things Paul Millar … on behalf of the dCache team.
Chapter 27 WWW and HTTP.
Presentation transcript:

IT-SDC : Support for Distributed Computing An HTTP federation prototype for LHCb Fabrizio Furano 1

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC Introduction  In September we started setting up an HTTP fed for LHCb  Stefan Roiser  Fabrizio Furano  Very good results in a short time  We present here the challenges, the results and the status of the prototype 2

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC HTTP/WEBDAV federation  The HTTP/WebDAV LHCb prototype fed for an user appears as just a huge, distributed repository with a friendly feel  is accessible from a browser or with a decent HTTP client (curl, wget, davix, …)  works quickly and reliably  takes realtime redirection choices, considering the worldwide status (instead of a static catalogue)  never out of sync with the storage elements’ content  can scale up the size of the repo  can scale up the number of clients  A huge data repository accessible with a browser, fast and always exact  Exact means “taking into account the status of the endpoints in that moment”  It means that the endpoints that are down are not shown 3

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC Dynamic Federations  A project started a few years ago  Goal: a frontend that presents what a certain number of endpoints would present together  Without indexing them beforehand  These endpoints can be a very broad range of objects that act as data or metadata stores  We prefer to use HTTP/WebDAV things, yet that’s not a constraint 4

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 5 Aggregation /dir1 /dir1/file1 /dir1/file2 /dir1/file3.../dir1/file1.../dir1/file2 Storage/MD endpoint 1.../dir1/file2.../dir1/file3 Storage/MD endpoint 2 This is What we want to see as users Sites remain independent and participate to a global view All the metadata interactions are hidden and done on the fly NO metadata Persistency needed here, just efficiency and parallelism With 2 replicas

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC Dynamic Federations  Opens to a multitude of use cases, by composing a worldwide system from macro building blocks speaking HTTP and/or WebDAV  Federate natively all the LHCb storage elements  Add third party outsourced HTTP/DAV servers  Add the content of fast changing things, like file caches  Add native S3 storage backends (a supported dialect)  Accommodate whatever metadata sources, even two or more remote catalogues at the same time  Clients are redirected to the replica closer to them  Redirect only to working endpoints  Accommodate whatever other Cloud-like storage endpoint 6

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC Why HTTP/DAV?  It’s there, whatever platform we consider  A very widely adopted technology  We (humans) like browsers, they give an experience of simplicity  Mainstream and sophisticated clients: curl, wget, Davix, …  ROOT works out of the box with HTTP access (LCG release >= 69)  Goes towards convergence  Users can use their devices to access their data easily, out of the box  Web applications development can meet Grid computing  Jobs and users just access data directly, in the same way  Can more easily be connected to commercial systems and apps 7

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC LHCb replica management  The first action was me asking Stefan to see the directory trees of a few LHCb SEs  They look the same everywhere, modulo a string prefix depending on the site  This is the simplest case that the Dynafeds can handle. My appreciation to whoever made this choice and kept it so clean.  Example: 8 /lhcb/LHCb/Collision12/BHADRONCOMPLETEEVENT.DST/ /0000/ _ _1.bhadroncompleteevent.dst remains constant, despite the prefix it may have, like: or

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC Look and feel  What we see in the browser is an HTML rendering of a listing  Everything is done on the fly  Click on a file to download it (if your client is authorized by the endpoint SE through X509)  Feed the URL of that file to any other client to download it  Click on the strange icon to get a metalink  A standard representation of the locations of a file sorted by increasing distance from the requestor  (Plugin-based, any other metric is possible)  It’s supported by multi-source download apps 9

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 10 Look and feel, like a normal list

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 11 Metalink example

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC LHCb HTTP SE harvesting  This step was performed by Stefan  Looking at BDII and SRM TURLs to harvest the LHCb SEs that had a working HTTP access  Enough for setting up the first little prototype in the machine of our DESY cooperators   Then Stefan wrote to everyone and we started keeping track of them 12

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC Status  13 sites out of 19  Missing:   Contacted and exchanging information.   Will join in Spring ‘15   CNAF working on a solution  PIC   Progress: seems to be available since today, still some config missing there  RAL-HEP (dCache) 13

IT-SDC : Support for Distributed Computing The Tech corner 14

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 15 Federator Plugin Frontend (Apache2+DMLite) Plugin SE Metadata cache The cache remembers what happened The next metadata interactions will very likely be fed by the cache The cache can be shared among federators SE

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC LHCb fed and metadata catalogues  A fed and a catalogue fulfil different use cases  A fed is dynamic: interacts with what’s available in that moment  Sites up/down, disappeared files, distance of alive sites from the client, …  A catalogue is static: it tells us what’s supposed to be there (data losses… dark data…)  Static/dynamic examples:  checking which site is supposed to have something needs a catalogue  selecting datasets for a run needs a catalogue  selecting files for a job will be more resilient with a fed providing fresh metalinks  running a job at a site will be more resilient with a fed providing fresh metalinks  downloading a file will be more resilient with a fed, and easier to do 16

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC What to do with LFC/DFC & C.  Keep them, they are useful because they keep trace of where LHCb pushed a file to and they feed the current workflow  They can be used to track file availability, comparing with the reality, manually or programmatically At the same time…  Technically, one can “mount” a catalogue anywhere in the namespace of a fed, even merging more catalogues…  The exercise will become a federation of catalogues, not of SEs  The federator will TRUST the catalogue for file listings, so the result will be less dynamic  It will be easy IF the catalogue has DAV access AND provides HTTP URLs for the replicas  LFC could also be mounted natively without WebDAV 17

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC Mounting a catalogue into a fed  Philippe asked me about putting LFC in the fed  Catalogues can be mounted, they would act as:  Static listing providers  Static providers of replica TURLs for namespaces that are not algorithmic (luckily not the LHCb case)  The dynafeds can translate on the fly SRM TURLs into HTTP, yet it’s a complex configuration  The dynafeds can check the translated static replica lists against the SEs, again it’s a complex configuration  The reliability of the fed will be linked to the reliability of the catalogue  My opinion…  So far, the LHCb federation does not need this, as everything is so clean without it  makes sense only if we just want to have an HTTP/DAV frontend to the catalogue itself… a legitimate use case to be kept separated from quick, dynamic data access 18

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC What about xrootd ?  Seems that LHCb is transitioning to using the Xrootd protocol for data access.  We see all the advantages of the direct data access approach supported by HTTP and Xrootd in all the Grid SE techs.  Many good reasons to grow an HTTP ecosystem that can happily coexist with a preexisting xrootd one  Native Xrootd4 sites can join it too, as Xrootd4 natively supports HTTP/WebDAV (tested with feds too)  A door open towards user-friendly, industry standard interfaces  A decise step towards opportunistic resource exploitation. We could federate an S3 backend today, together with the LHCb data.  In fact we already did in the /lhcb parent directory… 19

05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC 05 Nov 2014 HTTP Dynamic Federations - LHCb prototype IT-SDC Conclusion  A r/o R&D prototype that exceeded expectations  13 sites out of 19, the others are coming  Official site downtimes were always spotted on the log  Cleanness of LHCb repos helped  Please evaluate it and help us improve  This is likely to be an actor of a next evolution in large scale DM, HEP meeting the Web through proper tools  New features are coming. Smarter site detection, write support, logging, monitoring, …  High flexibility/scalability of the concept, able to deal with a broad range of endpoints  Can be made to work with WebFTS to find the “right” sources  Also endpoint prioritization is pluggable  Looking at exploiting the potential of mixing S3 storage with other techs  We are cooking a r/w prototype for BOINC 20