GT Dynamic Federations

Slides:



Advertisements
Similar presentations
Single Sign-On with GRID Certificates Ernest Artiaga (CERN – IT) GridPP 7 th Collaboration Meeting July 2003 July 2003.
Advertisements

GGF Toronto Spitfire A Relational DB Service for the Grid Peter Z. Kunszt European DataGrid Data Management CERN Database Group.
SPRING 2011 CLOUD COMPUTING Cloud Computing San José State University Computer Architecture (CS 147) Professor Sin-Min Lee Presentation by Vladimir Serdyukov.
Data Management Kelly Clynes Caitlin Minteer. Agenda Globus Toolkit Basic Data Management Systems Overview of Data Management Data Movement Grid FTP Reliable.
M i SMob i S Mob i Store - Mobile i nternet File Storage Platform Chetna Kaur.
The Data Grid: Towards an Architecture for the Distributed Management and Analysis of Large Scientific Dataset Caitlin Minteer & Kelly Clynes.
IT-SDC : Support for Distributed Computing An HTTP federation prototype for LHCb Fabrizio Furano 1.
ArcGIS Server for Administrators
CERN IT Department CH-1211 Geneva 23 Switzerland GT WG on Storage Federations First introduction Fabrizio Furano
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
6 th dCache WS | Daniel Becker| 18 April 2012 | 1 Daniel Becker 6 th dCache workshop, Zeuthen, April 18, 2012 The HTTP Federation.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM Collaboration Motivation and proposal Oliver Keeble CERN On.
Evaluating distributed EOS installation in Russian Academic Cloud for LHC experiments A.Kiryanov 1, A.Klimentov 2, A.Zarochentsev 3. 1.Petersburg Nuclear.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
CERN IT Department CH-1211 Geneva 23 Switzerland GT Davix A toolkit for efficient data access with HTTP/DAV based protocols Fabrizio Furano.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
CERN IT Department CH-1211 Geneva 23 Switzerland GT HTTP solutions for data access, transfer, federation Fabrizio Furano (presenter) on.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
EMI INFSO-RI Catalogue synchronization & ACL propagation Fabrizio Furano (CERN IT-GT)
CERN IT Department CH-1211 Geneva 23 Switzerland GT WG on Storage Federations Sept 2012 Usages and Goals Summary Fabrizio Furano on behalf.
CERN IT Department CH-1211 Genève 23 Switzerland t ALICE XROOTD news New xrootd bundle release Fixes and caveats A few nice-to-know-better.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Overview of DMLite Ricardo Rocha ( on behalf of the LCGDM team.
IT-SDC : Support for Distributed Computing Dynamic Federations: scalable, high performance Grid/Cloud storage federations Fabrizio Furano - Oliver Keeble.
DMLite GridFTP frontend Andrey Kiryanov IT/SDC 13/12/2013.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
EMI is partially funded by the European Commission under Grant Agreement RI Roadmap & Future Work Ricardo Rocha ( on behalf of the DPM team )
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
Andrea Manzi CERN EGI Conference on Challenges and Solutions for Big Data Processing on cloud 24/09/2014 Storage Management Overview 1 24/09/2014.
IT-SDC : Support for Distributed Computing Dynafed FTS3 Human Brain Project use cases Fabrizio Furano Alejandro Alvarez.
1 EMI INFSO-RI Dynamic Federations Seamless aggregation of standard-protocol-based storage endpoints Fabrizio Furano Patrick Fuhrmann Paul Millar.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Protocols in DPM Ricardo Rocha.
EMI INFSO-RI Catalogue synchronization & ACL propagation Fabrizio Furano (CERN IT-GT-DMS)
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Solutions for WAN data access: xrootd and NFSv4.1 Andrea Sciabà.
EMI INFSO-RI Catalogue synchronization & ACL propagation Fabrizio Furano (CERN IT-GT)
EMI is partially funded by the European Commission under Grant Agreement RI DPM in EMI-II HTTP and NFS interfaces Oliver Keeble On behalf of DPM.
IT-SDC : Support for Distributed Computing Dynamic Federation of Grid and Cloud Storage Fabrizio Furano, Oliver Keeble, Laurence Field Speaker: Fabrizio.
EMI is partially funded by the European Commission under Grant Agreement RI Future Proof Storage with DPM Oliver Keeble (on behalf of the CERN IT-GT-DMS.
EMI INFSO-RI Catalogue synchronization & ACL propagation Fabrizio Furano (CERN IT-GT)
1 EMI INFSO-RI Dynamic Federations Seamless aggregation of standard-protocol-based storage endpoints Fabrizio Furano Patrick Fuhrmann Paul Millar.
Dynamic Federation of Grid and Cloud Storage Fabrizio Furano, Oliver Keeble, Laurence Field Speaker: Fabrizio Furano.
Implementation of GLUE 2.0 support in the EMI Data Area Elisabetta Ronchieri on behalf of JRA1’s GLUE 2.0 Working Group INFN-CNAF 13 April 2011, EGI User.
CERN IT Department CH-1211 Geneva 23 Switzerland GT Dynamic Federations Seamless aggregation of open-protocol-based storage endpoints Fabrizio.
CERN IT-Storage Strategy Outlook Alberto Pace, Luca Mascetti, Julien Leduc
Federating Data in the ALICE Experiment
Argus EMI Authorization Integration
Jean-Philippe Baud, IT-GD, CERN November 2007
CernVM-FS vs Dataset Sharing
File Syncing Technology Advancement in Seafile -- Drive Client and Real-time Backup Server Johnathan Xu CTO, Seafile Ltd.
Dynamic Storage Federation based on open protocols
Ricardo Rocha ( on behalf of the DPM team )
Introduction to Distributed Platforms
Vincenzo Spinoso EGI.eu/INFN
(on behalf of the POOL team)
Large Scale Parallel Print Service
Dynafed, DPM and EGI DPM workshop 2016 Speaker: Fabrizio Furano
Open Source distributed document DB for an enterprise
StoRM Architecture and Daemons
Introduction to Data Management in EGI
Grid Computing.
Dirk Düllmann CERN Openlab storage workshop 17th March 2003
Processes The most important processes used in Web-based systems and their internal organization.
GFAL 2.0 Devresse Adrien CERN lcgutil team
EGI UMD Storage Software Repository (Mostly former EMI Software)
Ákos Frohner EGEE'08 September 2008
DCache things Paul Millar … on behalf of the dCache team.
Cloud Computing.
Presentation transcript:

GT Dynamic Federations A demo with dCache, DPM and a cloud storage provider Fabrizio Furano (Did the work) Patrick Fuhrmann (presenter) Paul Millar Daniel Becker Adrien Devresse (did the work) Oliver Keeble Ricardo Brito da Rocha Alejandro Alvarez Credits to ShuTing Liao (ASGC) Partially funded by and CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it EMI is partially funded by the European Commission under Grant Agreement INFSO-RI-261611

Storage Federations: Motivations GT Currently data lives on islands of storage catalogues are the maps FTS/gridFTP are the delivery companies Experiment frameworks populate the island Jobs are directed to places where the needed data is or should be ...... Almost all data lives on more than one island Assumption: perfect storage ( unlikely to impossible) perfect experiment workflow and catalogues ( unlikely ) Strict locality has some limitations a single missing file can derail the whole job or series of jobs Failover to data on another island could help  Federation CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it 23 Apr 2012 - Dynamic Federations

GT Storage federations What has to be done? How should this be done? Make different storage clusters be seen as one Make global file-based data access seamless How should this be done? No strange APIs, everything looks “banal” Use dynamic systems that are easy to setup/maintain: no complex metadata persistency no DB babysitting (keep it for the experiment’s metadata) no central catalogue inconsistencies, by design Use systems that exhibit high performance Goal is hundreds of Ks interactions per second (peak) Try to optimize by using proximity etc. CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it 23 Apr 2012 - Dynamic Federations

Vision: Everything Everywhere GT Transparent access to remote data should feel natural This includes interactive browsing of files and directories Seamless storage federations of: Official Storage Elements, LFCs, catalogues... Cached data (i.e. SQUID-like things, not registered in any catalogue) HTTP/DAV-based servers Cloud storage services HTTP-enabled XROOTD/EOS clusters, sharing the data. Base everything on open ‘just works’ technologies Local SE as a preference, give the freedom to point to an efficient and reliable global federation Optimize redirections based on on-the-fly client-data proximity Avoid inconsistencies, just looking at where the files are now. Limit complexity: read only Usually writes happen to well-known, close islands CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it 23 Apr 2012 - Dynamic Federations

GT Dynamic Federations We federate (meta)data repositories that are ‘compatible’ Name space (modulo simple prefixes) Permissions (they don’t contradict across sites) Content (same key or filename means same file) Dynamically and transparently discovering metadata looks like a unique, very fast file metadata system properly presenting the aggregated metadata views redirecting clients to the geographically closest endpoint As a plus: can be used by client tools that everybody knows focus on HTTP/DAV, we can use it by ROOT via Browswer, OS’es to smartphones CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it 23 Apr 2012 - Dynamic Federations

GT Dynamic Federations Technically TODAY we can aggregate: dCache DAV/HTTP instances DPM DAV/HTTP instances LFC DAV/HTTP instances Cloud DAV/HTTP services Native LFC and DPM databases (through DMLite used as a client) Can be extended to other sources The system also can load a “Geo” plugin Gives a geographical location to replicas and clients Allows the core to choose the replica that is closer to the client The one that’s available uses GeoIP (free) CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it 23 Apr 2012 - Dynamic Federations

GT Demo The demo federates three storages via HTTP/DAV: a DPM instance at ASGC (Taiwan) a dCache instance in DESY a Cloud storage account by Deutsche Telecom The feeling it gives is surprising performance is in avg higher than contacting the endpoints I put one test file in 3 sites, i.e. 3 replicas. /myfed/atlas/fabrizio/hand-shake.JPG Clients in EU get the one from DESY/DT Clients in Asia get the one from ASGC CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it 23 Apr 2012 - Dynamic Federations

/dir1 /dir1/file1 /dir1/file2 /dir1/file3 The basic idea Grid Technology /dir1 /dir1/file1 /dir1/file2 /dir1/file3 Aggregation We see this All the metadata interactions are hidden NO persistency needed here, just efficiency and parallelism With 2 replicas /dir1/file1 /dir1/file2 Storage/MD endpoint 1 /dir1/file2 /dir1/file3 Storage/MD endpoint 2 23 Apr 2012 - Dynamic Federations

GT Dynamic Federations Technically “loosely coupled storage systems” Idea: a single entry point for a federation of endpoints “lonely” storage clusters (e.g. dCache, DPM, plain HTTP servers) site/VO catalogues (e.g. LFCs) pointing to storage elements This entry point knows its endpoints, can redirect clients to them, it can present their metadata to clients Many interesting possibilities Federate third party outsourced HTTP/DAV servers (also clouds) Federate the content of SQUID caches Federate them together with the information of some experiment’s DB When requesting a file we would get it from an endpoint that is good for us, it could be a cache or a non-cache one. See as one experiment’s DBs (e.g. two LFCs), also considering what’s in the SQUID caches worldwide Transparent, direct access to the official replicas AND the cached ones as well CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it 23 Apr 2012 - Dynamic Federations

Dynamic Federations Grid Technology The endpoints are a federation, hence they are homogeneous Same access protocol (e.g. HTTP/DAV) Same name space (not necessarily same content!) The same file/replica has the same (or compatible) path/name They grant access to the same groups of users This entry point learns dynamically, automatically about their metadata content As clients contact it to get access to files It can ask the endpoints for information on the fly This entry point redirects each client to the proper endpoint Ev. applying some smart criteria, e.g. geographical proximity In principle it would work for any data access protocol that works over WAN supports redirections The system core is fully protocol-agnostic Our focus is towards HTTP/DAV DPM and dCache are releasing support for it (StoRM is on it’s way) Work in progress, priority is read access As, in general, write access is done in the local site 23 Apr 2012 - Dynamic Federations

Example Client LFC or DB SE LFC SE Grid Technology Aggregator (UGR) Plugin DMLite Frontend (Apache2+DMLite) Plugin DAV/HTTP Plugin HTTP LFC or DB SE LFC SE Plain DAV/HTTP Plain DAV/HTTP 23 Apr 2012 - Dynamic Federations

GT Why HTTP/DAV? It’s there, whatever platform we consider A very widely adopted technology We (humans) like browsers, they give an experience of simplicity Goes towards convergence Users can use their devices to access their data easily, out of the box Jobs just go straight to the data CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it 23 Apr 2012 - Dynamic Federations

GT System design A system that only works is not sufficient To be usable, it must privilege speed, parallelism, scalability The core component is a plugin-based component called originally “Uniform Generic Redirector” (Ugr) Can plug into an Apache server thanks to the DMLITE and DAV- DMLITE modules (by IT-GT) Composes on the fly the aggregated metadata views by managing parallel tasks of information location Never stacks up latencies! Able to redirect clients to replicas By construction, the responses are a data structure that models a partial, volatile namespace Keep them in an LRU fashion and we have a fast 1st level namespace cache Peak performance is ~500K->1M hits/second per core by now CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it 23 Apr 2012 - Dynamic Federations

Focus: performance GT Performance and scalability have primary importance Otherwise it’s useless... Full parallelism No limit to the number of outstanding clients/tasks No global locks/serializations! The endpoints are treated in a completely independent way Thread pools, prod/consumer queues used extensively (e.g. to stat N items in M endpoints while X clients wait for some items) Aggressive metadata caching A relaxed, hash-based, in-memory partial name space Juggles info in order to always contain what’s needed Stalls clients the minimum time that is necessary to juggle their information bits Peak perf per CPU core: 0.5~1M stats/sec High performance DAV client implementation (DAVIX) Loaded by the core as a “location” plugin Uses libneon w/ sessions caching Compound list/stat operations CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it 23 Apr 2012 - Dynamic Federations

Dynamic Feds vs XROOTD feds GT XROOTD federations are focused on the “redirection” concept Very light at the meta-manager, just redirect clients away as soon as possible If not possible, the penalty is 5 seconds per jump Global listing is implemented in the client, slow, a bit more difficult, less robust Some details do not match yet very well with geography-aware redirections Dynamic Federations support both the “redirection” concept and the “browsing” concept by design Cache metadata for the clients, in-memory Designed for scalability, performance and features Extendable plugin architecture, geography-aware redirection Can speak any protocol, our focus in on http-based things CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it 23 Apr 2012 - Dynamic Federations

Next steps GT Implement the missing bits of security, without hurting scalability and performance. More massive tests, with many endpoints, possibly distant Precise performance measurements Handle gracefully the ‘death’ of the endpoints Understanding the impact of the production workloads Does the machinery need a 2nd level of caching? Which characteristics? Immediate sensing of changes in the endpoints’ content, e.g. add, delete SEMsg in EMI2 already has some useful notifications (default off) Some more practical experience (getting used to the idea, using SQUIDs, CVMFS, EOS, clouds,... <put your item here>* ) Power users helping in getting the best out of the system CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it 23 Apr 2012 - Dynamic Federations

Conclusions GT Dynamic Federations: an efficient, persistency- free, easily manageable approach to federate remote storage endpoints Usable for fast changing caches and clouds Gives ways to solve some nasty Data Management problems Peak performance is very high: O(105) hits/s Opens our Data world to a large variety of already available clients, by using standard protocols. Work in progress, first milestone was now. CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it 23 Apr 2012 - Dynamic Federations

GT Thank you Questions? CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it Partially funded by EMI is partially funded by the European Commission under Grant Agreement INFSO-RI-261611