Taming the protocol zoo

Slides:



Advertisements
Similar presentations
Jens G Jensen CCLRC/RAL hepsysman 2005Storage Middleware SRM 2.1 issues hepsysman Oxford 5 Dec 2005.
Advertisements

Wahid Bhimji SRM; FTS3; xrootd; DPM collaborations; cluster filesystems.
Data Management TEG Status Dirk Duellmann & Brian Bockelman WLCG GDB, 9. Nov 2011.
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
WLCG Operations and Tools TEG Monitoring – Experiment Perspective Simone Campana and Pepe Flix Operations TEG Workshop, 23 January 2012.
Data & Storage Management TEGs Summary of recommendations Wahid Bhimji, Brian Bockelman, Daniele Bonacorsi, Dirk Duellmann GDB, CERN 18 th April 2012.
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
LHCb input to DM and SM TEGs. Remarks to DM and SM TEGS Introduction m We have already provided some input during our dedicated session of the TEG m Here.
PhysX CoE: LHC Data-intensive workflows and data- management Wahid Bhimji, Pete Clarke, Andrew Washbrook – Edinburgh And other CoE WP4 people…
Data Management The GSM-WG Perspective. Background SRM is the Storage Resource Manager A Control protocol for Mass Storage Systems Standard protocol:
Δ Storage Middleware GridPP10 What’s new since GridPP9? CERN, June 2004.
Your university or experiment logo here Storage and Data Management - Background Jens Jensen, STFC.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
Data and Storage Evolution in Run 2 Wahid Bhimji Contributions / conversations / s with many e.g.: Brian Bockelman. Simone Campana, Philippe Charpentier,
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
MW Readiness Verification Status Andrea Manzi IT/SDC 21/01/ /01/15 2.
GLUE 2 Open Issues in Storage Information Providers 16 th May 2014.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
Your university or experiment logo here The Protocol Zoo A Site Presepective Shaun de Witt, STFC (RAL)
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
CERN IT Department CH-1211 Geneva 23 Switzerland GT HTTP solutions for data access, transfer, federation Fabrizio Furano (presenter) on.
Storage Interfaces Introduction Wahid Bhimji University of Edinburgh Based on previous discussions with Working Group: (Brian Bockelman, Simone Campana,
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
Storage Classes report GDB Oct Artem Trunov
LCG Accounting Update John Gordon, CCLRC-RAL WLCG Workshop, CERN 24/1/2007 LCG.
Storage Interfaces and Access pre-GDB Wahid Bhimji University of Edinburgh On behalf of all those who participated.
SRM v2.2 Production Deployment SRM v2.2 production deployment at CERN now underway. – One ‘endpoint’ per LHC experiment, plus a public one (as for CASTOR2).
Testing Infrastructure Wahid Bhimji Sam Skipsey Intro: what to test Existing testing frameworks A proposal.
Storage Interfaces and Access: Interim report Wahid Bhimji University of Edinburgh On behalf of WG: Brian Bockelman, Philippe Charpentier, Simone Campana,
Wahid Bhimji (Some slides are stolen from Markus Schulz’s presentation to WLCG MB on 19 June Apologies to those who have seen some of this before)
EGEE-II INFSO-RI Enabling Grids for E-sciencE Architecture of LHC File Catalog Valeria Ardizzone INFN Catania – EGEE-II NA3/NA4.
CERN IT Department CH-1211 Genève 23 Switzerland t DPM status and plans David Smith CERN, IT-DM-SGT Pre-GDB, Grid Storage Services 11 November.
HEPiX IPv6 Working Group David Kelsey (STFC-RAL) GridPP33 Ambleside 22 Aug 2014.
Riccardo Zappi INFN-CNAF SRM Breakout session. February 28, 2012 Ingredients 1. Basic ingredients (Fabric & Conn. level) 2. (Grid) Middleware ingredients.
Operations Coordination Team Maria Girone, CERN IT-ES GDB, 11 July 2012.
Accounting Review Summary and action list from the (pre)GDB Julia Andreeva CERN-IT WLCG MB 19th April
EGEE Data Management Services
Federating Data in the ALICE Experiment
dCache Paul Millar, on behalf of the dCache Team
Evolution of storage and data management
a brief summary for users
WLCG IPv6 deployment strategy
Main objectives To have by the end of August :
WLCG Network Discussion
Ricardo Rocha ( on behalf of the DPM team )
Status of the SRM 2.2 MoU extension
Future of WAN Access in ATLAS
Dynafed, DPM and EGI DPM workshop 2016 Speaker: Fabrizio Furano
Storage Interfaces and Access: Introduction
dCache – protocol developments and plans
Storage / Data TEG Introduction
Gfal/lcg-util -> Gfal2/gfal2-util
dCache Scientific Cloud
SRM Developers' Response to Enhancement Requests
Storage Protocol overview
SRM2 Migration Strategy
The ADC Operations Story
Update from the HEPiX IPv6 WG
Storage Resource Reporting Proposal
GFAL 2.0 Devresse Adrien CERN lcgutil team
EGI UMD Storage Software Repository (Mostly former EMI Software)
WLCG and support for IPv6-only CPU
DPM and SRM-less operation
dCache: new and exciting features
Data Management cluster summary
Australia Site Report Sean Crosby DPM Workshop – 13 December 2013.
WLCG Collaboration Workshop: Outlook for 2009 – 2010
Data services in gLite “s” gLite and LCG.
Presentation transcript:

Taming the protocol zoo Wahid Bhimji pre-GDB 13th January 2015

The topics for today Whether we are on track to allow non-SRM disk-only sites within WLCG in Run 2 Reducing / limiting the overall number of services needed by WLCG disk-only storage (the so-called ‘protocol zoo’)

History – non-SRM sites Recommendation from Storage/Data ‘TEGs’ (see backup) And followed up in ‘Storage interfaces (SI)’ working group Annecy pre-GDB (Oct 2012!) summarized quite well: https://indico.cern.ch/event/208241/ Not designing an replacement to SRM but bring together people and activities to communicate and guide Needed activities were defined (see next slide).. SI group became just one part of more general ‘data’ activity. Some short updates since e.g. May 2014 ‘data access’ GDB Things have progressed in that time … FTS3, Federations, gfal2, Rucio all in production; and Davix available Much development for all activities in table done .. BUT largely not pushed into production use

Annecy GDB –development areas at the time Updates in Red https://indico.cern.ch/event/155073/other-view?view=standard Needed by? Issue Solution ATLAS/ LHCb Reporting of space used in space tokens. JSON publishing currently used in some places on ATLAS – probably temporary. WebDav quotas? Use RFC 4331 (reporting). Calculating per directory also needs action. Targeting upload to space token. Could just use namespace but certain SEs would need to change the way they report space to reflect. (Or use e.g. http) SEs are enabling – but ongoing as above. ATLAS/LHCb Deletion gFal2 will help. gFal2 in production. ATLAS deletion via dav in Rucio (not tested) LHCb (ATLAS) Surl->Turl Require a redirecting protocol and SURL = Turl for sites that want no SRM. See LHCb update Any? Checksum check – confirm not needed? Some service query is needed by ATLAS – as is some “srm-ls”. gFal2 will help (OK for gridftp) All? pure gridFTP on different storage types DPM at least willing to look at this. Enabled for DPM (not used in prod.). dCache had it already

Protocol Zoo – History and problem Linked somewhat with SRM item - sites currently still asked for SRM as well as new protocols. E.g DPM/ATLAS site currently needs SRM, gridftp, xrootd, rfio, WebDav..  Also increase interest in (for example) Ceph – and don’t want to develop (too many) plugins for different wlcg specific protocols. That’s a site perspective but also load on storage system providers and locks us in to storage systems with wlcg-specifc effort. Also Experiments – e.g. ATLAS has expressed (in Dec Jamboree – see slides later) that zoo of local access protocols is something they could do without Note this is a zoo across sites rather than within so its a slightly different issue. Personally: I think it would be great (and feasible) if WLCH could use sites that only had 1 (e.g. http/dav). But for ‘normal’ WLCG sites, 2-3 not unreasonable. Short-term minimum is gridftp (FTS), xrootd (federation and data access), and, for those experiments who need it, Dav or SRM (metadata ops: space reporting , deletion). Xrootd-only feasible if metadata not needed…

This meeting Experiments – progress towards non-SRM (if applicable) and protocol requirements. Sites – protocol opinions – RAL as example but also open forum for comments. Storage providers – current protocol development plus roadmap and opinions. Discussion -> Action Plan -> WLCG Ops

BACKUP SLIDES - History

TEG recommendation reminder  Maintain SRM at archive sites  Experiments, middleware experts and sites should agree on alternatives to be considered for testing and deployment, targeting not the full SRM functionality but the subset detailed here, determined by its actual usage. … recommend a small working group be formed,reporting to the GDB, to evaluate alternatives as they emerge Develop future interfaces: … different approaches to integrate cloud-based storage resources need to be investigated …

Table of used functions from TEG Somewhat simplified and removed those only relevant for Archive/T1 A couple of observations: Not that much is needed – e.g. space management is only querying and not even that for CMS

Brief functionality table from Annecy pre-GDB: (see also LHCb talks) Used by ATLAS CMS LHCb Is there an existing Alternative or Issue (to SRM) Transfer: 3rd Party (FTS) YES Using just gridFTP in EOS (ATLAS) and Nebraska (CMS) What about on other SEs? Transfer: Job in/out (LAN) ATLAS and CMS using LAN protocols directly Negotiate a transport protocol NO LHCb use lcg-getturls; Transfer: Direct Download ATLAS use SRM via lcg-cp, Alternative plugins in rucio Namespace: Manipulation / Deletion ATLAS: Deletion would need plugin for an alternative Space Query YES? Development Required Space Upload Minor Development Required