Wahid Bhimji (Some slides are stolen from Markus Schulz’s presentation to WLCG MB on 19 June Apologies to those who have seen some of this before)

Slides:



Advertisements
Similar presentations
Storage Workshop Summary Wahid Bhimji University Of Edinburgh On behalf all of the participants…
Advertisements

Wahid Bhimji SRM; FTS3; xrootd; DPM collaborations; cluster filesystems.
Data Management TEG Status Dirk Duellmann & Brian Bockelman WLCG GDB, 9. Nov 2011.
Operations Coordination Team Maria Girone, CERN IT-ES Kick-off meeting 24 th September 2012.
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
WLCG Operations and Tools TEG Monitoring – Experiment Perspective Simone Campana and Pepe Flix Operations TEG Workshop, 23 January 2012.
Data & Storage Management TEGs Summary of recommendations Wahid Bhimji, Brian Bockelman, Daniele Bonacorsi, Dirk Duellmann GDB, CERN 18 th April 2012.
WebFTS as a first WLCG/HEP FIM pilot
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
SRM 2.2: status of the implementations and GSSD 6 th March 2007 Flavia Donno, Maarten Litmaath INFN and IT/GD, CERN.
PhysX CoE: LHC Data-intensive workflows and data- management Wahid Bhimji, Pete Clarke, Andrew Washbrook – Edinburgh And other CoE WP4 people…
Storage Wahid Bhimji DPM Collaboration : Tasks. Xrootd: Status; Using for Tier2 reading from “Tier3”; Server data mining.
Your university or experiment logo here Storage and Data Management - Background Jens Jensen, STFC.
Data and Storage Evolution in Run 2 Wahid Bhimji Contributions / conversations / s with many e.g.: Brian Bockelman. Simone Campana, Philippe Charpentier,
MW Readiness Verification Status Andrea Manzi IT/SDC 21/01/ /01/15 2.
CERN IT Department CH-1211 Geneva 23 Switzerland GT WG on Storage Federations First introduction Fabrizio Furano
OSG Technology Area Brian Bockelman Area Coordinator’s Meeting February 15, 2012.
MW Readiness WG Update Andrea Manzi Maria Dimou Lionel Cons 10/12/2014.
INFSO-RI Enabling Grids for E-sciencE Enabling Grids for E-sciencE Pre-GDB Storage Classes summary of discussions Flavia Donno Pre-GDB.
WLCG Grid Deployment Board, CERN 11 June 2008 Storage Update Flavia Donno CERN/IT.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
Your university or experiment logo here The Protocol Zoo A Site Presepective Shaun de Witt, STFC (RAL)
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
ATLAS XRootd Demonstrator Doug Benjamin Duke University On behalf of ATLAS.
Report from the WLCG Operations and Tools TEG Maria Girone / CERN & Jeff Templon / NIKHEF WLCG Workshop, 19 th May 2012.
Evolution of storage and data management Ian Bird GDB: 12 th May 2010.
Jan 2010 OSG Update Grid Deployment Board, Feb 10 th 2010 Now having daily attendance at the WLCG daily operations meeting. Helping in ensuring tickets.
Storage Interfaces Introduction Wahid Bhimji University of Edinburgh Based on previous discussions with Working Group: (Brian Bockelman, Simone Campana,
The new FTS – proposal FTS status. EMI INFSO-RI /05/ FTS /05/ /05/ Bugs fixed – Support an SE publishing more than.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Data Placement Intro Dirk Duellmann WLCG TEG Workshop Amsterdam 24. Jan 2012.
Storage Classes report GDB Oct Artem Trunov
Data management demonstrators Ian Bird; WLCG MB 18 th January 2011.
Storage Interfaces and Access pre-GDB Wahid Bhimji University of Edinburgh On behalf of all those who participated.
FTS monitoring work WLCG service reliability workshop November 2007 Alexander Uzhinskiy Andrey Nechaevskiy.
DMLite GridFTP frontend Andrey Kiryanov IT/SDC 13/12/2013.
Enabling Grids for E-sciencE INFSO-RI Enabling Grids for E-sciencE Gavin McCance GDB – 6 June 2007 FTS 2.0 deployment and testing.
SRM v2.2 Production Deployment SRM v2.2 production deployment at CERN now underway. – One ‘endpoint’ per LHC experiment, plus a public one (as for CASTOR2).
ASCC Site Report Eric Yen & Simon C. Lin Academia Sinica 20 July 2005.
1 November 17, 2005 FTS integration at CMS Stefano Belforte - INFN-Trieste CMS Computing Integration Coordinator CMS/LCG integration Task Force leader.
Testing Infrastructure Wahid Bhimji Sam Skipsey Intro: what to test Existing testing frameworks A proposal.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Storage Interfaces and Access: Interim report Wahid Bhimji University of Edinburgh On behalf of WG: Brian Bockelman, Philippe Charpentier, Simone Campana,
The Grid Storage System Deployment Working Group 6 th February 2007 Flavia Donno IT/GD, CERN.
Evolution of WLCG infrastructure Ian Bird, CERN Overview Board CERN, 30 th September 2011 Accelerating Science and Innovation Accelerating Science and.
1 DIRAC Data Management Components A.Tsaregorodtsev, CPPM, Marseille DIRAC review panel meeting, 15 November 2005, CERN.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
The HEPiX IPv6 Working Group David Kelsey (STFC-RAL) EGI OMB 19 Dec 2013.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
Outcome should be a documented strategy Not everything needs to go back to square one! – Some things work! – Some work has already been (is being) done.
VO Experiences with Open Science Grid Storage OSG Storage Forum | Wednesday September 22, 2010 (10:30am)
Dynamic Federation of Grid and Cloud Storage Fabrizio Furano, Oliver Keeble, Laurence Field Speaker: Fabrizio Furano.
CERN IT Department CH-1211 Genève 23 Switzerland t DPM status and plans David Smith CERN, IT-DM-SGT Pre-GDB, Grid Storage Services 11 November.
HEPiX IPv6 Working Group David Kelsey (STFC-RAL) GridPP33 Ambleside 22 Aug 2014.
Operations Coordination Team Maria Girone, CERN IT-ES GDB, 11 July 2012.
Ian Bird, CERN WLCG Project Leader Amsterdam, 24 th January 2012.
Federating Data in the ALICE Experiment
Evolution of storage and data management
WLCG IPv6 deployment strategy
Storage Interfaces and Access: Introduction
dCache – protocol developments and plans
Storage / Data TEG Introduction
Storage Protocol overview
Taming the protocol zoo
Final summary Ian Bird Amsterdam, DAaM 18th June 2010.
WLCG Demonstrator R.Seuster (UVic) 09 November, 2016
DCache things Paul Millar … on behalf of the dCache team.
Presentation transcript:

Wahid Bhimji (Some slides are stolen from Markus Schulz’s presentation to WLCG MB on 19 June Apologies to those who have seen some of this before)

 SRM (v2.2) API is near ubiquitous at present  There are issues with relying on SRM for management: Not all storage is manageable through SRM – particularly storage implementations outside the HEP niche do not integrate SRM. Not all Storage Element implementations provide complete implementation of SRM spec. Very little of SRM v2.2 spec. is used. Performance concerns – e.g. overhead for transfers of small files.  The TEG identified a subset of functionality of SRM that is required for the LHC experiments and what of that functionality was provided by alternatives. For disk-only systems alternatives are feasible... “Allowing” other interfaces could add to simplification, flexibility or performance.  But this is not about pushing sites, developers or experiments to abandon SRM; nor about designing a top down alternative  Where it works well and people are happy – that’s fine.

 Maintain SRM at archive sites  Experiments, middleware experts and sites should agree on alternatives to be considered for testing and deployment, targeting not the full SRM functionality but the subset detailed here, determined by its actual usage. … recommend a small working group be formed, reporting to the GDB, to evaluate alternatives as they emerge…  Develop future interfaces: … different approaches to integrate cloud-based storage resources need to be investigated …

 Based on the TEG report specify for disk only (and archive and hierarchical systems)* the minimal required functionality based on the work in the Storage TEG.  Evaluate alternatives as they emerge, call for the need of tests whenever interesting, and recommend those shown to be interoperable, scalable and supportable. **  Ensure that these alternatives can be supported by FTS and lcg_utils to allow interoperability ** (I prefer not to cover archive sites – see later slides) ** Text as in TEG Report

As presented to MB:  The WG should not attempt to design a replacement interface to SRM  The WG should bring together and coordinate existing and emerging activities such as the support of “gridFTP-only” sites Added by me:  Focus initially on non-archive use

Should be a small WG: Not excluding anyone but need to remain focussed pragmatic – will report to GDB for wider discussion  Experiment representation: Esp. those with experience in “non-SRM” storage usage Simone Campana for ATLAS Brian Bockelman for CMS Philippe Charpentier for LHCb  Site representation: especially those with non SRM storage  OSG sites (Brian), CERN (Dirk) … Also me..  Middleware and Tool Developers: FTS etc. Markus Xrootd (link to other WG) (Brian), WebDav (Paul Millar),

 Start: 1 st full meeting of WG in Sept. (next week hopefully) Material in this talk:  Review of activities on ATLAS / CMS from 1-1 discussions  Distillation of tables and other material from TEG report Oct pre-GDB – f2f discussion with other storage WGs,  Duration (as presented by Markus to MB): Since the WG monitors and coordinates emerging activities it isn’t a short term WG Report every 3 months on the progress of monitored activities and suggest additional evaluations and tests After 12 months propose a roadmap

Somewhat simplified and removed those only relevant for Archive/T1 Still probably can’t read it (!) but an observation: Not that much is needed – e.g. space management is only querying and not even that for CMS

FunctionAlt.?IssueNotes Transfer: 3 rd Party (FTS) ✔✗ using gridFTP only in EOS Transfer: Job in/out (lan protocol) ✔✗ using xrootd only in EOS Transfer: Download (DQ2-Get) ✗✔ Currently use lcg-cp (alt plugins in rucio - dev) Namespace: Creating Directories/Deletion ✗✔ Deletion would need plugin for an alternative.. Dev work Namespace: Checking Checksums??? Space Management: Query Use ✗✔✔ Need something Space Management: Target Upload ✗✔ Need something (could use namespace) Service: Availability and Info ✗✗ Probably Not needed And listing where there are alternatives demonstrated

 Cloud Storage Some work but early stages  Xrootd – federation Use expanding but not likely to replace FTS placement There is another WLCG WG on this: liaison important  FTS3 Prototype exists that supports gridFTP session reuse. Xrootd and http are expected in next prototype. Will test.  Possible ATLAS perspective on (issues for) the WG See advantages in SRM alternatives (e.g. reducing complexity; local access performance; dq2-get overheads) Need to ensure interoperability – not a proliferation of interfaces Ensure lcg-cp etc. support interfaces (make transition easier) Need alternative for space used querying.  Namespace on top-level dirs under /atlas/ good enough? Manpower on ATLAS development side needs to be considered

 No real blockers on moving non-SRM for non-archive FunctionAlt.?IssueNotes Transfer: 3 rd Party (FTS) ✔✗ gridFTP only in Nebraska (interested in xrootd or HTTP) Transfer: Job in/out (lan protocol) ✔✗ xrootd only in Nebraska Transfer: Download (DQ2-Get) ✗✗ Not really used on CMS Namespace: Creating Directories/Deletion ✗✗ Deletion or query not needed on CMS, site approves transfers. Space Management: Query Use ✗✗ Space Management: Target Upload ✗✗ STs not required on CMS Service: Availability and Info ✗✗ Probably Not needed

 Experience with using site with gridftp / xrootd IP Load balancing for gridftp - one giant endpoint. Integrating with FTS and phedex Some other advantages over srm - can take failing servers out easily - one host cert. Interested in trying xrootd with FTS.  Need work on xrootd for 3rd party transfers - will be following that. Interested in 3rd party transfers over HTTP.  Asking all sites to provide xrootd interface by end of year.  Don’t have same issues / need for space tokens monitoring or direct download: All transfers approved by the site. So they don’t overflow… No real direct download (lcg-cp); move output data to site then access it Spacetokens - only for sites benefit  No real blockers on incorporating non-SRM non-archive sites that provide gridFTP and xrootd interfaces

 For non-archive sites there is an opportunity for simplification of the storage management interface.  There is a need for some oversight to evaluate/incorporate any transitions: For ATLAS at least, some coordination and development needed on issues of space querying; deletion; upload to space tokens. Overseeing that FTS and lcg-utils support different interface in same way (e.g. error codes) will ease transition.