LHCb input to DM and SM TEGs. Remarks to DM and SM TEGS Introduction m We have already provided some input during our dedicated session of the TEG m Here.

Slides:



Advertisements
Similar presentations
Data & Storage Management TEGs Summary of recommendations Wahid Bhimji, Brian Bockelman, Daniele Bonacorsi, Dirk Duellmann GDB, CERN 18 th April 2012.
Advertisements

Data Grid Web Services Chip Watson Jie Chen, Ying Chen, Bryan Hess, Walt Akers.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
Storage TEG “emerging” observations and recommendations Wahid Bhimji With contributions from the SM editors (listed in intro)
Tier 3 Data Management, Tier 3 Rucio Caches Doug Benjamin Duke University.
PhysX CoE: LHC Data-intensive workflows and data- management Wahid Bhimji, Pete Clarke, Andrew Washbrook – Edinburgh And other CoE WP4 people…
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
SRM workshop – September’05 1 SRM: Expt Reqts Nick Brook Revisit LCG baseline services working group Priorities & timescales Use case (from LHCb)
Towards a Global Service Registry for the World-Wide LHC Computing Grid Maria ALANDES, Laurence FIELD, Alessandro DI GIROLAMO CERN IT Department CHEP 2013.
LHCb The LHCb Data Management System Philippe Charpentier CERN On behalf of the LHCb Collaboration.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
INFSO-RI Enabling Grids for E-sciencE gLite Data Management and Interoperability Peter Kunszt (JRA1 DM Cluster) 2 nd EGEE Conference,
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
1 LHCb File Transfer framework N. Brook, Ph. Charpentier, A.Tsaregorodtsev LCG Storage Management Workshop, 6 April 2005, CERN.
Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 An update on SRM Mumbai & subsequent discussion summary –
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
Managing Data DIRAC Project. Outline  Data management components  Storage Elements  File Catalogs  DIRAC conventions for user data  Data operation.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
Evolution of storage and data management Ian Bird GDB: 12 th May 2010.
David Adams ATLAS ATLAS distributed data management David Adams BNL February 22, 2005 Database working group ATLAS software workshop.
The CMS Top 5 Issues/Concerns wrt. WLCG services WLCG-MB April 3, 2007 Matthias Kasemann CERN/DESY.
Storage Interfaces Introduction Wahid Bhimji University of Edinburgh Based on previous discussions with Working Group: (Brian Bockelman, Simone Campana,
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Data Placement Intro Dirk Duellmann WLCG TEG Workshop Amsterdam 24. Jan 2012.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
Storage Classes report GDB Oct Artem Trunov
Author - Title- Date - n° 1 Partner Logo WP5 Status John Gordon Budapest September 2002.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Data Management Highlights in TSA3.3 Services for HEP Fernando Barreiro Megino,
Storage Interfaces and Access pre-GDB Wahid Bhimji University of Edinburgh On behalf of all those who participated.
Handling of T1D0 in CCRC’08 Tier-0 data handling Tier-1 data handling Experiment data handling Reprocessing Recalling files from tape Tier-0 data handling,
EGEE is a project funded by the European Union under contract IST Enabling bioinformatics applications to.
DMLite GridFTP frontend Andrey Kiryanov IT/SDC 13/12/2013.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
Wahid Bhimji (Some slides are stolen from Markus Schulz’s presentation to WLCG MB on 19 June Apologies to those who have seen some of this before)
1 DIRAC Data Management Components A.Tsaregorodtsev, CPPM, Marseille DIRAC review panel meeting, 15 November 2005, CERN.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
LHCb Computing activities Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
Riccardo Zappi INFN-CNAF SRM Breakout session. February 28, 2012 Ingredients 1. Basic ingredients (Fabric & Conn. level) 2. (Grid) Middleware ingredients.
EGEE Data Management Services
Federating Data in the ALICE Experiment
Evolution of storage and data management
Jean-Philippe Baud, IT-GD, CERN November 2007
WLCG IPv6 deployment strategy
Ricardo Rocha ( on behalf of the DPM team )
ATLAS Use and Experience of FTS
StoRM: a SRM solution for disk based storage systems
Vincenzo Spinoso EGI.eu/INFN
Future of WAN Access in ATLAS
Gfal/lcg-util -> Gfal2/gfal2-util
Introduction to Data Management in EGI
SRM Developers' Response to Enhancement Requests
Taming the protocol zoo
Final summary Ian Bird Amsterdam, DAaM 18th June 2010.
SRM2 Migration Strategy
Evolution of the distributed computing model The case of CMS
GFAL 2.0 Devresse Adrien CERN lcgutil team
EGI UMD Storage Software Repository (Mostly former EMI Software)
Ákos Frohner EGEE'08 September 2008
Data services in gLite “s” gLite and LCG.
Architecture of the gLite Data Management System
DIRAC Data Management: consistency, integrity and coherence of data
INFNGRID Workshop – Bari, Italy, October 2004
Presentation transcript:

LHCb input to DM and SM TEGs

Remarks to DM and SM TEGS Introduction m We have already provided some input during our dedicated session of the TEG m Here are a list of questions we want to address for each session o Not exhaustive, limited to one slide per session m It would have been useful to hear what are the proposals for evolution of SM and DM as well o These should be proposals, not diktat o Extensive discussions are needed with users and sites before embarking on full scale “prototypes” that no longer are prototypes m Whatever the future is, WLCG must ensure the long term support (no external “firm”) o Do not underestimate the amount of work needed for users to adapt o Therefore plan well ahead, gather functionality requirements… F2F Data and Storage Management TEG, Amsterdam2

Remarks to DM and SM TEGS F2F Data and Storage Management TEG, Amsterdam3 Data Placement and Federation m Does it imply replica catalogs are no longer needed? o How is brokering of jobs done? o Random or “where most files are” m When is data transfer performed? o By the WMS: implies a priori brokering, incompatible with pilot jobs o By the job: inefficiency of slots o Is it a cache (life time?) or just a download facility? m What is the advantage compared to placement using popularity? o Limited number of “master replicas” (e.g. 2) o Add replicas when popularity increases o Remove replicas when popularity decreases o … but still with a catalog and job brokering m What is the minimal number of sites for which it becomes worth it?

Remarks to DM and SM TEGS WAN protocols and FTS m We need third party transfer! o http 3 rd party transfer? OK if commercial, why support it ourselves? m We need a transfer “batch” system! o Asynchronous bulk files transfer m Whatever reliable and efficient underlying protocol is used is just fine… m There is a need to define the service class where the file has to be put (or one service class per SE) m What about the dedicated network (OPN)? o Requires a service for using it? o Not all bells and whistles may be necessary m The real point is for a user (experiment): o Transfer this list of LFNs to this SE (SE = storage class at site) P The actual physical source is irrelevant P The TS should discover whether there is an online replica, if not it should bring it online before making the transfer P Ideally it (or the SE) should register the new replica (keep consistency) m All this was already said in… Mumbai (February 2006)! m FTS 3 was looking promising… why is it dead? F2F Data and Storage Management TEG, Amsterdam4

Remarks to DM and SM TEGS Management of Catalogues and Namespaces m See Data placement… m Do we need a replica catalog? o LHCb answer is YES: we want to be able to do brokering of jobs o May only contain information on the SEs (+ file metadata + usability flags) m Do we need a catalog with URLs? o Not necessarily: the URL can be formed from the SE information and the LFN (trivial catalog), as SE information is quite static. m Do we need a single URL (used for transfers and for protocol access)? o No problem as long as the access is transparent and fast m See SRM slide for more comments… o Namespace vs storage class? F2F Data and Storage Management TEG, Amsterdam5

Remarks to DM and SM TEGS Security and Access Control m We MUST protect our data from deletion o LHCb doesn’t care about protecting from access so much m The current situation is INACCEPTABLE o Anyone with little knowledge (available on the web) can delete all files in Castor! o VOMS (or equivalent) identification and authorisation is a MUST! What about ARGUS? P Identity and role P Currently in Castor we have only 2 (uid, gid)! P Protection done by the LFC, but all backdoors are open o Backdoors should be closed (nsrm, stager_xxx active commands…) o Explicit “delete” permission would be desirable m Change of DN should be straightforward (not trivial, but OK in LFC, DPM) o Action from VO administrator F2F Data and Storage Management TEG, Amsterdam6

Remarks to DM and SM TEGS Separation of disk and tape m We need two (and only two) storage classes: o T1D0 and T0D1 o This is because no space (storage class) change is possible in some implementations of SRM m T1D0 has two functions: o Archive of T0D1 o Permanent storage of read-few data (RAW, RECO) P For this the BringOnline functionality is mandatory P We need to access the data directly from the MSS without a need for replication onto another storage P Pinning is also a must (suboptimal usage of tape drives without it) d Help to the garbage collector F2F Data and Storage Management TEG, Amsterdam7

Remarks to DM and SM TEGS Storage Interfaces: SRM and Clouds m Clouds??? m We need an interface for: o BringOnline o Pinning o Defining the storage class (unless different endpoints are used, i.e. different SEs) o Currently (Mumbai) this is done by gfal (what is its future?) m SRM is far from perfect but… o It provides the above o All efforts put into defining a standard were a miserable failure… don’t expect any other interface will be any better m … but… o We could probably wrap the minimal functionality on top of the SE native interface, if available P BringOnline and pinning not available for dCache except in SRM P Can xroot provide this functionality? P Isn’t there the danger it becomes as clumsy as SRM depending on the implementation? F2F Data and Storage Management TEG, Amsterdam8

Remarks to DM and SM TEGS Storage IO, LAN Protocols m What is wrong with POSIX file: protocol? o Very efficient in the StoRM-GPFS implementation used at CNAF o Of course abuse is could be a danger (recursive ls) but this could be taken into account in the implementation (throttling) o Almost anybody can now write a fuse plugin to make it happen, so why not use a powerful commercial protocol? m Should access protocols be more than protocols? o i.e. interact behind the scene with the MSS, discover the file location etc… o Can a tURL be just an access point? P ://diskserver.cern.ch// P … or better file: P Avoid accepting URLs like “/castor/cern.ch/…” d Needs to be fixed in application layer? No “guesses”? o Do we need different URLs for different operations? P Transfer and posix-like access F2F Data and Storage Management TEG, Amsterdam9

Remarks to DM and SM TEGS Conclusion m Whatever the future is: o Consider we need a permanently running system o No disruption of service of more than 24 hours for migration o Millions of replicas are in the current system and… it works (even when not optimal) o Any drastic change requires a lot of work on the user side (experiments framework) o Old and new systems must exist in parallel m Requirements may be different depending on the Computing Model of experiments o 7 to 12 analysis centers (LHCb) is different from 50 to 70 centers o Solutions may not be universal and complication may not be required F2F Data and Storage Management TEG, Amsterdam10