Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 An update on SRM Mumbai & subsequent discussion summary –http://agenda.cern.ch/fullAgenda.php?ida=a056461.

Slides:



Advertisements
Similar presentations
Jens G Jensen CCLRC/RAL hepsysman 2005Storage Middleware SRM 2.1 issues hepsysman Oxford 5 Dec 2005.
Advertisements

1 User Analysis Workgroup Update  All four experiments gave input by mid December  ALICE by document and links  Very independent.
EGEE is a project funded by the European Union under contract IST Using SRM: DPM and dCache G.Donvito,V.Spinoso INFN Bari
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
Storage Issues: the experiments’ perspective Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 9 September 2008.
– n° 1 StoRM latest performance test results Alberto Forti Otranto, Jun
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
Maarten Litmaath (CERN), EGEE User Forum, CERN, 2006/03/02 (v3) Use of the SRM interface Use case What is the SRM? –Who develops it? –Is it a standard?
Data management in grid. Comparative analysis of storage systems in WLCG.
Status report on SRM v2.2 implementations: results of first stress tests 2 th July 2007 Flavia Donno CERN, IT/GD.
SRM 2.2: status of the implementations and GSSD 6 th March 2007 Flavia Donno, Maarten Litmaath INFN and IT/GD, CERN.
LHCb input to DM and SM TEGs. Remarks to DM and SM TEGS Introduction m We have already provided some input during our dedicated session of the TEG m Here.
Write-through Cache System Policies discussion and A introduction to the system.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
Andrew C. Smith – Storage Resource Managers – 10/05/05 Functionality and Integration Storage Resource Managers.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
Maarten Litmaath (CERN), GDB meeting, CERN, 2006/02/08 VOMS deployment Extent of VOMS usage in LCG-2 –Node types gLite 3.0 Issues Conclusions.
SRM workshop – September’05 1 SRM: Expt Reqts Nick Brook Revisit LCG baseline services working group Priorities & timescales Use case (from LHCb)
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
Report on Installed Resource Capacity Flavia Donno CERN/IT-GS WLCG GDB, CERN 10 December 2008.
LHCb The LHCb Data Management System Philippe Charpentier CERN On behalf of the LHCb Collaboration.
New stager commands Details and anatomy CASTOR external operation meeting CERN - Geneva 14/06/2005 Sebastien Ponce, CERN-IT.
SRM Monitoring 12 th April 2007 Mirco Ciriello INFN-Pisa.
INFSO-RI Enabling Grids for E-sciencE Enabling Grids for E-sciencE Pre-GDB Storage Classes summary of discussions Flavia Donno Pre-GDB.
WLCG Grid Deployment Board, CERN 11 June 2008 Storage Update Flavia Donno CERN/IT.
SRM & SE Jens G Jensen WP5 ATF, December Collaborators Rutherford Appleton (ATLAS datastore) CERN (CASTOR) Fermilab Jefferson Lab Lawrence Berkeley.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
1 LHCb File Transfer framework N. Brook, Ph. Charpentier, A.Tsaregorodtsev LCG Storage Management Workshop, 6 April 2005, CERN.
Maarten Litmaath (CERN), GDB meeting, CERN, 2006/06/07 SRM v2.2 working group update Results of the May workshop at FNAL
CERN SRM Development Benjamin Coutourier Shaun de Witt CHEP06 - Mumbai.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
David Adams ATLAS ATLAS distributed data management David Adams BNL February 22, 2005 Database working group ATLAS software workshop.
Storage Classes report GDB Oct Artem Trunov
INFSO-RI Enabling Grids for E-sciencE SRMv2.2 in DPM Sophie Lemaitre Jean-Philippe.
Author - Title- Date - n° 1 Partner Logo WP5 Status John Gordon Budapest September 2002.
CASTOR project status CASTOR project status CERNIT-PDP/DM October 1999.
WLCG Grid Deployment Board CERN, 14 May 2008 Storage Update Flavia Donno CERN/IT.
Handling of T1D0 in CCRC’08 Tier-0 data handling Tier-1 data handling Experiment data handling Reprocessing Recalling files from tape Tier-0 data handling,
SRM v2.2 Production Deployment SRM v2.2 production deployment at CERN now underway. – One ‘endpoint’ per LHC experiment, plus a public one (as for CASTOR2).
Dynamic staging to a CAF cluster Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE CAF / PROOF Workshop,
INFSO-RI Enabling Grids for E-sciencE Enabling Grids for E-sciencE Storage Element Model and Proposal for Glue 1.3 Flavia Donno,
SRM-iRODS Interface Development WeiLong UENG Academia Sinica Grid Computing 1.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Grid Deployment Board 5 December 2007 GSSD Status Report Flavia Donno CERN/IT-GD.
1 SRM v2.2 Discussion of key concepts, methods and behaviour F. Donno CERN 11 February 2008.
The Grid Storage System Deployment Working Group 6 th February 2007 Flavia Donno IT/GD, CERN.
The GridPP DIRAC project DIRAC for non-LHC communities.
CASTOR in SC Operational aspects Vladimír Bahyl CERN IT-FIO 3 2.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Architecture of LHC File Catalog Valeria Ardizzone INFN Catania – EGEE-II NA3/NA4.
SRM 2.2: experiment requirements, status and deployment plans 6 th March 2007 Flavia Donno, INFN and IT/GD, CERN.
Enabling Grids for E-sciencE EGEE-II INFSO-RI Status of SRB/SRM interface development Fu-Ming Tsai Academia Sinica Grid Computing.
Enabling Grids for E-sciencE EGEE-II INFSO-RI The Development of SRM interface for SRB Fu-Ming Tsai Academia Sinica Grid Computing.
EGEE Data Management Services
Federating Data in the ALICE Experiment
Status of the SRM 2.2 MoU extension
Flavia Donno CERN GSSD Storage Workshop 3 July 2007
GDB 8th March 2006 Flavia Donno IT/GD, CERN
SRM v2.2 / v3 meeting report SRM v2.2 meeting Aug. 29
SRM v2.2 planning Critical features for WLCG
SRM Developers' Response to Enhancement Requests
SRM2 Migration Strategy
Ákos Frohner EGEE'08 September 2008
The INFN Tier-1 Storage Implementation
Stephen Burke, PPARC/RAL Jeff Templon, NIKHEF
Data Management cluster summary
Architecture of the gLite Data Management System
INFNGRID Workshop – Bari, Italy, October 2004
The LHCb Computing Data Challenge DC06
Presentation transcript:

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 An update on SRM Mumbai & subsequent discussion summary – See the Storage Types BOF at the end of the agenda, with the material attached. –Workshop summary (version 2): v2.x shopping list More…

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Mumbai SC4 in a nutshell Tony Cass: “we are the masters and can decide what words mean in our environment. This we did in Mumbai: permanent means with tape copy, durable means without. ” Permanent –System manages cache –Access sometimes slow Durable –User (VO) manages cache –Access fast

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 How to implement it Let info system publish an SAPath + SAType per storage class –Path can be the same or different for both classes Modify clients to pick up the correct path as instructed by user –Backward compatible default behavior Client shall set “wantPermanent” flag corresponding to SAType –True  permanent, false  durable, null  server decides

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Example for dCache dn: GlueSALocalID=dteam-durable,GlueSEUniqueID=srm.my_domain,mds-vo-name=... [...] GlueSARoot: dteam:/pnfs/my_domain/foo/dteam GlueSAPath: /pnfs/my_domain/foo/dteam GlueSAType: durable [...] GlueChunkKey: GlueSEUniqueID=srm.my_domain [...] dn: GlueSALocalID=dteam-permanent,GlueSEUniqueID=srm.my_domain,mds-vo-name=... [...] GlueSARoot: dteam:/pnfs/my_domain/bar/dteam GlueSAPath: /pnfs/my_domain/bar/dteam GlueSAType: permanent [...] GlueChunkKey: GlueSEUniqueID=srm.my_domain [...]

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Example for CASTOR dn: GlueSALocalID=dteam-durable,GlueSEUniqueID=srm.my_domain,mds-vo-name=... [...] GlueSARoot: dteam:/castor/my_domain/foobar/dteam GlueSAPath: /castor/my_domain/foobar/dteam GlueSAType: durable [...] GlueChunkKey: GlueSEUniqueID=srm.my_domain [...] dn: GlueSALocalID=dteam-permanent,GlueSEUniqueID=srm.my_domain,mds-vo- name=... [...] GlueSARoot: dteam:/castor/my_domain/foobar/dteam GlueSAPath: /castor/my_domain/foobar/dteam GlueSAType: permanent [...] GlueChunkKey: GlueSEUniqueID=srm.my_domain [...]

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Example usage Lcg-utils/GFAL $ lcg-cr --stype durable file:... -d castorsrm.cern.ch -l lfn:... Default: $ lcg-cr --stype permanent file:... -d castorsrm.cern.ch -l lfn:... For SRM v2.x it might become “--stype durability:accessibility:lifetime” FTS –Option to set storage class per transfer –VO plug-in can determine correct SURL Could cache results of BDII query to find all SAPaths Default could use “generated” directory under correct SAPath

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Graeme Stewart’s matrix Storage Type Storage Types Action Requested Found permanent permanent copy to permanent path permanent durable error durable permanent error durable durable copy to durable path null permanent copy to permanent path null durable copy to durable path null both copy to permanent path Mixed client/server responsibility Some servers may not recognize the indicated error conditions

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Don Petravick’s Storage Ontology Abstract SRM v1 web svc SRM v2 web svc GLUE concept API notes API notes notes File retention not modeled permanent, permanent, time durable, volatile durable, volatile, other (SAType) Quality of not modeled not modeled tape, (multi)disk, retention other (implicit in SE architecture) Uniformity of not discoverable, not discoverable, tape, (multi)disk, transfer manageable via manageable via other (implicit in performance pin prepareToGet SE architecture) (with side effects)

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Experiment’s view LHCb and others distinguish these classes: –Files that are rarely used but of high value High quality of retention  tape –Files that are frequently used, less costly to regenerate or retrieve from other replica Disk (precious files also on tape) –Files that are frequently used, are highly mutable, often small, but need fast access and resilience User files  disk and tape (unless scratch) LHCb want to put storage type in path –Can be done in VO namespace, not a requirement on SE implementation

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Tier-1 view Lionel Schwarz (IN2P3 Lyon): –Big sites may need multiple SAPaths per class! The clients had better deal with that Jos van Wezel (FZK): –“At the moment no one is charging real money but the moment you do that users must have the ability to choose the storage based on price/speed/reliability etc.” May need to extend the ontology even further

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Critical features for WLCG Result of WLCG Baseline Services Working Group – Originally planned to be implemented by WLCG Service Challenge 4 –Delayed until autumn 2006 Features from version critical subset of version 2.1 (Nick Brook, SC3 planning meeting – June ’05) –File types –Space reservation –Permission functions –Directory functions –Data transfer control functions –Relative paths –Query supported protocols

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 File types Volatile –Temporary and sharable copy of an MSS resident file –If not pinned it can be removed by the garbage collector as space is needed (typically according to LRU policy) Durable –File can only be removed if the system has copied it to an archive Permanent –System cannot remove file Users can always explicitly delete files For SC4 the experiments only want durable and permanent

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Space reservation v1.1 –Space reservation done on file-by-file basis –User does not know in advance if SE will be able to store all files in multi-file request v2.1 –Allows for a user to reserve space But can 100 GB be used by a single 100 GB file or by 100 files of 1 GB each? MSS space vs. disk cache space –Reservation has a lifetime –“PrepareToGet(Put)” requests fail if not enough space v3.0 –Allows for “streaming” When space is exhausted requests wait until space is released –Not needed for SC4 What about quotas? –Strong interest from LHC VOs, but not yet accepted as task for SRM

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Permission functions v2.1 allows for POSIX-like ACLs –Can be associated per directory and per file –Parent directory ACLs inherited by default –Can no longer let a simple UNIX file system deal with all the permissions Need file system with ACLs or ACL-aware permission manager in SRM –May conflict with legacy applications LHC VOs desire storage system to respect permissions based on VOMS roles and groups –Currently only supported by DPM File ownership by individual users not needed in SC4 –Systems shall distinguish production managers from unprivileged users Write access to precious directories, dedicated stager pools Supported by all implementations

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Directory functions Create/remove directories Delete files –v1.1 only has an “advisory” delete Interpreted differently by different implementations –Complicates applications like the File Transfer Service Rename directories or files (on the same SE) List files and directories –Output will be truncated to implementation-dependent maximum size Full (recursive) listing could tie up or complicate server (and client) –May return huge result –Could return chunks with cookies  server would need to be stateful It is advisable to avoid very large directories No need for “mv” between SEs

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Data transfer control functions StageIn, stageOut type functionality –prepareToGet, prepareToPut Pinning and unpinning files –Avoid untimely cleanup by garbage collector –Pin has a lifetime, but can be renewed by client Avoid dependence on client to clean up Monitor status of request –How many files ready –How many files in progress –How many files left to process Suspend/resume request –Not needed for SC4 Abort request

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Relative paths Everything should be defined with respect to the VO base directory Example: srm://castorsrm.cern.ch/castor/cern.ch/grid/lhcb/DC04/prod0705/0705_123.dst SE defined by protocol and hostname VO base directory is the storage root for the VO –Advertized in information system, but unnecessary detail Clutters catalog entries SRM could insert VO base path automatically –Available in dCache VO namespace below base directory

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 Query supported protocols List of transfer protocols per SE available from information system –Workaround, complicates client –SRM knows what it supports, can inform client Client always sends SRM a list of acceptable protocols –gsiftp, (gsi)dcap, rfio, xrootd, root, … –SRM returns TURL with protocol applicable to site Query not needed for SC4

Maarten Litmaath (CERN), GDB meeting, CERN, 2006/03/08 More coordination items SRM compatibility tests –Test suite of Jiri Mencak (RAL) seems to do the job for us Clients need to keep supporting v1.1 –First try v2.x? v2.x on separate port –8444 standard? xrootd and rootd integration rfio incompatibility Quotas for user files …