Stephen Burke, PPARC/RAL Jeff Templon, NIKHEF

Slides:



Advertisements
Similar presentations
DataTAG WP4 Meeting CNAF Jan 14, 2003 Interfacing AliEn and EDG 1/13 Stefano Bagnasco, INFN Torino Interfacing AliEn to EDG Stefano Bagnasco, INFN Torino.
Advertisements

WP2: Data Management Gavin McCance University of Glasgow November 5, 2001.
Stephen Burke - WP8 Status - 9/5/2002 Partner Logo WP8 Status Stephen Burke, PPARC/RAL.
Stephen Burke - WP8 Status - 14/2/2002 Partner Logo WP8 Status Stephen Burke, PPARC/RAL.
Chapter 12: File System Implementation
Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
Author - Title- Date - n° 1 GDMP The European DataGrid Project Team
Stephen Burke – Heidelberg - 26/9/2003 Partner Logo Overview of applications view of the data management middleware Stephen Burke.
NIKHEF Testbed 1 Plans for the coming three months.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
DataGrid is a project funded by the European Union CHEP 2003 – March 2003 – Title – n° 1 Grid Data Management in Action Experience in Running and.
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
WP9 – Earth Observation Applications – n° 1 Experiences with Testbed1, plans and objectives for Testbed 2 Testbed retreat th August 2002
KNMI Applications on Testbed 1 …and other activities.
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
LHCb Applications and GRID Integration Domenico Galli Catania, April 9, st INFN-GRID Workshop.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL U.S. ATLAS Physics and Computing Advisory Panel Review Argonne National Laboratory Oct 30, 2001.
File and Object Replication in Data Grids Chin-Yi Tsai.
LHCb and DataGRID - the workplan for 2001 Eric van Herwijnen Wednesday, 28 march 2001.
CMS Stress Test Report Marco Verlato (INFN-Padova) INFN-GRID Testbed Meeting 17 Gennaio 2003.
MAGDA Roger Jones UCL 16 th December RWL Jones, Lancaster University MAGDA  Main authors: Wensheng Deng, Torre Wenaus Wensheng DengTorre WenausWensheng.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
Author - Title- Date - n° 1 Partner Logo WP5 Summary Paris John Gordon WP5 6th March 2002.
Magda Distributed Data Manager Status Torre Wenaus BNL ATLAS Data Challenge Workshop Feb 1, 2002 CERN.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
First attempt for validating/testing Testbed 1 Globus and middleware services WP6 Meeting, December 2001 Flavia Donno, Marco Serra for IT and WPs.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Data Access for Analysis Jeff Templon PDP Groep, NIKHEF A. Tsaregorodtsev, F. Carminati, D. Liko, R. Trompert GDB Meeting 8 march 2006.
11/5/2001WP5 UKHEPGRID1 WP5 Mass Storage UK HEPGrid UCL 11th May Tim Folkes, RAL
Stephen Burke – Data Management - 3/9/02 Partner Logo Data Management Stephen Burke, PPARC/RAL Jeff Templon, NIKHEF.
GridPP Collaboration Meeting 5 th November 2001 Dan Tovey, University of Sheffield Non-LHC and Non-US-Collider Experiments’ Requirements Dan Tovey, University.
SRM & SE Jens G Jensen WP5 ATF, December Collaborators Rutherford Appleton (ATLAS datastore) CERN (CASTOR) Fermilab Jefferson Lab Lawrence Berkeley.
1 LHCb File Transfer framework N. Brook, Ph. Charpentier, A.Tsaregorodtsev LCG Storage Management Workshop, 6 April 2005, CERN.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
ATLAS Magda Distributed Data Manager Torre Wenaus BNL PPDG Robust File Replication Meeting Jefferson Lab January 10, 2002.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
Oxana Smirnova LCG/ATLAS/Lund September 3, 2002, Budapest 5 th EU DataGrid Conference ATLAS-EDG Task Force status report.
10 May 2001WP6 Testbed Meeting1 WP5 - Mass Storage Management Jean-Philippe Baud PDP/IT/CERN.
Data Management The European DataGrid Project Team
Magda Distributed Data Manager Prototype Torre Wenaus BNL September 2001.
Data Management The European DataGrid Project Team
Author - Title- Date - n° 1 Partner Logo WP5 Status John Gordon Budapest September 2002.
The DataGrid Project NIKHEF, Wetenschappelijke Jaarvergadering, 19 December 2002
Stephen Burke – Sysman meeting - 22/4/2002 Partner Logo The Testbed – A User View Stephen Burke, PPARC/RAL.
Storage Element Security Jens G Jensen, WP5 Barcelona, May 2003.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Architecture of LHC File Catalog Valeria Ardizzone INFN Catania – EGEE-II NA3/NA4.
What to expect in TB 2.0 J. Templon, NIKHEF/WP8. Jeff Templon – AWG Meeting, NIKHEF, WP1 u New resource broker architecture Big potato.
Magda Distributed Data Manager Torre Wenaus BNL October 2001.
J Jensen / WP5 /RAL UCL 4/5 March 2004 GridPP / DataGrid wrap-up Mass Storage Management J Jensen
File-System Management
Federating Data in the ALICE Experiment
The EDG Testbed Deployment Details
Work Package 9 – EO Applications
Oxana Smirnova, Jakob Nielsen (Lund University/CERN)
Classic Storage Element
Status of the SRM 2.2 MoU extension
Moving the LHCb Monte Carlo production system to the GRID
Report of Dubna discussion
The European DataGrid Project Team
John Gordon EDG Conference Barcelona, May 2003
ALICE Physics Data Challenge 3
Data Management in Release 2
The INFN Tier-1 Storage Implementation
Chapter 2: System Structures
The EU DataGrid Data Management
Architecture of the gLite Data Management System
Status and plans for bookkeeping system and production tools
Presentation transcript:

Stephen Burke, PPARC/RAL Jeff Templon, NIKHEF Data Management Stephen Burke, PPARC/RAL Jeff Templon, NIKHEF

Overview Current situation + some outstanding problems: GDMP status – Jeff Templon Directing jobs to data Replica Manager Mass storage

GDMP Status Jeff Templon

Sending jobs to data The JDL can specify a list of LFNs and an access protocol, and the address of the Replica Catalogue. Protocols are “file” for local (NFS) access, “gridftp” and RFIO – but Castor files can’t easily appear in an RC so RFIO is not so useful (see later). Currently all SEs support all protocols. The job will go to a CE which has all files locally. Information about the local system and filenames is available through the BrokerInfo interface.

Replica Manager No real Replica Manager yet (i.e. no code to replicate files automatically), but some prototype commands are available. There are commands to copy a file into or between SEs, and to update the Replica Catalogue. There are commands to delete files and RC entries . Designed for single files rather than batches. Does not write full file information into the RC. See the tutorial by Heinz Stockinger.

Mass Storage: RFIO The basic solution for accessing tape stores is RFIO. Files can be copied between Castor and disks at CERN using rfcp as normal. Anyone with a CERN tape account is mapped to that account, otherwise to generic accounts with enough privilege to read and write. rfcp can be used from a WN at CERN. It is not currently on the UIs. This usage can be extended to HPSS at Lyon. RFIO can do direct file access with API calls.

RFIO drawbacks RFIO only works on machines local to the tape store (security issues). Files in Castor can’t easily be recorded in a Replica Catalogue, and even if they could the WP2 tools would not know how to deal with them. There is no GridFTP access to tape files (coming soon?). RAL has a completely different system (virtual tapes), although this can also be used in a similar way.

GDMP hooks GDMP has some hooks to allow MSS access. When a file is replicated from another site to CERN it can be automatically copied to a fixed location in Castor. However, locally-created files are not copied automatically. When GDMP gets a request to replicate a CERN file, if the file is not on disk it tries to stage it first (this currently fails due to a bug). Again, local access does nothing.

WP5 staging scripts The current testbed has some staging scripts which use rfcp to copy files to/from a fixed directory in Castor. Drawbacks: file names can clash, Castor has a maximum file name length of 130 (?) characters, the scripts are not callable by users. New scripts will solve all these problems, at the cost of using a relatively opaque file naming scheme (MD5 hashes).

MSS Problems No disk space management, users need to manage the space themselves. Automatic staging is limited to replication, in general users have to stage in and out themselves. Files are staged to a specific place with opaque filenames, which may be hard to integrate with an external file catalogue. Files in other places in Castor cannot be registered in Replica Catalogues or manipulated by Grid tools. Experiments need to define their needs to see if they can live with this in TB1.2.