Introduction to Storage Element Hsin-Wei Wu Academia Sinica Grid Computing Center, Taiwan.

Slides:



Advertisements
Similar presentations
Data Management Expert Panel - WP2. WP2 Overview.
Advertisements

Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
1 SRM-Lite: overcoming the firewall barrier for large scale file replication Arie Shoshani Alex Sim Lawrence Berkeley National Laboratory April, 2007.
1 CHEP 2003 Arie Shoshani Experience with Deploying Storage Resource Managers to Achieve Robust File replication Arie Shoshani Alex Sim Junmin Gu Scientific.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
Data Grid Web Services Chip Watson Jie Chen, Ying Chen, Bryan Hess, Walt Akers.
EGEE-II INFSO-RI Enabling Grids for E-sciencE gLite Data Management System Yaodong Cheng CC-IHEP, Chinese Academy.
INFSO-RI Enabling Grids for E-sciencE gLite Data Management Services - Overview Mike Mineter National e-Science Centre, Edinburgh.
A. Sim, CRD, L B N L 1 Oct. 23, 2008 BeStMan Extra Slides.
1 School of Computer, National University of Defense Technology A Profile on the Grid Data Engine (GridDaEn) Xiao Nong
Data management in grid. Comparative analysis of storage systems in WLCG.
ESP workshop, Sept 2003 the Earth System Grid data portal presented by Luca Cinquini (NCAR/SCD/VETS) Acknowledgments: ESG.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Data Grid Services/SRB/SRM & Practical Hai-Ning Wu Academia Sinica Grid Computing.
1 Use of SRMs in Earth System Grid Arie Shoshani Alex Sim Lawrence Berkeley National Laboratory.
Data Management The GSM-WG Perspective. Background SRM is the Storage Resource Manager A Control protocol for Mass Storage Systems Standard protocol:
The LCG File Catalog (LFC) Jean-Philippe Baud – Sophie Lemaitre IT-GD, CERN May 2005.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware Data Management in gLite.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
Author - Title- Date - n° 1 Partner Logo WP5 Summary Paris John Gordon WP5 6th March 2002.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
Author: Andrew C. Smith Abstract: LHCb's participation in LCG's Service Challenge 3 involves testing the bulk data transfer infrastructure developed to.
INFSO-RI Enabling Grids for E-sciencE Experiences with LFC and comparison with RNS Erwin Laure Jean-Philippe.
E-science grid facility for Europe and Latin America Data Management Services E2GRIS1 Rafael Silva – UFCG (Brazil) Universidade Federal.
1 SRM-Lite: overcoming the firewall barrier for data movement Arie Shoshani Alex Sim Viji Natarajan Lawrence Berkeley National Laboratory SDM Center All-Hands.
1 LHCb File Transfer framework N. Brook, Ph. Charpentier, A.Tsaregorodtsev LCG Storage Management Workshop, 6 April 2005, CERN.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
INFSO-RI Enabling Grids for E-sciencE Αthanasia Asiki Computing Systems Laboratory, National Technical.
Managing Data DIRAC Project. Outline  Data management components  Storage Elements  File Catalogs  DIRAC conventions for user data  Data operation.
SEE-GRID-SCI Storage Element Installation and Configuration Branimir Ackovic Institute of Physics Serbia The SEE-GRID-SCI.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
Padova, 5 October StoRM Service view Riccardo Zappi INFN-CNAF Bologna.
The new FTS – proposal FTS status. EMI INFSO-RI /05/ FTS /05/ /05/ Bugs fixed – Support an SE publishing more than.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Data management in LCG and EGEE David Smith.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks gLite Data Management Components Presenter.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache Implementation at FZK Forschungszentrum Karlsruhe.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Data management in EGEE.
1 DIRAC Data Management Components A.Tsaregorodtsev, CPPM, Marseille DIRAC review panel meeting, 15 November 2005, CERN.
INFSO-RI Enabling Grids for E-sciencE University of Coimbra gLite 1.4 Data Management System Salvatore Scifo, Riccardo Bruno Test.
INFSO-RI Enabling Grids for E-sciencE University of Coimbra Data Management System gLite – LCG – FiReMan Salvatore Scifo INFN Catania.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Architecture of LHC File Catalog Valeria Ardizzone INFN Catania – EGEE-II NA3/NA4.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) Algiers, EUMED/Epikh Application Porting Tutorial, 2010/07/04.
Enabling Grids for E-sciencE EGEE-II INFSO-RI Status of SRB/SRM interface development Fu-Ming Tsai Academia Sinica Grid Computing.
Security recommendations DPM Jean-Philippe Baud CERN/IT.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Data Management Maha Metawei
A. Sim, CRD, L B N L 1 Production Data Management Workshop, Mar. 3, 2009 BeStMan and Xrootd Alex Sim Scientific Data Management Research Group Computational.
FP6−2004−Infrastructures−6-SSA E-infrastructure shared between Europe and Latin America LFC Server Installation and Configuration.
Riccardo Zappi INFN-CNAF SRM Breakout session. February 28, 2012 Ingredients 1. Basic ingredients (Fabric & Conn. level) 2. (Grid) Middleware ingredients.
EGEE Data Management Services
CASTOR: possible evolution into the LHC era
Jean-Philippe Baud, IT-GD, CERN November 2007
GFAL Grid File Access Library
gLite Basic APIs Christos Filippidis
StoRM: a SRM solution for disk based storage systems
Vincenzo Spinoso EGI.eu/INFN
gLite Data management system overview
GGF OGSA-WG, Data Use Cases Peter Kunszt Middleware Activity, Data Management Cluster EGEE is a project funded by the European.
StoRM Architecture and Daemons
Introduction to Data Management in EGI
Data Management cluster summary
Data Management Ouafa Bentaleb CERIST, Algeria
Data services in gLite “s” gLite and LCG.
Architecture of the gLite Data Management System
gLite Data and Metadata Management
INFNGRID Workshop – Bari, Italy, October 2004
Presentation transcript:

Introduction to Storage Element Hsin-Wei Wu Academia Sinica Grid Computing Center, Taiwan

 SRM  A union form standard for SEs  Storage Elements  BeStMan  Castor  dCash  DPM  StoRM  LFC (LHC File Catalog)  Demonstration  Data Management Commands 2 Outline

3 API Access Job Mgmt. Services Computing Element Workload Management Metadata Catalog Data Services Storage Element Data Movement File & Replica Catalog Authorization Security Services Authentication Information & Monitoring Information & Monitoring Services Service Discovering Accounting Auditing Job Provenance Package Manager CLI Network Monitoring Overview of gLite Middleware

Data service in gLtie Ref gLite-3.2 User Guide 4  Storage(SE): The files are physical located  Storage URL or SURL  srm://as-ds01.euasiagrid.org/dpm/euasiagrid.org/data  Catalogs: High level hierarchical namespace, maps the “physical” files to a virtual “logical” filename  Logical File Name or LFN  lfn:/grid/twgrid/data  Movement: put/get files into grid SE's, move/replicate files between SE's  File Transfer Service or FTS  Transport URL or TRUL  gsiftp://sc003.grid.sinica.edu.tw:2811/data

5  A Storage Element (SE) is an aggregate of Grid services that allows Grid users to store and manage files together with the space assigned to them  Data are stored on disk pool servers or Mass Storage Systems  Storage resource management needs to take into account  Transparent access to files (migration to/from disk pool)  File pinning  Space reservation  File status notification  Life time management  The SRM (Storage Resource Manager) takes care of all these details  The SRM is a single interface that takes care of local storage interaction and provides a Grid interface to the outside world Storage Resource Management

Berkeley Storage Manager (BeStMan) LBNL  Java implementation  Designed to work with unix- based disk systems  As well as MSS to stage/archive from/to its own disk (currently HPSS)  Adaptable to other file systems and storages (e.g. NCAR MSS, VU L-Store, TTU Lustre, NERSC GFS)  Uses in-memory database (BerkeleyDB) Multiple transfer protocolsMultiple transfer protocols Space reservationSpace reservation Directory management (no ACLs)Directory management (no ACLs) Can copy files from/to remote SRMs or GridFTP ServersCan copy files from/to remote SRMs or GridFTP Servers Can copy entire directory recursivelyCan copy entire directory recursively Large scale data movement of thousands of files Recovers from transient failures (e.g. MSS maintenance, network down) Local PolicyLocal Policy Fair request processing File replacement in disk Garbage collection

Castor-SRM CERN and Rutherford Appleton Laboratory  CASTOR is the HSM in production at CERN  Support for multiple tape robots  Support for Disk-only storage recently added  Designed to meet Large Hadron Collider Computing requirements  Maximize throughput from clients to tape (e.g. LHC experiments data taking) C++ ImplementationC++ Implementation Reuse of CASTOR software infrastructureReuse of CASTOR software infrastructure Derived SRM specific classes Configurable number of thread pools for both front- and back- endsConfigurable number of thread pools for both front- and back- ends ORACLE centricORACLE centric Front and back ends can be distributed on multiple hostsFront and back ends can be distributed on multiple hosts

dCache-SRM FNAL and DESY  Strict name space and data storage separation  Automatic file replication based on access patterns  HSM Connectivity (Enstore, OSM, TSM, HPSS, DMF)  Automated HSM migration and restore  Scales to Peta-byte range on 1000’s of disks  Supported protocols:  (gsi/krb)FTP, (gsi/krb)dCap, xRoot, NFS 2/3  Separate I/O queues per protocol  Resilient dataset management  Command line and graphical admin interface  Variety of Authorization mechanisms including VOMS  Deployed in a large number of institutions worldwide Support SRM 1.1 and SRM 2.2Support SRM 1.1 and SRM 2.2 Dynamic Space ManagementDynamic Space Management Request queuing and schedulingRequest queuing and scheduling Load balancingLoad balancing Robust replication using srmCopy functionality via SRM, (gsi)FTP and http protocolsRobust replication using srmCopy functionality via SRM, (gsi)FTP and http protocols

Disk Pool Manager (DPM) CERN  Provide a reliable, secure and robust storage system  Manages storage on disks only  Security  GSI for authentication  VOMS for authorization  Standard POSIX permissions + ACLs based on user’s DN and VOMS roles  Virtual ids  Accounts created on the fly  Full SRMv2.2 implementation  Standard disk pool manager capabilities  Garbage collector  Replication of hot files  Transfer protocols  GridFTP (v1 and v2)  Secure RFIO  https  Xroot  Works on Linux 32/64 bits machines  Direct data transfer from/to disk server (no bottleneck)  Support DICOM backend  Requirement from Biomed VO  Storage of encrypted files in DPM on the fly + local decryption  Use of GFAL/srm to get TURLs and decrypt the file Supported database backendsSupported database backends MySQL Oracle High availabilityHigh availability All servers can be load balanced (except the DPM one) Resilient: all states are kept in the DB at all times

Storage Resource Manager (StoRM) INFN/CNAF - ICTP/EGRID  It's designed to leverage the advantages of high performing parallel file systems in Grid.  Different file systems supported through a driver mechanism:  generic POSIX FS  GPFS  Lustre  XFS  It provides the capability to perform local and secure access to storage resources (file:// access protocol + ACLs on data).file:// StoRM architecture: FrontendsFrontends: C/C++ based, expose the SRM interface BackendsBackends: Java based, execute SRM requests. DBDB: based on MySQL DBMS, stores requests data and StoRM metadata. replicated and instantiated on a dedicated machine.Each component can be replicated and instantiated on a dedicated machine.

LHC File Catalog 5th NRENs and Grids Workshop, Paris June  LFC maps LFNs to SURLs  Logical File Name (LFN): user file name  in VO namespace, aliases supported  Glbally Unique IDentifier (GUID)  unique string assigned by the system to the file  Site URL (SURL): identifies a replica  A Storage Element and the logical name of the file inside it  GSI security: ACLs (based on VOMS)  To each VOMS group/role corresponds a virtual group identifier  Support for secondary groups  Web Service query interface: Data Location Interface (DLI)  Hierarchical Namespace  Supports sessions and bulk operations LFC GUID SURL 1 SURL 2 ACL LFN 1 LFN 2 lfc-ls –l /grid/vo/ /grid /vo /data file LFC DLI lfc-getacl /grid/vo/data

Data Management Command Ref gLite-3.2 User Guide 12 SRMLFCDPM listlcg-ls SRMlfc-ls LFNdpns-ls DPNS upload file lcg-cp LOCAL SRM globus-url-copy LOCAL SRM lcg-cr -d SE -l LFN LOCALrfcp LOCAL DPNS download file lcg-cp SRM LOCAL globus-url-copy SRM LOCAL lcg-cp LFN LOCALrfcp DPNS LOCAL delete file lcg-del --nolfc SRMlcg-del -a LFNrfrm DPNS environment variable LFC_HOST LFC_HOME DPM_HOST DPNS_HOST DPMS_HOME

DPM Demonstration Ref gLite-3.2 User Guide 13  Initial Proxy  Check DPM_HOST and DPNS_HOST

DPM Demonstration Ref gLite-3.2 User Guide 14  dpns-ls  Upload file to SE

DPM Demonstration Ref gLite-3.2 User Guide 15  Remove file

SRM Demonstration Ref gLite-3.2 User Guide 16  Initial Proxy  Check SE

SRM Demonstration Ref gLite-3.2 User Guide 17  Ls  Upload fire to SE

SRM Demonstration Ref gLite-3.2 User Guide 18  Remove file

Thank you for your listening and patience Ref gLite-3.2 User Guide 19