The INFN Tier-1 Storage Implementation

Slides:



Advertisements
Similar presentations
LCG Tiziana Ferrari - SC3: INFN installation status report 1 Service Challenge Phase 3: Status report Tiziana Ferrari on behalf of the INFN SC team INFN.
Advertisements

T1-NREN Luca dell’Agnello CCR, 21-Ottobre The problem Computing for LHC experiments –Multi tier model (MONARC) –LHC computing based on grid –Experiment.
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
– n° 1 StoRM latest performance test results Alberto Forti Otranto, Jun
INFN-T1 site report Giuseppe Misurelli On behalf of INFN-T1 staff HEPiX Spring 2015.
Luca dell’Agnello INFN-CNAF FNAL, May
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
Toward new HSM solution using GPFS/TSM/StoRM integration Vladimir Sapunenko (INFN, CNAF) Luca dell’Agnello (INFN, CNAF) Daniele Gregori (INFN, CNAF) Riccardo.
CERN, 29 August 2006 Status Report Riccardo Zappi INFN-CNAF, Bologna.
StoRM Some basics and a comparison with DPM Wahid Bhimji University of Edinburgh GridPP Storage Workshop 31-Mar-101Wahid Bhimji – StoRM.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Federico Ruggieri INFN-CNAF GDB Meeting 10 February 2004 INFN TIER1 Status.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
1 INFN-T1 site report Andrea Chierici On behalf of INFN-T1 staff 28 th October 2009.
CASTOR: CERN’s data management system CHEP03 25/3/2003 Ben Couturier, Jean-Damien Durand, Olof Bärring CERN.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
Padova, 5 October StoRM Service view Riccardo Zappi INFN-CNAF Bologna.
Handling of T1D0 in CCRC’08 Tier-0 data handling Tier-1 data handling Experiment data handling Reprocessing Recalling files from tape Tier-0 data handling,
StoRM: status report A disk based SRM server.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
IT-INFN-CNAF Status Update LHC-OPN Meeting INFN CNAF, December 2009 Stefano Zani 10/11/2009Stefano Zani INFN CNAF (TIER1 Staff)1.
Storage & Database Team Activity Report INFN CNAF,
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
Bologna, March 30, 2006 Riccardo Zappi / Luca Magnoni INFN-CNAF, Bologna.
Stato del Tier1 Luca dell’Agnello 11 Maggio 2012.
Validation tests of CNAF storage infrastructure Luca dell’Agnello INFN-CNAF.
Riccardo Zappi INFN-CNAF SRM Breakout session. February 28, 2012 Ingredients 1. Basic ingredients (Fabric & Conn. level) 2. (Grid) Middleware ingredients.
EGEE Data Management Services
status, usage and perspectives
CASTOR: possible evolution into the LHC era
Dynamic Extension of the INFN Tier-1 on external resources
Extending the farm to external sites: the INFN Tier-1 experience
DPM at ATLAS sites and testbeds in Italy
Luca dell’Agnello INFN-CNAF
LCG Service Challenge: Planning and Milestones
GEMSS: GPFS/TSM/StoRM
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
StoRM: a SRM solution for disk based storage systems
Vincenzo Spinoso EGI.eu/INFN
Status of the SRM 2.2 MoU extension
Andrea Chierici On behalf of INFN-T1 staff
CC - IN2P3 Site Report Hepix Spring meeting 2011 Darmstadt May 3rd
Service Challenge 3 CERN
SRM v2.2 / v3 meeting report SRM v2.2 meeting Aug. 29
StoRM Architecture and Daemons
Introduction to Data Management in EGI
Bernd Panzer-Steindel, CERN/IT
CERN Lustre Evaluation and Storage Outlook
Luca dell’Agnello INFN-CNAF
Olof Bärring LCG-LHCC Review, 22nd September 2008
Ákos Frohner EGEE'08 September 2008
Stephen Burke, PPARC/RAL Jeff Templon, NIKHEF
CTA: CERN Tape Archive Overview and architecture
WLCG Storage planning Andrea Sciabà
Vladimir Sapunenko On behalf of INFN-T1 staff HEPiX Spring 2017
Experience with GPFS and StoRM at the INFN Tier-1
Storage resources management and access at TIER1 CNAF
CASTOR: CERN’s data management system
Architecture of the gLite Data Management System
INFNGRID Workshop – Bari, Italy, October 2004
StoRM disk management middleware
Presentation transcript:

The INFN Tier-1 Storage Implementation Luca dell’Agnello INFN-CNAF February, 6 2007

WAN connectivity T1 CNAF LAN GEANT GARR FZK 10 Gbps default 10 Gbps LHCOPN Juniper GARR T1 CNAF LAN

Storage layout LHCOPN gridftp 10 gb/s rfio nx2 gb/s nx1 gb/s 2x2 gb/s 2x1 gb/s each rack nx2 gb/s rfio gpfs wn FC Ethernet

Storage & Mass storage Total disk space at Tier1: ~ 600 TB raw (~ 500 TB net disk space) Completely allocated New storage (400 TB raw) arrived – prod. December Mainly based on SATA-FC technology (for general use) Good performances and scalability Raid 5 + hot spare Possibility to have redundant path to the data Access via gridftp, rfio, native GPFS Still some NFS storage (to be phased off) HMS based on CASTOR (v2) StorageTek L5500 library (up to 20 drives, 1 PB on line) 6 LTO2 drives (20-30 MB/s) with 1300 tapes 7 (+3) 9940B drives (25-30 MB/s) with 1350 (+600) tapes CASTOR file system hides tape level Es. /castor/cnaf.infn.it/lcg/cms/prova.dat Access to HSM Native access protocol: rfio srm interface for grid fabric available (rfio/gridftp) 192 FC ports 2x2 Gbps A1 A2 B2 2 x 2GB Interlink connections RAID5 B1

Hw evolution (2007) New storage system (400 TB) ready for production 3 Clarion CX3-80 (EMC) SATA – FC technology 36 new disk servers Increase of disk capacity depending of infrastructural upgrade (cooling and power) Probably additional 200 TB of disk in the meantime Increase of # of disk servers Tender for new library starting now

CASTOR CASTOR v. 2.1.1-9 1 stager instance for production (supporting other VOs besides LHC) 30 (25 for LHC VOs) disk servers supporting rfio, gridftp protocols and interconnected to storage via FC 12 tape servers 2 srm 1.1 end-points (1 instance with no tape back-end, no gc) At present 1 pool per storage class for each LHC experiment Only 3 disk servers for each pool (need probably to increase number) 1 stager instance for tests 1srm 2.2 end-point for basic tests 3 disk servers (access to internal disks only at the moment)

GPFS GPFS cluster including all WNs and GPFS disk servers 100 TB of disk served V 3.1release installed (stable since a few months) 12 disk servers interconnected via FC to SAN No SRM 1.1 interface provided Only possible to use as classic SE A small cluster for test purpose available

StoRM SRM 2.2 interface to POXIS file systems with ACL support (GPFS, XFS) Provides support for disk-only storage systems (D1T0) It is possible to have different Storage Areas in the same SC instance e.g.srm://storm02.cr.cnaf.infn.it:8444//srm/managerv2?SFN=/lhcb/SA1/dir/test.txt where /lhcb/SA1 identifies the SA In next version a "targetSpaceToken" will identify the SA A test SRM 2.2 instance (StoRM) installed First tests started (using ad hoc clients)

Planned (basic) tests Verification (certification) of storage infrastructure (SAN, LAN, servers) Throughput tests for both CASTOR and GPFS Verification of scalability of GPFS Comparison tests between xrootd, GPFS, rfio

Storage classes implementation Disk0Tape1 is and will be CASTOR Space managed by system Data migrated to tapes and deleted from when staging area full Disk1tape1 will be (probably!) CASTOR Space managed by VO (i.e. if disk is full, copy fails) Large buffer of disk with tape back end (and gc with an high threshold?) Disk1tape0 also investigating GPFS/StoRM Space managed by VO Open issue: efficiency of data moving to and from CASTOR (some VOs asked for this and also for some sort of backup  at least in the first phase) Still need to investigate clearly experiments needs E.g. LAN, WAN differences

StoRM StoRM is a storage resource manager for disk based storage systems. It implements the SRM interface version 2.x. StoRM is designed to support guaranteed space reservation and direct access (native POSIX I/O call), as well as other standard libraries (like RFIO). StoRM take advantage from high performance parallel file systems. Also standard POSIX file systems are supported. A modular architecture decouples StoRM logic from the supported file system. Strong security framework with VOMS support. by L. Magnoni

StoRM General Considerations 1/2 File system currently supported by StoRM GPFS from IBM. XFS from SGI. Any other File System with POSIX interface and ACLs support. Light and flexible namespace structure The namespace of the files managed by StoRM relies upon the underlying file systems. StoRM does not need to query any DB to know the physical location of a requested SURL. by L. Magnoni

StoRM General Considerations 2/2 ACLs Usage StoRM enforce ACL entries on physical files for the local user corresponding to the grid-credential. Standard grid applications (such as GridFTP, RFIO, etc.) can access the storage on behalf of the user. Scalability and high availability. FE, DB, and BE can be deployed in 3 different machines. StoRM is designed to be configured with n FE and m BE, with a common DB. But more tests are needed to validate this scenario. by L. Magnoni

StoRM Grid usage scenario StoRM dynamically manages files and space in the storage system. Applications can directly access the Storage Element (SE) during the computational process. by L. Magnoni File metadata are managed (and stored) by underlying file system. No replica of metadata at application level.. That is a file system job! In this way StoRM gain in performance. Data access is performed without interacting with an external service, with great performance improvement (POSIX calls). Otherwise, standard data access using I/O Server (such as RFIO) is also fully supported.

StoRM status and SRM issues Migration to SRM v2.2 completed. All functions requested by the SRM WLCG usage agreement are implemented. New version of StoRM available. StoRM SRM tests StoRM is involved in interoperability tests made by SRM-WG, the results are available here: http://sdm.lbl.gov/srm-tester/v22-progress.html StoRM is involved also in other SRM tests made with S2 test suite: http://gdrb02.cern.ch:25000/srms2test/scripts/protos/srm/2.2/basic/s2_logs/ by L. Magnoni