dCache Status and Plans – Proposals for SC3

Slides:



Advertisements
Similar presentations
30-31 Jan 2003J G Jensen, RAL/WP5 Storage Elephant Grid Access to Mass Storage.
Advertisements

Steve Traylen Particle Physics Department Experiences of DCache at RAL UK HEP Sysman, 11/11/04 Steve Traylen
GridKa January 2005 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann 1 Mass Storage at GridKa Forschungszentrum Karlsruhe GmbH.
Data Management Expert Panel - WP2. WP2 Overview.
Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
Author - Title- Date - n° 1 Partner Logo WP5 Summary Paris John Gordon WP5 6th March 2002.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
 CASTORFS web page - CASTOR web site - FUSE web site -
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
DGC Paris WP2 Summary of Discussions and Plans Peter Z. Kunszt And the WP2 team.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
Oct 24, 2002 Michael Ernst, Fermilab DRM for Tier1 and Tier2 centers Michael Ernst Fermilab February 3, 2003.
ALCF Argonne Leadership Computing Facility GridFTP Roadmap Bill Allcock (on behalf of the GridFTP team) Argonne National Laboratory.
The new FTS – proposal FTS status. EMI INFSO-RI /05/ FTS /05/ /05/ Bugs fixed – Support an SE publishing more than.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
10 May 2001WP6 Testbed Meeting1 WP5 - Mass Storage Management Jean-Philippe Baud PDP/IT/CERN.
GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache Implementation at FZK Forschungszentrum Karlsruhe.
NeST: Network Storage John Bent, Venkateshwaran V Miron Livny, Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau.
DMLite GridFTP frontend Andrey Kiryanov IT/SDC 13/12/2013.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
A Data Handling System for Modern and Future Fermilab Experiments Robert Illingworth Fermilab Scientific Computing Division.
E-commerce Architecture Ayşe Başar Bener. Client Server Architecture E-commerce is based on client/ server architecture –Client processes requesting service.
Cofax Scalability Document Version Scaling Cofax in General The scalability of Cofax is directly related to the system software, hardware and network.
9/20/04Storage Resource Manager, Timur Perelmutov, Jon Bakken, Don Petravick, Fermilab 1 Storage Resource Manager Timur Perelmutov Jon Bakken Don Petravick.
J Jensen / WP5 /RAL UCL 4/5 March 2004 GridPP / DataGrid wrap-up Mass Storage Management J Jensen
Riccardo Zappi INFN-CNAF SRM Breakout session. February 28, 2012 Ingredients 1. Basic ingredients (Fabric & Conn. level) 2. (Grid) Middleware ingredients.
CERN IT-Storage Strategy Outlook Alberto Pace, Luca Mascetti, Julien Leduc
EGEE Data Management Services
Federating Data in the ALICE Experiment
CASTOR: possible evolution into the LHC era
Jean-Philippe Baud, IT-GD, CERN November 2007
Grid Computing: Running your Jobs around the World
Scalable sync-and-share service with dCache
StoRM: a SRM solution for disk based storage systems
Vincenzo Spinoso EGI.eu/INFN
(on behalf of the POOL team)
Status of the SRM 2.2 MoU extension
Data Bridge Solving diverse data access in scientific applications
dCache “Intro” a layperson perspective Frank Würthwein UCSD
GGF OGSA-WG, Data Use Cases Peter Kunszt Middleware Activity, Data Management Cluster EGEE is a project funded by the European.
StoRM Architecture and Daemons
Abhishek Singh Rana UC San Diego
SRM Developers' Response to Enhancement Requests
Sergio Fantinel, INFN LNL/PD
CMS analysis job and data transfer test results
Artem Trunov and EKP team EPK – Uni Karlsruhe
THE STEPS TO MANAGE THE GRID
Ákos Frohner EGEE'08 September 2008
Enabling High Speed Data Transfer in High Energy Physics
The INFN Tier-1 Storage Implementation
Large Scale Test of a storage solution based on an Industry Standard
A Web-Based Data Grid Chip Watson, Ian Bird, Jie Chen,
Chapter 2: The Linux System Part 1
INFNGRID Workshop – Bari, Italy, October 2004
Presentation transcript:

dCache Status and Plans – Proposals for SC3 Michael Ernst For the dCache Team LCG Data Management Workshop

dCache is a joint effort between DESY and Fermilab 4/05/2005 LCG-DM-WS M.Ernst

Agenda Architecture and Components dCache Functionality Layers and Basic Design Data Access Methods HSM Interface Pool Selection Mechanism SRM/dCache as LCG SE Installation & Management dCache Support Model Plans Extensions to Information Provider to support Job & Data Co-Scheduling “The Mirror Cache” 4/05/2005 LCG-DM-WS M.Ernst

dCache – The Architecture and Components Name Space uniquely represented within single file system tree Strictly separates filename space of data repository from physical location File Namespace in DB & accessible to application by NFS, SRM, GridFTP etc. Replicas of given file may exist on multiple storage nodes and MSS for e.g. load balancing (pool-to-pool transfers) Scalable Architecture Fully distributed Architecture w/ Autodiscovery for Components Integrates Heterogeneous Disk Storage and Server Technology with multiple hundred individual (commodity) nodes Automatic load balancing by cost metric and inter pool transfers Multiple Distributed Data Access Points (pluggable Door/Mover pairs) supporting different Standards for Data Access dCache distributes files autonomously across Disk Servers Selection depends on available space and server load Fine-grained steering directives to control data flow and utilization of Storage Resources (Pool Attraction) 4/05/2005 LCG-DM-WS M.Ernst

dCache – The Architecture and Components Scalable Architecture (cont’d.) Configurable as Disk-only and as part of a Storage Hierarchy (e.g. with Tape Back-end) Automatic Migration and Staging between dCache and underlying Storage System Very Flexible and Modular MSS interface - Integration done for Enstore, TSM, OSM and HPSS Resiliency Manager allows to automatically create and maintain a configurable number of copies of a given file on different storage nodes – eases maintenance (adjusts replica count on scheduled pool maintenance) and improves availability in case of pool failures Management Rich set of Admin commands coming with individual dCache Modules (Cells) CLI (via SSH) and GUI (Web interface) allow to navigate through and login to the Modules (e.g. SRM, GridFTP door, PoolManager, Storage Pools) Allow to add/remove components to/from the active system To customize/tune the system while in operation 4/05/2005 LCG-DM-WS M.Ernst

Resilient dCache Developed by Fermilab Goal is managed reliable storage without a tape backend Reliability is achieved through replication Expect pools to go in/out of service and files are replicated as this happens. Can also schedule pools offline to smooth replication process Active replica checksum comparison with replacement when necessary User web interface 4/05/2005 LCG-DM-WS M.Ernst

dCache Functionality Layers GFAL Storage Element (LCG) Storage Resource Mgr. GRIS Wide Area dCache FTP Server (GSI, Kerberos) Resilient Cache Resilient Manager Basic Cache System dCap Client (GSI, Kerberos) dCap Server PNFS dCache Core HSM Adapter Cell Package (concept by P. Fuhrmann) 4/05/2005 LCG-DM-WS M.Ernst

dCache Basic Design Door Name Space Provider Pool Manager Pool Mover Components involved in Data Storage and Data Access Pool Provides specific end point for client connection Exists as long as client process is alive Client’s proxy used within dCache Interface to a file system name space Maps dCache name space operations to filesystem operations Stores extended file metadata Performs pool selection Data repository handler Launches requested data transfer protocols Data transfer handler (gsi)dCap, (Grid)FTP, http, HSM hooks (concept by P. Fuhrmann) 4/05/2005 LCG-DM-WS M.Ernst

Data Access Methods dCap – the native protocol Library for regular file access via POSIX calls Can be linked against applications and available as preload lib Supports pluggable security mechanisms Implemented are GssApi (Kerberos, GSI) and ssl Adds resiliency by protecting clients from intermediate network and storage node failures (lib reconnects) Provides URL style addressing (alternative to mounting the pnfs namespace provider) FTP – support for multiple dialects GridFTP with GSI security GssFTP with Kerberos security Other Protocols can be easily integrated through well defined, well documented Interface 4/05/2005 LCG-DM-WS M.Ernst

Where starts your HSM ? 4/05/2005 LCG-DM-WS M.Ernst

HSM Interface 4/05/2005 LCG-DM-WS M.Ernst

HSM Interface Very flexible Interface to various MSSs (scripts) Precious data is separately collected per storage class Each storage class queue has individual steering parameters for HSM flush operation Elapsed time a file is allowed to be precious per storage class Total data volume that is precious per storage class Maximum number of precious files per storage class Maximum number of concurrent HSM flush operations is configurable Multiple HSM instances and HSM classes are simultaneously supported 4/05/2005 LCG-DM-WS M.Ernst

Deferred HSM flush 4/05/2005 LCG-DM-WS M.Ernst

Pool Selection Mechanism 4/05/2005 LCG-DM-WS M.Ernst

Pool Selection 4/05/2005 LCG-DM-WS M.Ernst

Pool Selection “Satellite” Cache Central Cache Location X Location Y 4/05/2005 LCG-DM-WS M.Ernst

Pool Selection Sub Cache in DMZ Central Cache Farm Subnet 4/05/2005 LCG-DM-WS M.Ernst

dCache Components dCache Components HSMs Pool Nodes I/O Door Nodes Admin Doors GFtp (Krb5,ssl) (K)Ftp Http SRM Admin dCap File Name Space File Name Space NFS Server Pool Manager Provider Database pnfs Pool Nodes HSMs OSM Enstore TSM 4/05/2005 LCG-DM-WS M.Ernst

Data Grid Functionality – dCache as LCG SE Aiming at Seamless Integration with LCG Infrastructure SRM related aspects covered in Timur’s talk Information Provider (GRIS - currently fairly detached, working on full integration w/ dCache system) Full compliance with LCG demanded functionalities & supplied Client Utilities, and other SE Implementations Primarily concerns SRM, GridFTP, dCap (GFAL) Working with the LCG GDT on improving packaging & documentation, installation procedure and manageability to make dCache a suitable solution for any kind of LHC relevant Computing Facility 4/05/2005 LCG-DM-WS M.Ernst

Installation & Management Few RPMs (<= 4) to install PNFS (Namespace Manager) dCache Core (for Admin, Pool, Door node) dCache Optional S/W (SRM, GridFTP, gsidcap) SRM needs PostgreSQL DB dCache Client Central Configuration provides suitable defaults & needs only minor customization Auto Configuration to add & remove pools Disk pools/nodes request to add themselves to dCache Allows simple reconfiguration of the Disk Pools Administrator can temporarily remove pools from dCache if a disk has crashed and is being repaired PoolManager takes out a Pool automatically in case it doesn’t respond any longer Medium Tier-2 site w/o tape backend and ~20 pool nodes needs approx. 2h to install dCache from scratch Upgrade is usually much faster 4/05/2005 LCG-DM-WS M.Ernst

Installable Server Components PNFS Name Space Manager Choice of two Databases GDBM (default) Postgres (not part of the dCache distribution) Admin Node Provides all central Management Components LocationManager, PoolManager, httpd, Can have Storage Pools (small configurations) Can have Access Points for a variety of Management (SRM) and Data Transfer Protocols (automatically inserted/removed) Pool Node Provides Storage Space managed by dCache Automatically inserted/removed in default configuration Can have Access Points (automatically inserted/removed) Door Node Provides Access Points for SRM and a variety of Data Transfer Protocols 4/05/2005 LCG-DM-WS M.Ernst

dCache Licensing Note: The following applies to dCache only. The SRM as it is distributed by Fermilab has different terms Current Conditions (work in progress) Free, non-exclusive license to use dCache w/o limitations The source code is available to selected partners (to be approved by dCache Collaboration) 4/05/2005 LCG-DM-WS M.Ernst

LCG - dCache Support Model 4/05/2005 LCG-DM-WS M.Ernst

Make Storage a seamlessly integrated Resource Particularly interesting for access to data on “Permanent Storage” (Tape based MSS) Access to information about the location of data sets Information about access and transfer costs Scheduling of data transfers and data availability optimize data transfers w.r.t. storage space, data access costs etc. Perfectly fits into general grid scheduling: access to similar services interaction necessary 4/05/2005 LCG-DM-WS M.Ernst

Resources Execution success Data Resource Usage Network Profiling Interactive Data Analysis Tools Usage Job + Result Steering Toolbox Grid User/Application Job Success Optimizer Service Application Specific Scheduling Service Query run-time collected information Information Service static & run-time collected scheduled / forecasted Query for resources Scheduling Service Reservation Resources Execution success Data Management Service Job Supervisor Service Data Resource Usage Maintain information Network Profiling Accounting and Billing Service Network Management Service Maintain information Compute Manager Data Manager Network Manager Management System Compute Management System Storage Management System Network Compute-Resources Data-Resources Network-Resources 4/05/2005 LCG-DM-WS M.Ernst

Extensions to SE Information Providers A “Cost Prediction Module” calculates load development of the local Storage Element (SE) and makes the information available using an agreed upon schema (i.e. GLUE, additional dynamic records req.) By using this Information the local SE can predict the time required to make requested data sets available, and publish it to the Grid Scheduler The local SE can calculate a point in time a collection of data sets can be optimally made available This approach allows to only schedule jobs to the Compute Element (CE) when the associated SE has flagged all required data sets online 4/05/2005 LCG-DM-WS M.Ernst

Extensions to improve Job/Data Co-Scheduling Incoming Jobs Grid Scheduler Query current load Jobs Scheduled Query load prediction “make ready” Query “optimized submission time” Local Site Load / Cost Compute Element Local Scheduler Prediction Module Storage Element Load Info Local Cache Tertiary Storage System (dCache) RED : modules to improve 4/05/2005 LCG-DM-WS M.Ernst

“The Mirror Cache” Almost all Tape Data on Mirror Cache Mirror Cache built from low cost Components Mirror Cache adds another level in Hierarchy (no cartridge handling penalties) Managed number of high performance streams between Mirror & Regular Cache Mirror Cache disks spin only if accessed Tape is treated as Archive 4/05/2005 LCG-DM-WS M.Ernst