EOS, DPM and FTS developments and plans

Slides:



Advertisements
Similar presentations
30-31 Jan 2003J G Jensen, RAL/WP5 Storage Elephant Grid Access to Mass Storage.
Advertisements

Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
HDFS and S3 plugins Andrea Manzi Martin Hellmich 13/12/2013.
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
WebFTS as a first WLCG/HEP FIM pilot
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
M i SMob i S Mob i Store - Mobile i nternet File Storage Platform Chetna Kaur.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
The Replica Location Service The Globus Project™ And The DataGrid Project Copyright (c) 2002 University of Chicago and The University of Southern California.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Interfaces to Grid Storage DPM and LFC Update Ricardo.
CERN IT Department CH-1211 Geneva 23 Switzerland GT HTTP solutions for data access, transfer, federation Fabrizio Furano (presenter) on.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Overview of DMLite Ricardo Rocha ( on behalf of the LCGDM team.
1 Xrootd-SRM Andy Hanushevsky, SLAC Alex Romosan, LBNL August, 2006.
DMLite GridFTP frontend Andrey Kiryanov IT/SDC 13/12/2013.
EMI is partially funded by the European Commission under Grant Agreement RI Roadmap & Future Work Ricardo Rocha ( on behalf of the DPM team )
CASTOR in SC Operational aspects Vladimír Bahyl CERN IT-FIO 3 2.
Andrea Manzi CERN EGI Conference on Challenges and Solutions for Big Data Processing on cloud 24/09/2014 Storage Management Overview 1 24/09/2014.
An Analysis of Data Access Methods within WLCG Shaun de Witt, Andrew Lahiff (STFC)
DPM: Future Proof Storage Ricardo Rocha ( on behalf of the DPM team ) EMI INFSO-RI
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Protocols in DPM Ricardo Rocha.
EMI is partially funded by the European Commission under Grant Agreement RI Future Proof Storage with DPM Oliver Keeble (on behalf of the CERN IT-GT-DMS.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS CASTOR and EOS status and plans Giuseppe Lo Presti on behalf.
Riccardo Zappi INFN-CNAF SRM Breakout session. February 28, 2012 Ingredients 1. Basic ingredients (Fabric & Conn. level) 2. (Grid) Middleware ingredients.
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
CERN IT-Storage Strategy Outlook Alberto Pace, Luca Mascetti, Julien Leduc
EGEE Data Management Services
Federating Data in the ALICE Experiment
PaaS services for Computing and Storage
Onedata Eventually Consistent Virtual Filesystem for Multi-Cloud Infrastructures Michał Orzechowski (CYFRONET AGH)
Jean-Philippe Baud, IT-GD, CERN November 2007
WLCG IPv6 deployment strategy
DPM at ATLAS sites and testbeds in Italy
Dynamic Storage Federation based on open protocols
CASTOR Giuseppe Lo Presti on behalf of the CASTOR dev team
Ricardo Rocha ( on behalf of the DPM team )
StoRM: a SRM solution for disk based storage systems
Vincenzo Spinoso EGI.eu/INFN
Status of the SRM 2.2 MoU extension
Section 6 Object Storage Gateway (RADOS-GW)
BNL Tier1 Report Worker nodes Tier 1: added 88 Dell R430 nodes
Future of WAN Access in ATLAS
Data Bridge Solving diverse data access in scientific applications
dCache “Intro” a layperson perspective Frank Würthwein UCSD
Service Challenge 3 CERN
dCache – protocol developments and plans
StoRM Architecture and Daemons
Gfal/lcg-util -> Gfal2/gfal2-util
Introduction to Data Management in EGI
Taming the protocol zoo
DPM Python tools Andrea Manzi CERN DPM Workshop 07th December 2015.
GFAL 2.0 Devresse Adrien CERN lcgutil team
DPM releases and platforms status
Ákos Frohner EGEE'08 September 2008
The INFN Tier-1 Storage Implementation
DCache things Paul Millar … on behalf of the dCache team.
DPM and SRM-less operation
CTA: CERN Tape Archive Overview and architecture
Data Management cluster summary
20409A 7: Installing and Configuring System Center 2012 R2 Virtual Machine Manager Module 7 Installing and Configuring System Center 2012 R2 Virtual.
UMD update 11 Joao Pina (EGI/LIP).
INFNGRID Workshop – Bari, Italy, October 2004
Rucio & objectstores James Perry.
Presentation transcript:

EOS, DPM and FTS developments and plans Andrea Manzi - on behalf of the IT Storage Group, AD section HEPIX Fall 2016 Workshop – LBNL 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans Outline CERN IT Storage group, AD section EOS Namespace on Redis DPM DOME FTS New optimizer Object Stores Integration 19/10/16 EOS, DPM and FTS developments and plans

CERN IT-ST group, AD Section 16 members Main activities Development EOS, DPM, FTS, Data management clients (Davix, gfal2, Xrootd client) Operations FTS Analytics WG Effort in WLCG ops 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans Outline CERN IT Storage group, AD section EOS Namespace on Redis DPM DOME FTS New optimizer Object Stores Integration 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans EOS architecture Disk only file storage Designed for Massive storage with high performances ( > 50PB, 500 M files and Khz metadata operations) In memory hierarchical namespace File layouts (default 2 replicas) Low latency access 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans 1260 43170 19/10/16 EOS, DPM and FTS developments and plans

EOS World-Wide Deployment ASGC AARNET CERN 19/10/16 EOS, DPM and FTS developments and plans

EOS releases and branches Production version Branch: beryl_aquamarine Release number: >= 0.3.210 Development version (master) Branch: citrine Release number: >= 4.1.4 Requires XRootD 4.4.0 Feature branches get merged into master e.g. kinetic, geo-scheduling, namespace devel. etc. 19/10/16 EOS, DPM and FTS developments and plans

What is the EOS namespace? C++ library used by the EOS MGM node single-threaded Provides API for dealing with hierarchical collections of files Filesystem elements Containers & files Views Aggregate info about filesystem elem. E.g QuotaView, FileSystemView etc. Persistence objects Objects responsible for reading and storing filesystem elements Implemented as binary change-logs 19/10/16 EOS, DPM and FTS developments and plans

Namespace architectures pros/cons Using hashes all in memory  extremely fast Every change is logged  low risk of data loss Views rebuilt at each boot  high consistency Cons: For big instances it requires a lot of RAM Booting the namespace from the change-log takes long 19/10/16 EOS, DPM and FTS developments and plans

EOS Namespace Interface Prepare the setting for different namespace implementations Abstract a Namespace Interface to avoid modifying other parts of the code EOS citrine 4.* Plugin manager – able not only to dynamically load but also stack plugins if necessary libEosNsInMemory.so – the original in-memory namespace implementation libEosNsOnFilesystem.so – not existing based on a Linux filesystem 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans Why Redis? Redis – in-memory data structure store Separate data from the application logic and user interface Supports various data structures: strings, hashes, lists, sets, sorted sets etc. Namespace implementation: libEosOnRedis.so 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans XRootD and Redis Replace Redis backend with XRootD Implemented as an XRootD protocol plugin – to be contributed upstream XRootD can use RocksDB as persistent key-value store 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans Namespace HA Ensure high-availability using the Raft consensus algorithm 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans Outline CERN IT Storage group , AD section EOS Namespace on Redis DPM DOME FTS New optimizer Object stores Integration 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans Disk Pool Manager DPM is a system for managing disk storage at small and medium sites Over 70PB stored in the system at around 150 sites DPM has just finished a little revolution. We now have our definitive platform for the future SRM-less operations Caching 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans The 3rd Generation DPM 1st generation – derived from Castor 2nd generation – introduced dmlite Internal framework abstracting many functions, enabling multiple new frontends 3rd generation – introduces Dome DPM 1.9 “Legacy flavour” Legacy services still running DPM 1.9 “Dome flavour” Legacy services retired No more dpmd, dpns, rfio, srm, csec 19/10/16 EOS, DPM and FTS developments and plans

DPM Headnode architecture DMLite section dmlite core WebDAV Xrootd mysql profiler adapter gsiftp mem cache Legacy daemons rfio dpns srm Legacy clients CSec dpm (daemon) Communication with DPMd can now be dropped DOME Adapter DOME 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans DPM 1.9 “Dome flavour” Activate Dome, what do you get? SRM free operation Quotas “space token” concept generalised and mapped into namespace Space reporting – used/free via HTTP/DAV Reporting on “space tokens” and subdirectories GridFTP redirection enables scalability Caching hooks Simplified system All internal communication over HTTP Control and data tunneling Improved dmlite-shell 19/10/16 EOS, DPM and FTS developments and plans

Disk server (simplified) DOME Apache HTTPd Requests to HEAD Timed Logic xrootd Workers mod_proxy_fcgi gsiftp /domedisk/… AuthN Ext Pull() requests (configurable script) Task Executor Request logic Dmlite: - IO - DiskCatalog lcgdm_dav /<diskpaths>/… Dome_checksum Script (bundled with DOME) Disks Dome can do IO tunnelling if clients don’t support redirection

Used/free space: WebDAV $ davix-http -P grid -X PROPFIND --header 'Depth: 0' --header 'Content-Type: text/xml; charset=UTF-8' "https://domehead-trunk.cern.ch/dpm/cern.ch/home/dteam" --data '<?xml version="1.0" ?><D:propfind xmlns:D="DAV:"><D:prop><D:quota-used-bytes/><D:quota-available-bytes/></D:prop></D:propfind>’ <?xml version="1.0" encoding="utf-8"?> <D:multistatus xmlns:D="DAV:" xmlns:ns0="DAV:"> <D:response xmlns:lp1="DAV:" xmlns:lp2="http://apache.org/dav/props/" xmlns:lp3="LCGDM:"> <D:href>/dpm/cern.ch/home/dteam/</D:href> <D:propstat> <D:prop> <lp1:quota-used-bytes>24677181319</lp1:quota-used-bytes> <lp1:quota-available-bytes>75322818681</lp1:quota-available-bytes> </D:prop> <D:status>HTTP/1.1 200 OK</D:status> </D:propstat> </D:response> </D:multistatus> 19/10/16 EOS, DPM and FTS developments and plans

Going SRM-less with your DPM ADM Configures quotatokens AND primes the space calcs DmLite::DomeAdapter substitutes Adapter (config change) Experiments gradually stop using SRM 1.9.0 in EPEL Site updates the system Load on SRM goes to zero Sysadmin Experiments DPM team …continuous support… Q4/2016 Good moment for a minor release DPM/dpns/rfio still used (through DMLite::Adapter and SRM) DOME used only for dmlite-shell New powerful dmlite-shell is available Can report used space per path Free space reporting is like before (Only SRM can report on free spc) All the DOME features are enabled WebDAV reports used AND free spc per path Quotas are properly enforced Can create the ATLAS space report SRM/DPM still report on spacetokens SRM/DPM/dpns/rfio are now optional DPM/dpns/rfio/SRM can be uninstalled

EOS, DPM and FTS developments and plans Caching laboratory DPM 1.9 with Dome will allow investigation of operating WLCG storage as a cache Scenarios Data origin a local federation of associate sites Data origin the global federation Hybrid cache/conventional setup A volatile pool can be defined which calls out to a stager on a miss Caching logic implemented in a pluggable way Questions to investigate Cache management logic Different client strategies on miss blocking read, async read, redirection to origin Authentication solutions Workflow adaptation for locality 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans Outline CERN IT Storage group , AD section EOS Namespace DPM DOME FTS New optimizer Object stores Integration 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans FTS is the service responsible for distributing the majority of LHC data across the WLCG infrastructure Is a low level data movement service, responsible for moving sets of files from one site to another while allowing participating sites to control the network resource usage WLCG stats: Installed at: CERN, RAL, BNL, FNAL ~20PB monthly volume / ~2.2M files per day 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans FTS Architecture Optimizer Server Server Scheduler Bring online Bring online … FTS-Url-Copy REST REST Gfal2 plugins MySQL 19/10/16 EOS, DPM and FTS developments and plans

FTS 3.5: Optimizer changes Min/max active ranges can be configured per link Improved throughput calculation Softener Exponential Moving Average to reduce noise effects Throughput taken into account even with high success rates 1 stream per transfer by default Reduces resource consumption on the storages 19/10/16 EOS, DPM and FTS developments and plans

FTS 3.5: Optimizer evolution 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans Plans for FTS 3.6 Targeted towards Jan 2017 Remove SOAP Only CMS is using it and it’s already migrating to REST Remove Bulk Deletions New algorithm for multiple replicas jobs Database profiling and optimizations 19/10/16 EOS, DPM and FTS developments and plans

Object stores Integration Advantages Scalability and performance achieved through relaxing or abandoning many aspects of posix Applications must be aware or adapted How can such resources be plugged into existing WLCG workflows? Can apply to public or private cloud NB ceph at sites Data transfer -> FTS integration via davix/gfal2 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans davix davix-put /etc/services https://objbkt1.s3.amazonaws.com/file01 --s3secretkey <secret> --s3accesskey <access> davix-cp -P grid davs://dpm.cern.ch/dpm/cern.ch/home/dteam/file01 s3s://objbkt1.s3.amazonaws.com/file01 --s3secretkey <secret> --s3accesskey <access> 3rd party copy!! 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans gfal2/davix gfal-copy file:///etc/services s3://objbkt1.s3.amazonaws.com/file01 gfal-copy davs://dpm.cern.ch/dpm/cern.ch/home/dteam/file01 s3://objbkt1.s3.amazonaws.com/file01 3rd party copy!! 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans FTS: Pre-signed URL fts-transfer-submit --strict-copy -s https://fts3.cern.ch:8446 https://dpm.cern.ch/dpm/cern.ch/home/dteam/file01 'https://objbkt1.s3.amazonaws.com/tf_04?Signature=eFAyXMWlSY%2BWEVcqfvGvuxZF6ZQ%3D&Expires=2105774242&AWSAccessKeyId=AKIAJZZQ2TYSEBKNVWKA' 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans FTS: key management You can also allow FTS to hold the keys to your cloud storage $curl[…] https://fts3devel01.cern.ch:8446/config/cloud_storage -H "Content-Type: application/json" -X POST -d '{"storage_name":"S3:s3.domain.com"}‘ $curl[…]"https://fts3devel01.cern.ch:8446/config/cloud_storage/S3:s3.domain.com" -H "Content-Type: application/json" -X POST -d "${config}“ { "vo_name": "dteam", "access_key": "ACCESS_KEY", "secret_key": "SECRET_KEY" } 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans FTS: transport Solutions for import to and export from clouds Several S3 variants supported Various architectures possible FTS gateway SRM<->S3 3rd party transfer Multi-hop with tactical storage FTS Cloud Storage (S3 varieties) Existing Storage Tactical Storage, Multihop 19/10/16 EOS, DPM and FTS developments and plans

EOS, DPM and FTS developments and plans References EOS http://eos.web.cern.ch/ DPM http://lcgdm.web.cern.ch/ DPM Workshop 2016, 23-24 Nov, Paris https://indico.cern.ch/event/559673/ FTS http://fts3-service.web.cern.ch/ 19/10/16 EOS, DPM and FTS developments and plans