Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/i t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.

Slides:



Advertisements
Similar presentations
HEPiX GFAL and LCG data management Jean-Philippe Baud CERN/IT/GD.
Advertisements

Wahid Bhimji SRM; FTS3; xrootd; DPM collaborations; cluster filesystems.
Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
ATLAS Tier-3 in Geneva Szymon Gadomski, Uni GE at CSCS, November 2009 S. Gadomski, ”ATLAS T3 in Geneva", CSCS meeting, Nov 091 the Geneva ATLAS Tier-3.
HDFS and S3 plugins Andrea Manzi Martin Hellmich 13/12/2013.
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT dpm-xrootd v3 Creating Federated Data Stores for the LHC David.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Simplifying Configuration Ricardo Rocha ( on behalf of the LCGDM.
Operating Systems & Infrastructure Services CERN IT Department CH-1211 Geneva 23 Switzerland t OIS CERN Search Updates Eduardo Alvarez November.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Performant and Future Proof: MySQL, Memcache and Raspberry Pi.
ATLAS DQ2 Deletion Service D.A. Oleynik, A.S. Petrosyan, V. Garonne, S. Campana (on behalf of the ATLAS Collaboration)
INFSO-RI Enabling Grids for E-sciencE The US Federation Miron Livny Computer Sciences Department University of Wisconsin – Madison.
StoRM Some basics and a comparison with DPM Wahid Bhimji University of Edinburgh GridPP Storage Workshop 31-Mar-101Wahid Bhimji – StoRM.
Ricardo Rocha ( on behalf of the DPM team ) Standards, Status and Plans.
CERN IT Department CH-1211 Geneva 23 Switzerland t Storageware Flavia Donno CERN WLCG Collaboration Workshop CERN, November 2008.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
CERN IT Department CH-1211 Genève 23 Switzerland t Castor development status Alberto Pace LCG-LHCC Referees Meeting, May 5 th, 2008 DRAFT.
Author - Title- Date - n° 1 Partner Logo WP5 Summary Paris John Gordon WP5 6th March 2002.
MW Readiness Verification Status Andrea Manzi IT/SDC 21/01/ /01/15 2.
RAL Site Report Castor Face-to-Face meeting September 2014 Rob Appleyard, Shaun de Witt, Juan Sierra.
Towards a Global Service Registry for the World-Wide LHC Computing Grid Maria ALANDES, Laurence FIELD, Alessandro DI GIROLAMO CERN IT Department CHEP 2013.
GLUE 2 Open Issues in Storage Information Providers 16 th May 2014.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
INFSO-RI Enabling Grids for E-sciencE gLite Data Management and Interoperability Peter Kunszt (JRA1 DM Cluster) 2 nd EGEE Conference,
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM Collaboration Motivation and proposal Oliver Keeble CERN On.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
DPM Python tools Ivan Calvet IT/SDC-ID DPM Workshop 10 th October 2014.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Interfaces to Grid Storage DPM and LFC Update Ricardo.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
CERN IT Department CH-1211 Geneva 23 Switzerland GT HTTP solutions for data access, transfer, federation Fabrizio Furano (presenter) on.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
INFSO-RI Enabling Grids for E-sciencE SRMv2.2 in DPM Sophie Lemaitre Jean-Philippe.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Overview of DMLite Ricardo Rocha ( on behalf of the LCGDM team.
SRM-2 Road Map and CASTOR Certification Shaun de Witt 3/3/08.
DMLite GridFTP frontend Andrey Kiryanov IT/SDC 13/12/2013.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
Enabling Grids for E-sciencE INFSO-RI Enabling Grids for E-sciencE Gavin McCance GDB – 6 June 2007 FTS 2.0 deployment and testing.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Data Management cluster summary David Smith JRA1 All Hands meeting, Catania, 7 March.
Testing Infrastructure Wahid Bhimji Sam Skipsey Intro: what to test Existing testing frameworks A proposal.
EMI is partially funded by the European Commission under Grant Agreement RI Roadmap & Future Work Ricardo Rocha ( on behalf of the DPM team )
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
Andrea Manzi CERN EGI Conference on Challenges and Solutions for Big Data Processing on cloud 24/09/2014 Storage Management Overview 1 24/09/2014.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF Cluman: Advanced Cluster Management for Large-scale Infrastructures.
Security recommendations DPM Jean-Philippe Baud CERN/IT.
DPM: Future Proof Storage Ricardo Rocha ( on behalf of the DPM team ) EMI INFSO-RI
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Protocols in DPM Ricardo Rocha.
EMI is partially funded by the European Commission under Grant Agreement RI DPM in EMI-II HTTP and NFS interfaces Oliver Keeble On behalf of DPM.
EMI is partially funded by the European Commission under Grant Agreement RI Future Proof Storage with DPM Oliver Keeble (on behalf of the CERN IT-GT-DMS.
CERN IT Department CH-1211 Genève 23 Switzerland t DPM status and plans David Smith CERN, IT-DM-SGT Pre-GDB, Grid Storage Services 11 November.
Dynamic Storage Federation based on open protocols
Ricardo Rocha ( on behalf of the DPM team )
DPM Installation Configuration
CREAM Status and Plans Massimo Sgaravatto – INFN Padova
Taming the protocol zoo
DPM Python tools Andrea Manzi CERN DPM Workshop 07th December 2015.
GFAL 2.0 Devresse Adrien CERN lcgutil team
DPM releases and platforms status
DCache things Paul Millar … on behalf of the dCache team.
Data Management cluster summary
INFNGRID Workshop – Bari, Italy, October 2004
Presentation transcript:

Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the LCGDM team ) EMI INFSO-RI

CERN IT Department CH-1211 Geneva 23 Switzerland t GT Releases We are now able to release more often – 1 month between and – Probably still in December Smaller releases, less intrusive Available early in epel-testing (more later) But we can even do more – Independent component releases – DMLite plugins, individual frontends – Should we keep a global number anyway? 2

CERN IT Department CH-1211 Geneva 23 Switzerland t GT What’s there… now (1.8.5) GLUE2 info publishing New endpoints publishing on the infosys – HTTP/DAV, XROOT Some small bug fixes 3

CERN IT Department CH-1211 Geneva 23 Switzerland t GT What’s coming… now (1.8.6) Fix for a report on a security vulnerability Database connection pooling on the old daemons – Default enabled only on the LFC Other small bug fixes Beta: Plugin taking into account current I/O load – More later 4

CERN IT Department CH-1211 Geneva 23 Switzerland t GT What’s coming… next (1.8.7) Update for GridFTP/XROOTD/NFS frontends – With versions using DMLite – Faster… and able to handle multiple pool types Dropping pool accounts requirement – We never actually make use of them Update of DMLite – Improved replica handling methods, lots of bug fixes Beta state – Tool for hot file replication / rebalancing – Plugins for HDFS and S3 support (already there since a while, but first official release) 5

CERN IT Department CH-1211 Geneva 23 Switzerland t GT Some features worth discussing

CERN IT Department CH-1211 Geneva 23 Switzerland t GT Concurrency, rebalancing… Some sites have seen issues when under load Too many files on a single disk server – DPM redirects a high number of requests there – Filesystem then blocks under I/O wait Solutions – Poor man: DMLite plugin taking into account recent disk server access history (for replica selection) – Better: DMLite tool taking care of rebalancing the nodes (obsoletes filesystem weights?) – Also better: DMLite tool taking care of replicating hot files 7

CERN IT Department CH-1211 Geneva 23 Switzerland t GT GridFTP redirection HTTP, XROOT, NFS… all support client redirection – Which fits our model… separate metadata from data GridFTP doesn’t… or our plugin doesn’t – Not a problem if using the SRM… but bad if not! – But the specification does (or at least it looks like it!) Possibility 1: SPAS (striped access) – Should be enough for redirection – Should give us parallel data access too – Client support? Possibility 2: GridFTP v2 extensions – dCache seems to support extensions for this – Globus on the way to implement them – The calls should look the same as for SPAS 8

CERN IT Department CH-1211 Geneva 23 Switzerland t GT Preferred I/O protocols RFIO is still the default almost everywhere – But we’ve seen the performance numbers – We should slowly phase it out in favor of HTTP or XROOT (or later NFS) Not an issue when clients directly use protocols – Not the case when requesting via the SRM Thanks to – ASGC / ShuTing for the numbers – And UK / Wahid / Sam for the XROOT federation testing We now need larger scale deployment 9

CERN IT Department CH-1211 Geneva 23 Switzerland t GT Other I/O optimizations To allow HTTP as an alternative – Together with ASGC/ShuTing, we added support for it in Panda (ATLAS analysis) – ShuTing also following with Athena for some additional fixes required – Added HTTPS support in ROOT But we still need to improve ROOT HTTP further – Add support for metalink, for replica failover and parallel reads – Add multi-stream access (read and write) 10

CERN IT Department CH-1211 Geneva 23 Switzerland t GT VFS Plugin Number one plugin request for DMLite Idea – Expose a locally mounted clustered filesystem via all the grid protocols, with grid auth/authz, etc Implementation – DMLite plugin implementing Catalog and I/O – Redirection to gateways for the I/O Status – Catalog implementation is there – It could be picked up by anyone (even with no knowledge of DMLite) and finalised Volunteers? 11

CERN IT Department CH-1211 Geneva 23 Switzerland t GT Collaboration opportunities Check here Lots and growing… you can help! Really!! – Testing of new pool types (HDFS, S3, …) – Improved log analyzer – VFS plugin – Pool managing in the dmlite shell – Better drain process with dmlite tools – Dark data detection tool (integration) – Evaluating mysqltuner.pl – … Standard Interfaces to Grid Storage – DPM and LFC Update - 12

CERN IT Department CH-1211 Geneva 23 Switzerland t GT Summary Expect more frequent releases – But you don’t have to update each time Refactoring effort is over – A lot is in production, the rest just around the corner Next couple of months focus will be in the long standing fixes for I/O access issues Lots of parallel work on tools and plugins – You don’t need to know many details to contribute – If you have an idea/request, have a look at the APIs 13