Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/i t DBCF GT Overview of DMLite Ricardo Rocha ( on behalf of the LCGDM team.

Slides:



Advertisements
Similar presentations
30-31 Jan 2003J G Jensen, RAL/WP5 Storage Elephant Grid Access to Mass Storage.
Advertisements

Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS CASTOR Status Alberto Pace.
Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
DPM Status & Roadmap Ricardo Rocha ( on behalf of the DPM team ) EMI INFSO-RI
HDFS and S3 plugins Andrea Manzi Martin Hellmich 13/12/2013.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT dpm-xrootd v3 Creating Federated Data Stores for the LHC David.
Adding scalability to legacy PHP web applications Overview Mario A. Valdez-Ramirez.
O. Keeble, F. Furano, A. Manzi, A. Ayllon, I. Calvet, M. Hellmich DPM Workshop 2014 DPM Monitoring.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Simplifying Configuration Ricardo Rocha ( on behalf of the LCGDM.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
SWE 316: Software Design and Architecture – Dr. Khalid Aljasser Objectives Lecture 11 : Frameworks SWE 316: Software Design and Architecture  To understand.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Performant and Future Proof: MySQL, Memcache and Raspberry Pi.
Data Management Kelly Clynes Caitlin Minteer. Agenda Globus Toolkit Basic Data Management Systems Overview of Data Management Data Movement Grid FTP Reliable.
Heads in the cloud? GSM-WG at OGF31, Taipei Jens Jensen, RAL.
DPM CCRC - 1 Research and developments DPM status and plans Jean-Philippe Baud.
CERN IT Department CH-1211 Geneva 23 Switzerland t The Experiment Dashboard ISGC th April 2008 Pablo Saiz, Julia Andreeva, Benjamin.
Ricardo Rocha ( on behalf of the DPM team ) Standards, Status and Plans.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES PhEDEx Monitoring Nicolò Magini CERN IT-ES-VOS For the PhEDEx.
Operating Systems & Information Services CERN IT Department CH-1211 Geneva 23 Switzerland t OIS Drupal Database Selection Tim Bell 6 th June.
CERN IT Department CH-1211 Geneva 23 Switzerland t Storageware Flavia Donno CERN WLCG Collaboration Workshop CERN, November 2008.
CERN IT Department CH-1211 Geneva 23 Switzerland t Daniel Gomez Ruben Gaspar Ignacio Coterillo * Dawid Wojcik *CERN/CSIC funded by Spanish.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
PVSSProxy The first piece of the MACS procedure framework (ProShell) Angela Brett.
CERN - IT Department CH-1211 Genève 23 Switzerland t COOL Conditions Database for the LHC Experiments Development and Deployment Status Andrea.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
Developing Applications with the CSI Framework A General Guide.
CERN SRM Development Benjamin Coutourier Shaun de Witt CHEP06 - Mumbai.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM Collaboration Motivation and proposal Oliver Keeble CERN On.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Ricardo Rocha CERN (IT/GS) EGEE’08, September 2008, Istanbul, TURKEY Experiment.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
INFSO-RI Enabling Grids for E-sciencE The gLite File Transfer Service: Middleware Lessons Learned form Service Challenges Paolo.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Interfaces to Grid Storage DPM and LFC Update Ricardo.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
CERN IT Department CH-1211 Geneva 23 Switzerland GT HTTP solutions for data access, transfer, federation Fabrizio Furano (presenter) on.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS New tape server software Status and plans CASTOR face-to-face.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Andrea Sciabà Hammercloud and Nagios Dan Van Der Ster Nicolò Magini.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
Operating Systems & Information Services CERN IT Department CH-1211 Geneva 23 Switzerland t OIS Drupal at CERN Juraj Sucik Jarosław Polok.
INFSO-RI Enabling Grids for E-sciencE SRMv2.2 in DPM Sophie Lemaitre Jean-Philippe.
Overview of C/C++ DB APIs Dirk Düllmann, IT-ADC Database Workshop for LHC developers 27 January, 2005.
DMLite GridFTP frontend Andrey Kiryanov IT/SDC 13/12/2013.
EMI is partially funded by the European Commission under Grant Agreement RI Roadmap & Future Work Ricardo Rocha ( on behalf of the DPM team )
Web Cache. What is Cache? Cache is the storing of data temporarily to improve performance. Cache exist in a variety of areas such as your CPU, Hard Disk.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Andrea Manzi CERN EGI Conference on Challenges and Solutions for Big Data Processing on cloud 24/09/2014 Storage Management Overview 1 24/09/2014.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF Cluman: Advanced Cluster Management for Large-scale Infrastructures.
EMI is partially funded by the European Commission under Grant Agreement RI Common Authentication Library Daniel Kouril, for the CaNL PT EGI CF.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES The Common Solutions Strategy of the Experiment Support group.
Security recommendations DPM Jean-Philippe Baud CERN/IT.
DPM: Future Proof Storage Ricardo Rocha ( on behalf of the DPM team ) EMI INFSO-RI
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Protocols in DPM Ricardo Rocha.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Solutions for WAN data access: xrootd and NFSv4.1 Andrea Sciabà.
EMI is partially funded by the European Commission under Grant Agreement RI DPM in EMI-II HTTP and NFS interfaces Oliver Keeble On behalf of DPM.
EMI is partially funded by the European Commission under Grant Agreement RI Future Proof Storage with DPM Oliver Keeble (on behalf of the CERN IT-GT-DMS.
CERN IT Department CH-1211 Genève 23 Switzerland t DPM status and plans David Smith CERN, IT-DM-SGT Pre-GDB, Grid Storage Services 11 November.
UNICORE and Argus integration Krzysztof Benedyczak ICM / UNICORE Security PT.
Jean-Philippe Baud, IT-GD, CERN November 2007
Ricardo Rocha ( on behalf of the DPM team )
DPM Installation Configuration
Status of the SRM 2.2 MoU extension
Dynafed, DPM and EGI DPM workshop 2016 Speaker: Fabrizio Furano
Introduction to Data Management in EGI
FTS Monitoring Ricardo Rocha
GFAL 2.0 Devresse Adrien CERN lcgutil team
Data Management cluster summary
INFNGRID Workshop – Bari, Italy, October 2004
Summary of the dCache workshop
Presentation transcript:

Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Overview of DMLite Ricardo Rocha ( on behalf of the LCGDM team ) EMI INFSO-RI

Grid Technology Reasoning for DMLite ~1.5 years ago we performed a full DPM evaluation – Using PerfSuite, out testing framework – – ( most results presented in the workshop come from this framework too ) It showed the system had significant bottlenecks – Performance – Code maintenance (and complexity) – Extensibility 2

Grid Technology Dependency on NS/DPM daemons All calls to the system had to go via the daemons – Not only user / client calls – Also the case for our frontends (HTTP/DAV, NFS, XROOT, …) – Daemons were a bottleneck, and did not scale well Short term fix (available since 1.8.2) – Improve TCP listening queue settings to prevent timeouts – Increase number of threads in the daemon pools Previously statically defined to a rather low value Medium term (available since 1.8.4, with DMLite) – Refactor the daemon code into a library 3

Grid Technology Dependency on NS/DPM daemons All calls to the system had to go via the daemons – Not only user / client calls – Also valid for our new frontends (HTTP/DAV, NFS, XROOT, …) – Daemons were a bottleneck, and did not scale well Short term fix (available since 1.8.2) – Improve TCP listening queue settings to prevent timeouts – Increase number of threads in the daemon pools Previously statically defined Medium term (available since 1.8.4, with DMLite) – Refactor the daemon code into a library 4

Grid Technology GET asynchronous performance DPM used to mandate asynchronous GET calls – Introduces significant client latency – Useful when some preparation of the replica is needed – But this wasn’t really our case (disk only) Fix (available with 1.8.3) – Allow synchronous GET requests DMLite has the same sync behavior (but faster ) 5

Grid Technology GET asynchronous performance DPM used to mandate asynchronous GET calls – Introduces significant client latency – Useful when some preparation of the replica is needed – But this wasn’t really our case (disk only) Fix – Allow synchronous GET requests 6

Grid Technology Database Access No DB connection pooling, no bind variables  – DB connections were linked to daemon pool threads – DB connections would be kept for the whole life of the client Quicker fix (available with 1.8.6) – Add DB connection pooling to the old daemons – Good numbers, but needed extensive testing… took some time Medium term fix (available since for HTTP/DAV) – DMLite, which includes connection pooling – Among many other things… 7

Grid Technology Database Access No DB connection pooling, no bind variables  – DB connections were linked to daemon pool threads – DB connections would be kept for the whole life of the client Quicker fix – Add DB connection pooling to the old daemons – Good numbers, but needs extensive testing… Medium term fix (available since for HTTP/DAV) – DMLite, which includes connection pooling – Among many other things… 8

Grid Technology Dependency on the SRM SRM imposes significant latency for data access – It has its use cases, but is a killer for regular file access – For data access, only required for protocols not supporting redirection (file name to replica translation) Fix (all available from 1.8.4) – Keep SRM for space management only (usage, reports, …) – Add support for protocols natively supporting redirection HTTP/DAV, NFS 4.1/pNFS, XROOT And promote them widely… Investigating GridFTP redirection support (seems possible!) 9

Grid Technology Future Proof with DMLite

Grid Technology Future Proof with DMLite DMLite is our new plugin based library Meets goals resulting from the system evaluation – Refactoring of the existing code – Single library used by all frontends – Extensible, open to external contributions – Easy integration of standard building blocks Apache2, HDFS, S3, … 11

Grid Technology DMLite is our new plugin based library Meets goals resulting from the system evaluation – Refactoring of the existing code – Single library used by all frontends – Extensible, open to external contributions – Easy integration of standard building blocks Apache, HDFS, S3, … Future Proof with DMLite 12

Grid Technology DMLite is a single library used by all DPM components In production today – Already used by HTTP/DAV, soon by all frontends We’ve opened DPM to other systems – Many widely used in the industry (HDFS, S3, …) – And the work has just started Clean, well defined interfaces – And APIs in different languages, much easier to contribute Performance improved drastically! Plugin details come next… 13 Summary and Status