CERN IT Department CH-1211 Genève 23 Switzerland www.cern.ch/i t HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 The CASTOR SRM2 Interface Status and plans.

Slides:



Advertisements
Similar presentations
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS CASTOR Status Alberto Pace.
Advertisements

Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS TSM CERN Daniele Francesco Kruse CERN IT/DSS.
CASTOR Upgrade, Testing and Issues Shaun de Witt GRIDPP August 2010.
CERN IT Department CH-1211 Genève 23 Switzerland t Messaging System for the Grid as a core component of the monitoring infrastructure for.
Web Application Architecture: multi-tier (2-tier, 3-tier) & mvc
Client/Server Grid applications to manage complex workflows Filippo Spiga* on behalf of CRAB development team * INFN Milano Bicocca (IT)
CERN IT Department CH-1211 Genève 23 Switzerland t Integrating Lemon Monitoring and Alarming System with the new CERN Agile Infrastructure.
19 February CASTOR Monitoring developments Theodoros Rekatsinas, Witek Pokorski, Dennis Waldron, Dirk Duellmann,
Castor F2F Meeting Barbara Martelli Castor Database CNAF.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
CERN IT Department CH-1211 Genève 23 Switzerland t Tape-dev update Castor F2F meeting, 14/10/09 Nicola Bessone, German Cancio, Steven Murray,
CERN IT Department CH-1211 Geneva 23 Switzerland t The Experiment Dashboard ISGC th April 2008 Pablo Saiz, Julia Andreeva, Benjamin.
CERN IT Department CH-1211 Genève 23 Switzerland t Plans and Architectural Options for Physics Data Analysis at CERN D. Duellmann, A. Pace.
CERN - IT Department CH-1211 Genève 23 Switzerland CASTOR Operational experiences HEPiX Taiwan Oct Miguel Coelho dos Santos.
CERN - IT Department CH-1211 Genève 23 Switzerland Service Level Status (SLS) What it does The default SLS.
CERN IT Department CH-1211 Genève 23 Switzerland t EIS section review of recent activities Harry Renshall Andrea Sciabà IT-GS group meeting.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES PhEDEx Monitoring Nicolò Magini CERN IT-ES-VOS For the PhEDEx.
CERN IT Department CH-1211 Geneva 23 Switzerland t Storageware Flavia Donno CERN WLCG Collaboration Workshop CERN, November 2008.
CERN IT Department CH-1211 Genève 23 Switzerland t Tier0 Status - 1 Tier0 Status Tony Cass LCG-LHCC Referees Meeting 18 th November 2008.
CERN IT Department CH-1211 Genève 23 Switzerland t Castor development status Alberto Pace LCG-LHCC Referees Meeting, May 5 th, 2008 DRAFT.
CERN - IT Department CH-1211 Genève 23 Switzerland Castor External Operation Face-to-Face Meeting, CNAF, October 29-31, 2007 CASTOR2 Disk.
CERN IT Department CH-1211 Genève 23 Switzerland t MSG status update Messaging System for the Grid First experiences
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
4/5/2007Data handling and transfer in the LHCb experiment1 Data handling and transfer in the LHCb experiment RT NPSS Real Time 2007 FNAL - 4 th May 2007.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks GStat 2.0 Joanna Huang (ASGC) Laurence Field.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
CERN IT Department CH-1211 Genève 23 Switzerland t Multithreading in CASTOR Experiences from a real life application Giuseppe Lo Presti (IT/DM)
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
CERN - IT Department CH-1211 Genève 23 Switzerland t CASTOR Status March 19 th 2007 CASTOR dev+ops teams Presented by Germán Cancio.
Report from CASTOR external operations F2F meeting held at RAL in February Barbara Martelli INFN - CNAF.
CERN IT Department CH-1211 Geneva 23 Switzerland t GDB CERN, 4 th March 2008 James Casey A Strategy for WLCG Monitoring.
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
WLCG Grid Deployment Board, CERN 11 June 2008 Storage Update Flavia Donno CERN/IT.
CERN SRM Development Benjamin Coutourier Shaun de Witt CHEP06 - Mumbai.
CERN IT Department CH-1211 Genève 23 Switzerland t Load Testing Dennis Waldron, CERN IT/DM/DA CASTOR Face-to-Face Meeting, Feb 19 th 2009.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
XROOTD AND FEDERATED STORAGE MONITORING CURRENT STATUS AND ISSUES A.Petrosyan, D.Oleynik, J.Andreeva Creating federated data stores for the LHC CC-IN2P3,
Accounting Update John Gordon and Stuart Pullinger January 2014 GDB.
The CMS Top 5 Issues/Concerns wrt. WLCG services WLCG-MB April 3, 2007 Matthias Kasemann CERN/DESY.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF Agile Infrastructure Monitoring HEPiX Spring th April.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS New tape server software Status and plans CASTOR face-to-face.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
Distributed Logging Facility Castor External Operation Workshop, CERN, November 14th 2006 Dennis Waldron CERN / IT.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF CF Monitoring: Lemon, LAS, SLS I.Fedorko(IT/CF) IT-Monitoring.
CERN IT Department CH-1211 Genève 23 Switzerland t CERN IT Monitoring and Data Analytics Pedro Andrade (IT-GT) Openlab Workshop on Data Analytics.
MND review. Main directions of work  Development and support of the Experiment Dashboard Applications - Data management monitoring - Job processing monitoring.
FTS monitoring work WLCG service reliability workshop November 2007 Alexander Uzhinskiy Andrey Nechaevskiy.
WLCG Grid Deployment Board CERN, 14 May 2008 Storage Update Flavia Donno CERN/IT.
SRM-2 Road Map and CASTOR Certification Shaun de Witt 3/3/08.
Enabling Grids for E-sciencE INFSO-RI Enabling Grids for E-sciencE Gavin McCance GDB – 6 June 2007 FTS 2.0 deployment and testing.
CERN IT Department CH-1211 Genève 23 Switzerland t SL(C) 5 Migration at CERN CHEP 2009, Prague Ulrich SCHWICKERATH Ricardo SILVA CERN, IT-FIO-FS.
CERN - IT Department CH-1211 Genève 23 Switzerland CASTOR F2F Monitoring at CERN Miguel Coelho dos Santos.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
The GridPP DIRAC project DIRAC for non-LHC communities.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF CC Monitoring I.Fedorko on behalf of CF/ASI 18/02/2011 Overview.
CERN IT Department CH-1211 Genève 23 Switzerland t CERN Agile Infrastructure Monitoring Pedro Andrade CERN – IT/GT HEPiX Spring 2012.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF Cluman: Advanced Cluster Management for Large-scale Infrastructures.
EGEE-II INFSO-RI Enabling Grids for E-sciencE WLCG File Transfer Service Sophie Lemaitre – Gavin Mccance Joint EGEE and OSG Workshop.
CERN IT Department CH-1211 Genève 23 Switzerland t Towards end-to-end debugging for data transfers Gavin McCance Javier Conejero Banon Sophie.
CERN IT Department CH-1211 Genève 23 Switzerland t DPM status and plans David Smith CERN, IT-DM-SGT Pre-GDB, Grid Storage Services 11 November.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
StoRM Architecture and Daemons
CASTOR-SRM Status GridPP NeSC SRM workshop
CERN-Russia Collaboration in CASTOR Development
Ákos Frohner EGEE'08 September 2008
Presentation transcript:

CERN IT Department CH-1211 Genève 23 Switzerland t HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 The CASTOR SRM2 Interface Status and plans Towards databases consolidation Shaun De Witt (RAL), Giuseppe Lo Presti (CERN/IT)

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 Outline The SRM2 interface to CASTOR –Architecture overview –Status Current deployment –Tier 0 –Tier 1s Future plans –For SRM –For CASTOR databases consolidation

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 A Foreword on SRM Storage Resource Manager –The Storage Element (SE) of the Grid –An interface to expose a storage system, hiding the internal details of the different implementations Different versions of the specifications exist –In particular v1.1 and v2.2 This talk focuses on the CASTOR implementation of the v2.2 specifications –In particular, including the WLCG MoU document that has been agreed amongst all WLCG SEs implementations

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 CASTOR SRM2 Architecture Database centric –Stateless and redundant daemons Ease deployment and scalability –Inherits most of the design choices of the CASTOR core system A very special CASTOR client –The SRM software does not interact directly with the internal components of CASTOR –Makes SRM decoupled from the evolution of the internal CASTOR APIs

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 Architecture Overview RDBMS Request Handler Asynch Processor CASTOR Backend Data transfer (e.g. via GridFTP v2)... SRM Clients

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 SRM Components Request Handler (frontend) –Performs user authentication GSI-based, can generate high CPU load –Serves all synchronous requests –Stores asynchronous requests in the database Asynch Processor (backend) –No user interaction Very low CPU footprint –Processes all asynchronous requests e.g. srmBringOnLine –Updates db by polling the CASTOR backend –Handles srmCopy operations

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 Current Status Currently in production: release 1.3 (SLC3) –Passed first production tests during CCRC’08 –Developed and supported by RAL –Many known issues Being deployed: release 2.7 (SLC4) –Major review of the code to address all critical issues faced in the past months –Developed by RAL and CERN –Towards a release process consistent with the CASTOR core software

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 Release 2.7 Key Features Support for SLC4 Support for multiple frontends and backends Support for the SRM methods agreed in the WLCG MoU addendum (May ‘08) –srmPurgeFromSpace to allow better handling of Disk1 storage classes –srmCopy fully functional including support of source space token –srmChangeSpaceForFiles definitely deprecated –improved srmAbortRequest Improved management of space tokens –Admin tools for the service manager –Space allocation remains static Improved logging Improved internal garbage collection –Runs as a database job

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 Current Deployment At the Tier0 –One production endpoint per VO, running rel. 1.3 –Pre-production endpoints per VO being setup, running rel. 2.7 Will eventually become the production endpoints At the Tier1s –ASGC: one production endpoint for all VOs Recently upgraded to rel. 2.7? –RAL, CNAF: one production endpoint per VO Running rel. 1.3, upgrade foreseen in the coming weeks

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 Plans and Future Directions Short-medium term –Deployment and support of release 2.7 –Stabilizing, addressing remaining less critical bugs –Improving logging for better end-to-end monitoring Long term vision –The choice of an SRM interface loosely coupled with the CASTOR backend proved to be not optimal Processing time often degraded due to the number of layers and interactions between components It is often impossible to achieve the desired behaviour under given borderline conditions (e.g. aborting ongoing transfers)

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 Plans and Future Directions Long term vision (cont’d) –A more coupled integration is foreseen, where the SRM frontend daemon directly interacts with the CASTOR main stager database This would enable the stager to seamlessly process CASTOR requests and SRM requests Most of the SRM logic will also be integrated in the CASTOR stager logic –SRM will be seen as another gateway to the CASTOR system, similarly to the current CASTOR Request Handler –The SRM and Stager database schemas will be merged

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 Towards DBs Consolidation The CASTOR system nowadays includes a number of database schemas –The CASTOR namespace –The CASTOR stager and disk cache management –The CASTOR SRM interface –The CASTOR VDQM (Volume and Drive Queue Manager) for the tape archive It is envisaged to review those schemas in order to improve all internal processing –E.g. the namespace, stager and SRM databases share some information and should be merged –The VDQM database should evolve to a Tape archive management database

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 Conclusions The CASTOR SRM interface is now a consolidated component of the CASTOR system –Used in production (CCRC) –New version being deployed A development plan has been devised –Stabilization of the production release –Tighter integration with the stager foreseen Comments, questions?

CERN IT Department CH-1211 Genève 23 Switzerland t HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 CASTOR Monitoring Witek Pokorski, Theodoros Rekatsinas (CERN/IT) Presented by Giuseppe Lo Presti

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 CASTOR end-to-end monitoring Motivation –Provide the users with a dashboard presenting the usage of a CASTOR instance –Provide the developers with a mean to monitor the running of the system from the point of view of the specific implementation Scope –Provide end-to-end monitoring covering all the CASTOR sub-components gather the information from SRM, Stager, Nameserver, Garbage collector, diskservers, tapeservers, etc, all together –Combine and extend currently existing monitoring infrastructure based on Lemon and SLS

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 CASTOR monitoring architecture all the monitoring information comes from log messages sent by different CASTOR components current implementation uses DLF (Distributed Logging Facility) monitoring DB designed in a way to allow quick queries from the web interface future implementation will be based on SYSLOG allowing more efficient real-time parsing of the incoming messages Monitoring DB SYSLOG PHP DLF DB SQL WEB Interface Python

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 CASTOR monitoring web interface

CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 Conclusion and plans With the massive data-taking approaching, monitoring is an important element of CASTOR –need to extend current monitoring A prototype of CASTOR monitoring is being implemented and is under testing –only statistic gathering now, dashboard functionalities to come Plans for the future developments include: –further extensions of the monitoring DB schema –move to syslog-based transport Comments, questions?