GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache Implementation at FZK Forschungszentrum Karlsruhe.

Slides:



Advertisements
Similar presentations
Bernd Panzer-Steindel, CERN/IT WAN RAW/ESD Data Distribution for LHC.
Advertisements

GridKa May 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Installing dCache into an existing Storage environment at GridKa Forschungszentrum.
HEPiX GFAL and LCG data management Jean-Philippe Baud CERN/IT/GD.
Owen SyngeDCache Slide 1 D-Cache Status Report Owen Synge.
GridKa January 2005 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann 1 Mass Storage at GridKa Forschungszentrum Karlsruhe GmbH.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.
Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft gridKa Forschungszentrum Karlsruhe GmbH Institute for Scientific Computing P.O.
EGEE is a project funded by the European Union under contract IST Using SRM: DPM and dCache G.Donvito,V.Spinoso INFN Bari
CASTOR SRM v1.1 experience Presentation at SRM meeting 01/09/2004, Berkeley Olof Bärring, CERN-IT.
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
DCache at Tier3 Joe Urbanski University of Chicago US ATLAS Tier3/Tier2 Meeting, Bloomington June 20, 2007.
A. Sim, CRD, L B N L 1 OSG Applications Workshop 6/1/2005 OSG SRM/DRM Readiness and Plan Alex Sim / Jorge Rodriguez Scientific Data Management Group Computational.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
BNL Facility Status and Service Challenge 3 Zhenping Liu, Razvan Popescu, Xin Zhao and Dantong Yu USATLAS Computing Facility Brookhaven National Lab.
Data Management The GSM-WG Perspective. Background SRM is the Storage Resource Manager A Control protocol for Mass Storage Systems Standard protocol:
Data management for ATLAS, ALICE and VOCE in the Czech Republic L.Fiala, J. Chudoba, J. Kosina, J. Krasova, M. Lokajicek, J. Svec, J. Kmunicek, D. Kouril,
Working together towards a next generation storage element Surya D. Pathak Advanced Computing Center for Research and Education.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
08/30/05GDM Project Presentation Lower Storage Summary of activity on 8/30/2005.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
 CASTORFS web page - CASTOR web site - FUSE web site -
LCG Storage workshop at CERN. July Geneva, Switzerland. BNL’s Experience dCache1.8 and SRM V2.2 Carlos Fernando Gamboa Dantong Yu RHIC/ATLAS.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
BNL Wide Area Data Transfer for RHIC & ATLAS: Experience and Plans Bruce G. Gibbard CHEP 2006 Mumbai, India.
BNL Facility Status and Service Challenge 3 HEPiX Karlsruhe, Germany May 9~13, 2005 Zhenping Liu, Razvan Popescu, and Dantong Yu USATLAS/RHIC Computing.
From Digital Objects to Content across eInfrastructures Content and Storage Management in gCube Pasquale Pagano CNR –ISTI on behalf of Heiko Schuldt Dept.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.
Copyright 2007, Information Builders. Slide 1 Machine Sizing and Scalability Mark Nesson, Vashti Ragoonath June 2008.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
Oct 24, 2002 Michael Ernst, Fermilab DRM for Tier1 and Tier2 centers Michael Ernst Fermilab February 3, 2003.
David Adams ATLAS ATLAS distributed data management David Adams BNL February 22, 2005 Database working group ATLAS software workshop.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
DMLite GridFTP frontend Andrey Kiryanov IT/SDC 13/12/2013.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
Martina Franca (TA), 07 November Installazione, configurazione, testing e troubleshooting di Storage Element.
9/20/04Storage Resource Manager, Timur Perelmutov, Jon Bakken, Don Petravick, Fermilab 1 Storage Resource Manager Timur Perelmutov Jon Bakken Don Petravick.
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
Service Challenge, 2. Dec Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Bruno Hoeft Progress at GridKa Forschungszentrum Karlsruhe GmbH.
Road map SC3 preparation
Jean-Philippe Baud, IT-GD, CERN November 2007
The Data Grid: Towards an architecture for Distributed Management
Status of the SRM 2.2 MoU extension
BNL Tier1 Report Worker nodes Tier 1: added 88 Dell R430 nodes
Data Bridge Solving diverse data access in scientific applications
dCache “Intro” a layperson perspective Frank Würthwein UCSD
Service Challenge 3 CERN
Abhishek Singh Rana UC San Diego
dCache Status and Plans – Proposals for SC3
GFAL 2.0 Devresse Adrien CERN lcgutil team
Computing Infrastructure for DAQ, DM and SC
Acutelearn Technologies Tivoli Storage Manager(TSM) Training Tivoli Storage Manager Basics: Tivoli Storage Manager Overview Tivoli Storage Manager concepts.
Introduction to Informer
Australia Site Report Sean Crosby DPM Workshop – 13 December 2013.
DIRAC Data Management: consistency, integrity and coherence of data
INFNGRID Workshop – Bari, Italy, October 2004
The LHCb Computing Data Challenge DC06
Presentation transcript:

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache Implementation at FZK Forschungszentrum Karlsruhe GmbH Institute for Scientific Computing P.O. Box 3640 D Karlsruhe, Germany Dr. Doris Ressmann

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann Introduction What is dCache? Pool Selection mechanism dCache properties LCG connection Access to dCache – connection to CERN Tape Management Conclusion

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann What is dCache? Developed at DESY and FNAL Disk pool management with or without tape backend Data may be distributed among a huge amount of disk servers. Automatic load balancing by cost metric and inter pool transfers. Data removed only if space is needed Fine grained configuration of pool attraction scheme

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann Pool Selection Mechanism Pool Selection required for: Pool selection is done in 2 steps – Query configuration database : → which pools are allowed for requested operation (intern/extern) – Query 'allowed pool' for their vital functions : → find pool with lowest cost for requested operation Client dCache Tape dCache dCache dCache Client

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann LCG Storage Element DESY dCap lib incorporates with CERN GFAL library SRM version ~ 1.1 supported gsiFtp supported

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann Multiple access of one file Pool 1 Pool 2Pool 3 File 1

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann Access to dCache Mountpoint –ls –mv –rm –checksum,…. dCap –dccp –dc_open(...) –dc_read(...) Gridftp –Problematic when file needs to be staged first SRMCP InternExtern

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache environment Internal nodes file transfer head node pools file transfer tape library

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache environment Internal nodes file transfer head node gsiftp srm pools file transfer tape library

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache environment Internal nodes file transfer head node srm pools file transfer tape library gsiftp

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache environment Internal nodes file transfer head node gsiftp srm pools file transfer tape library file transfer gsiftp

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache environment Internal nodes file transfer head node srmcp srm pools file transfer tape library srmcp

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache environment Internal nodes file transfer head node srmcp srm pools file transfer tape library srmcp

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache environment Internal nodes file transfer head node srmcp srm pools file transfer tape library srmcp

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann PNFS Perfectly Normal File System gdbm databases Experiment specific databases Independent access Content of metadata: – User file name – File name within dCache – Information about the tape location (storage class…) – Pool name where the file is located real data F A E pool and tape database for filenames metadata pnfs

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann gsiftp Only registered dCache user!!! grid-proxy-init globus-url-copy –dbg \ file:///tmp/file1 \ gsiftp://srm1.fzk.de/grid/fzk.de/mounts/pnfs/cms/file1 dCache gridftp client and server in Java copy direct into available pool node ● pool: data is precious ● (can't be deleted) ● flush into tape ● data is cached (can be deleted from pool)

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann srmcp Only registered dCache user!!! grid-proxy-init srmcp –debug=true \ srm://srm.web.cern.ch:80//castor/cern.ch/grid/dteam/castorfile \ srm://srm1.fzk.de:8443//pnfs/gridka.de/data/ressmann/file2 srmcp –debug=true \ srm://srm1.fzk.de:8443//pnfs/gridka.de/data/ressmann/file2 file:////tmp/file2

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann Firewall issues Connection to headnode: Ports 8443 and 2811 Port Range to pool nodes: to

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann SRM Disk Version FNAL is currently developing a standalone SRM Disk version. The client uses a java version of gridftp The server uses a standard globus gridftp. It is far from production ready and needs: – SQL Database – jdbc driver

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann Tape Management Tivoli Storage Manager (TSM) library management TSM is not developed for archive  Interruption of TSM archive  No control what has been archived

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache tape access Convenient HSM connectivity (done for Enstore, OSM, TSM, bad for HPSS) Creates a separate session for every file Transparent access Allows transparent maintenance at HSM

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache pool node 20 GB 1h 800 GB

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache tape management  Precious data is separately collected per 'storage class’  Each 'storage class queue ' has individual parameters, steering the tape flush operation.  Maximum time, a file is allowed to be 'precious' per 'storage class'.  Maximum number of precious bytes per 'storage class‚  Maximum number of precious files per 'storage class‚  Maximum number of simultaneous ‘tape flush' operations can be configured

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann Conclusion and Future Work Low cost read pools Reliable write pools Write once never change a dCache file Single point of failure Working SRM connection between CERN and FZK Connection to openlab at CERN Adding 15 Pool nodes for the 10 Gbit test from SRM to SRM More at

GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann