DCache in the NDGF Distributed Tier 1 Gerd Behrmann Second dCache Workshop Hamburg, 18 th of January 2007.

Slides:



Advertisements
Similar presentations
Grids and Grid Technologies for Wide-Area Distributed Computing Mark Baker, Rajkumar Buyya and Domenico Laforenza.
Advertisements

CSC Grid Activities Arto Teräs HIP Research Seminar February 18th 2005.
E-Infrastructure hierarchy Networking and Computational facilities in Armenia ASNET AM Network Armenian National Grid Initiative Armenian ATLAS site (AM-04-YERPHI)
NORDUnet NORDUnet The Fibre Generation Lars Fischer CTO NORDUnet.
Dynamic Firewalls and Service Deployment Models for Grid Environments Gian Luca Volpato, Christian Grimm RRZN – Leibniz Universität Hannover Cracow Grid.
Data Management The GSM-WG Perspective. Background SRM is the Storage Resource Manager A Control protocol for Mass Storage Systems Standard protocol:
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Interoperability Grids, Clouds and Collaboratories Ruth Pordes Executive Director Open Science Grid, Fermilab.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
August 3, March, The AC3 GRID An investment in the future of Atlantic Canadian R&D Infrastructure Dr. Virendra C. Bhavsar UNB, Fredericton.
NORDUnet Nordic Infrastructure for Research & Education Workshop Introduction - Finding the Match Lars Fischer LHCONE Workshop CERN, December 2012.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
INFSO-RI Enabling Grids for E-sciencE The gLite File Transfer Service: Middleware Lessons Learned form Service Challenges Paolo.
Alex Read, Dept. of Physics Grid Activities in Norway R-ECFA, Oslo, 15 May, 2009.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Julia Andreeva on behalf of the MND section MND review.
NORDUnet NORDUnet e-Infrastrucure: Grids and Hybrid Networks Lars Fischer CTO, NORDUnet Fall 2006 Internet2 Member Meeting, Chicago.
NeST: Network Storage John Bent, Venkateshwaran V Miron Livny, Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau.
DMLite GridFTP frontend Andrey Kiryanov IT/SDC 13/12/2013.
Enabling Grids for E-sciencE INFSO-RI Enabling Grids for E-sciencE Gavin McCance GDB – 6 June 2007 FTS 2.0 deployment and testing.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Store and exchange data with colleagues and team Synchronize multiple versions of data Ensure automatic desktop synchronization of large files B2DROP is.
Enabling Grids for E-sciencE EGEE-II INFSO-RI Status of SRB/SRM interface development Fu-Ming Tsai Academia Sinica Grid Computing.
INFSO-RI Enabling Grids for E-sciencE Turkish Tier-2 Site Report Emrah AKKOYUN High Performance and Grid Computing Center TUBITAK-ULAKBIM.
NDGF – a Joint Nordic Production Grid Lars Fischer ICFA Workshop on HEP Networking, Grid, and Digital Divide Issues for Global e-Science Cracow, 2 October.
Grid Services for Digital Archive Tao-Sheng Chen Academia Sinica Computing Centre
Riccardo Zappi INFN-CNAF SRM Breakout session. February 28, 2012 Ingredients 1. Basic ingredients (Fabric & Conn. level) 2. (Grid) Middleware ingredients.
NDGF and the distributed Tier-I Center Michael Grønager, PhD Technical Coordinator, NDGF dCahce Tutorial Copenhagen, March the 27th, 2007.
A Nordic Tier-1 for LHC Mattias Wadenstein Systems Integrator, NDGF Grid Operations Workshop Stockholm, June the 14 th, 2007.
A Distributed Tier-1 for WLCG Michael Grønager, PhD Technical Coordinator, NDGF CHEP 2007 Victoria, September the 3 rd, 2007.
Ian Bird, CERN WLCG Project Leader Amsterdam, 24 th January 2012.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
Enabling Grids for E-sciencE EGEE-II INFSO-RI The Development of SRM interface for SRB Fu-Ming Tsai Academia Sinica Grid Computing.
CASTOR: possible evolution into the LHC era
Jean-Philippe Baud, IT-GD, CERN November 2007
Accessing the VI-SEEM infrastructure
Gene Oleynik, Head of Data Storage and Caching,
NDGF The Distributed Tier1 Site.
Status of BESIII Distributed Computing
WP18, High-speed data recording Krzysztof Wrona, European XFEL
Introduction to Distributed Platforms
StoRM: a SRM solution for disk based storage systems
Status of the SRM 2.2 MoU extension
LCG 3D Distributed Deployment of Databases
James Casey, IT-GD, CERN CERN, 5th September 2005
dCache “Intro” a layperson perspective Frank Würthwein UCSD
Open Source distributed document DB for an enterprise
Experiences with http/WebDAV protocols for data access in high throughput computing
dCache Scientific Cloud
SRM Developers' Response to Enhancement Requests
DEISA : integrating HPC infrastructures in Europe Prof
Clouds of JINR, University of Sofia and INRNE Join Together
CMS analysis job and data transfer test results
LQCD Computing Operations
Introduction to Networks
HEPiX IPv6 Working Group F2F Meeting
Data Management cluster summary
Management of Virtual Execution Environments 3 June 2008
Chapter 3: Windows7 Part 4.
An Introduction to Computer Networking
Pierre Girard ATLAS Visit
Grid Coordination by Using the Grid Coordination Protocol
INFNGRID Workshop – Bari, Italy, October 2004
Item 2.2 of the agenda IT Working Group meeting 2016
Presentation transcript:

dCache in the NDGF Distributed Tier 1 Gerd Behrmann Second dCache Workshop Hamburg, 18 th of January 2007

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, 2007 NORDUnet A/S The Nordic Regional Research and Educational Network (RREN) Owned by the 5 Nordic National RENs 25 Years of Nordic network collaboration Leverage National Initiatives Participates in major international efforts Represents Nordic NRENs internationally, gateway to the Nordic area

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, 2007 What is the NDGF? A Co-operative Nordic Data and Computing Grid facility  Nordic production grid, leveraging national grid resources  Common policy framework for Nordic production grid  Joint Nordic planning and coordination  Operate Nordic storage facility for major projects  Co-ordinate & host major e-Science projects (i.e., Nordic WLCG Tier-1)  Develop grid middleware and services NDGF  Funded (2 M.EUR/year) by National Research Councils of the Nordic countries  Builds on a history of Nordic grid collaboration  Strategic planning ongoing. NOS-N DKDK SFSF N S Nordic Data Grid Facility

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, 2007 NDGF Vision “...to establish a Nordic Data Grid Facility and to involve Nordic countries in European and global co-operation in data sharing in a variety of fields.” To coordinate and facilitate the creation of a Nordic e-Infrastructure sharing platform To enable Nordic researchers to participate in major international projects To optimize and standardize use of resources To optimize Nordic participation in international projects Think of NDGF as one big Super Computer Center – spanning the entire Nordic area

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, dCache at NDGF NDGF Tier 1 sites connected by dedicated 10Gbit fiber (end of 2007/ beginning of 2008) Storage resources not located at NDGF and not under direct control of NDGF Storage ressources not necessarily dedicated to dCache Still, we are required to expose all sites as a single Tier1 with a single entry point. Two installations are established, remaining sites have received funding and will be established during No “worker nodes”

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, Scenario A Central dCache installation with minimal storage (first-level dCache) Standalone dCache installation at each site (second-level dCache) Use HSM interface to stage data between first-level and second-level

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, Scenario A Problems:  Latency  Inherently centralized  Central buffering required  No standalone operation  Improvement possible by new type of pool Avoid central storage by storing directly on remote systems

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, Scenario B One uniform dCache spanning all sites dCache head nodes operated by NDGF and placed in Copenhagen (7 x Dell 1950, Dual Intel Core Xeon, 4GB RAM, 2 x External RAID boxes) dCache pools operated by site owners Still centralized We currently deploy scenario B

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, Challenges Security  Many administrative domains  Local and national rules  Internal node communication over WAN  Mounting NFS over WAN is out of the question Administration  Site administrators are worried about loosing control  Mechanisms for delegating control over local ressources

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, Challenges Maintenance  Platform (SL is not widely used in NorduGrid)  Upgradability  Autonomeous operation Reliability  dCache is fairly resilient against pool failures  Head nodes provide central point of failure  Network saparation in WAN  Disconnected operation (at least read-only) Brain dump ideas: Replicated name space provider, DHT  Long term hope that dCache becomes less centralised

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, Challenges Performance  No network model e.g. SRM door assummes all GridFTP doors are equal (except for current load)  Proxy operation of GridFTP Functionality  HSM without PNFS (done in head)  Heterogenous access to HSM Stage-in must happen to connected pool Tape0Disk1 -> Tape1Disk1 may require file migration to another pool  Tivoli (TSM) integration  User friendly view of logical name space without PNFS (beyond FTP access and beyond admin shell)

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, Problems with (Grid)FTP PASV/PORT before STOR/RETR  dCache in PASV mode will always use door as proxy In mode E (extended block mode), the sender is always active.  Uploads to dCache always use door as proxy

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, Solution A SRM performs SURL to TURL translation  Currently TURL points to some GridFTP using LFN. Instead, let TURL be: gsiftp://pool:port/pfn i.e. let TURL point directly to where the data is stored/is supposed to be stored. Consequences:  SRM door gets more work to do, and  we need to embed an FTP implementation in a mover  could make it possible to queue requests in SRM rather than in pools

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, Solution B Use a proxy, but do not place it at the door Problems  Requires a network model  Assumes that control channel and data channel is established from the same location

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, Solution C GridFTP version 2  Draft, GFD-R-P.047 Mandrichenko, Allcock, Perelmutov Relevant highlights  GET and PUT commands replace PASV/PORT and STOR/RETR: GET file=/foo/bar;pasv;mode=e; 127 PORT=(130,225,33,7,156,7)  New eXtended block mode: Mode X No restriction between direction of connection and direction of transfer, e.g. sender can be passive. Number of concurrent connections can adapt; check sum on blocks; concurrent transfers on shared data channels...

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, Solution C Problems:  Draft status. Some clarifications needed, IMO.  No signs of progress since June  No implementations, except dCache head has GETPUT feature and mode X is on my laptop Started developing a patch for GLOBUS.

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, Commitment 1-2 FTE for dCache development and initial deployment. As long as needed and as long as dCache moves in “the right direction”. We avoid long term promisses. Concrete development plans are only made for one month at a time. We currently focus on GridFTP2 (solution C). Next in line will likely be solution A and/or HSM issues.

dCache in the NDGF Distributed Tier 1 Second dCache Workshop, Hamburg, 18th of January, Summary WAN deployment of dCache at the NDGF distributed Tier 1 Provides unique and interesting problems NDGF is committed to contribute to dCache to resolve these problems... and continue to do so as long as we have the need and dCache continues to move in “the right direction”. Current focus is on immediate problems with data flow, HSM, security and administration.