DCache at Tier3 Joe Urbanski University of Chicago US ATLAS Tier3/Tier2 Meeting, Bloomington June 20, 2007.

Slides:



Advertisements
Similar presentations
Owen SyngeDCache Slide 1 D-Cache Status Report Owen Synge.
Advertisements

GridKa January 2005 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann 1 Mass Storage at GridKa Forschungszentrum Karlsruhe GmbH.
EGEE is a project funded by the European Union under contract IST Using SRM: DPM and dCache G.Donvito,V.Spinoso INFN Bari
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
Windows Server 2008 Chapter 8 Last Update
Hands-On Microsoft Windows Server 2008 Chapter 8 Managing Windows Server 2008 Network Services.
Hands-On Microsoft Windows Server 2008 Chapter 5 Configuring, Managing, and Troubleshooting Resource Access.
OSG End User Tools Overview OSG Grid school – March 19, 2009 Marco Mambelli - University of Chicago A brief summary about the system.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
70-291: MCSE Guide to Managing a Microsoft Windows Server 2003 Network Chapter 7: Domain Name System.
INSTALLING MICROSOFT EXCHANGE SERVER 2003 CLUSTERS AND FRONT-END AND BACK ‑ END SERVERS Chapter 4.
Home Media Network Hard Drive Training for Update to 2.0 By Erik Collett Revised for Firmware Update.
SRM at Clemson Michael Fenn. What is a Storage Element? Provides grid-accessible storage space. Is accessible to applications running on OSG through either.
Ted Hesselroth USATLAS Tier 2 and Tier 3 Workshop November 29, 2007 Abhishek Singh Rana and Frank Wuerthwein UC San Diego Installing and Using SRM-dCache.
Introduction to OSG Storage Suchandra Thapa Computation Institute University of Chicago March 19, 20091GSAW 2009 Clemson.
How to Install and Use the DQ2 User Tools US ATLAS Tier2 workshop at IU June 20, Bloomington, IN Marco Mambelli University of Chicago.
A. Sim, CRD, L B N L 1 OSG Applications Workshop 6/1/2005 OSG SRM/DRM Readiness and Plan Alex Sim / Jorge Rodriguez Scientific Data Management Group Computational.
Data management for ATLAS, ALICE and VOCE in the Czech Republic L.Fiala, J. Chudoba, J. Kosina, J. Krasova, M. Lokajicek, J. Svec, J. Kmunicek, D. Kouril,
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Support in setting up a non-grid Atlas Tier 3 Doug Benjamin Duke University.
SAM Installation Lauri Loebel Carpenter and the SAM Team February
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Maarten Litmaath (CERN), GDB meeting, CERN, 2006/02/08 VOMS deployment Extent of VOMS usage in LCG-2 –Node types gLite 3.0 Issues Conclusions.
 CASTORFS web page - CASTOR web site - FUSE web site -
LCG Storage workshop at CERN. July Geneva, Switzerland. BNL’s Experience dCache1.8 and SRM V2.2 Carlos Fernando Gamboa Dantong Yu RHIC/ATLAS.
Chapter 10 Chapter 10: Managing the Distributed File System, Disk Quotas, and Software Installation.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
June 22, 2007USATLAS T2-T3 DQ2 0.3 SiteServices Patrick McGuigan
DCache Basics Alessandro Usai, Sergio Maffioletti Grid Group CSCS.
Overview of Privilege Project at Fermilab (compilation of multiple talks and documents written by various authors) Tanya Levshina.
SRM Monitoring 12 th April 2007 Mirco Ciriello INFN-Pisa.
Role Based VO Authorization Services Ian Fisk Gabriele Carcassi July 20, 2005.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
OSG AuthZ components Dane Skow Gabriele Carcassi.
Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.
Configuring and Troubleshooting Identity and Access Solutions with Windows Server® 2008 Active Directory®
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
Tier 3 Status at Panjab V. Bhatnagar, S. Gautam India-CMS Meeting, July 20-21, 2007 BARC, Mumbai Centre of Advanced Study in Physics, Panjab University,
Oct 24, 2002 Michael Ernst, Fermilab DRM for Tier1 and Tier2 centers Michael Ernst Fermilab February 3, 2003.
OSG Site Admin Workshop - Mar 2008Using gLExec to improve security1 OSG Site Administrators Workshop Using gLExec to improve security of Grid jobs by Alain.
CSC414 “Introduction to UNIX/ Linux” Lecture 6. Schedule 1. Introduction to Unix/ Linux 2. Kernel Structure and Device Drivers. 3. System and Storage.
INFSO-RI Enabling Grids for E-sciencE SRMv2.2 in DPM Sophie Lemaitre Jean-Philippe.
GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache Implementation at FZK Forschungszentrum Karlsruhe.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
Martina Franca (TA), 07 November Installazione, configurazione, testing e troubleshooting di Storage Element.
Ted Hesselroth, OSG Site Administrators Meeting, December 13, 2007 Abhishek Singh Rana and Frank Wuerthwein UC San Diego dCache in OSG 1.0 and SRM 2.2.
Planning File and Print Services Lesson 5. File Services Role The File Services role and the other storage- related features included with Windows Server.
OSG STORAGE OVERVIEW Tanya Levshina. Talk Outline  OSG Storage architecture  OSG Storage software  VDT cache  BeStMan  dCache  DFS:  SRM Clients.
Security recommendations DPM Jean-Philippe Baud CERN/IT.
Abhishek Singh Rana and Frank Wuerthwein UC San Diegowww.opensciencegrid.org The Open Science Grid ConsortiumCHEP 2006 Mumbai INDIA February gPLAZMA:
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
9 Copyright © 2004, Oracle. All rights reserved. Getting Started with Oracle Migration Workbench.
STORAGE EXPERIENCES AT MWT2 (US ATLAS MIDWEST TIER2 CENTER) Aaron van Meerten University of Chicago Sarah Williams Indiana University OSG Storage Forum,
Open Science Grid Configuring RSV OSG Resource & Service Validation Thomas Wang Grid Operations Center (OSG-GOC) Indiana University.
Why you should care about glexec OSG Site Administrator’s Meeting Written by Igor Sfiligoi Presented by Alain Roy Hint: It’s about security.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
DCache.ORG Fuhrman n Wuppertal Atlas GridKa Cloud dCache.ORG dCache best practice additional funding, support or contributions by d-grid DGI.
Security recommendations for dCache
Jean-Philippe Baud, IT-GD, CERN November 2007
CCNA Routing and Switching Routing and Switching Essentials v6.0
dCache “Intro” a layperson perspective Frank Würthwein UCSD
dCache Status and Plans – Proposals for SC3
Chapter 10: Device Discovery, Management, and Maintenance
CCNA Routing and Switching Routing and Switching Essentials v6.0
Chapter 10: Device Discovery, Management, and Maintenance
INFNGRID Workshop – Bari, Italy, October 2004
Presentation transcript:

dCache at Tier3 Joe Urbanski University of Chicago US ATLAS Tier3/Tier2 Meeting, Bloomington June 20, 2007

What is dCache? From dCache.org: "A system for storing and retrieving huge amounts of data, distributed among a large number of heterogenous server nodes, under a single virtual filesystem tree”

Features & Advantages What can dCache do?  Capable of combining hundreds of commodity disk servers to get a huge petabyte scale data store  Allows several copies of a single file for distributed data access  Has internal load balancing using cost metrics and transfers between the site's pools  Has automatic file replication on high load ("hotspot detection")

What does dCache look like? The single virtual filesystem is provided by pnfs (Pretty Normal File System). Mounting and unmounting through mount and umount and /etc/fstab, much like traditional nfs. pnfs is “POSIX-like”:  can use: ls, mkdir, find  cannot use: cp, md5sum

Clients How do I access files in dCache?  dCap: dCache's native method, uses dccp. Easiest for local transfers.  gridFTP: via globus-url-copy with a valid X.509 proxy  SRM: via srmcp, srmls with a valid X.509 proxy

Architecture What about the backend?  Admin nodes: Provide basic admin services. One or more of these.  Door nodes: Provide I/O access via SRM or GridFTP. One or more, may reside on admin node on small installs.  pnfs node: Provides the unified namespace. Only one per install.  Pool nodes: Provides storage. Can be installed alongside any other type of node.

UC Tier3

UC Tier3 (cont'd)‏ 3 Admin nodes:  uct3-edge1: gridFTP, dCap  uct3-edge2: pnfs  uct3-edge3: admin, SRM 25 Pool nodes:  Currently: 22 compute nodes x 1.9TB + 3 admin nodes x 1.9TB = 47TB

Installing dCache The VDT Installer  Greatly simplifies and quickens the install process Automatically configures which services to run, and on what nodes to run them. Installs needed rpms. Configures dCache and its postgresql databases.  Latest vdt version is v1.1.8 Not to be confused with the dCache version, latest is v1.7.0.

Running the VDT Installer Download the latest vdt tarball, untar and cd into the install directory. Run node-config.pl to generate site-info.def file

config-node.pl install]./config-node.pl How many admin nodes (non-pool and non-door nodes) do you have? 2 The recommended services for node 1 are: lmDomain poolManager adminDoor httpDomain utilityDomain gplazmaService infoProvider srm The recommended services for node 2 are: pnfsManager dirDomain Enter the FQDN for the node 1: uct3-edge3.uchicago.edu Which services do you wish to run on node uct2-dc3.uchicago.edu (Enter for defaults)? Enter the FQDN for the node 2: uct3-edge2.uchicago.edu Which services do you wish to run on node uct2-dc2.uchicago.edu (Enter for defaults)? How many door nodes do you have? 1 Enter the FQDN of door number 1: uct3-edge1.uchicago.edu

node-config.pl (cont'd)‏ Enter the private network that the pools are in. If this does not apply, just press enter to skip: Enter the number of dcap doors to run on each door node [default 1]: 1 Enter a pool FQDN name(Press Enter when all are done): uct3-c001.mwt2.org Enter the first storage location (Press Enter when all are done)): /dcache Enter another storage location (Press Enter when all are done)): --SNIP-- Enter another pool FQDN name(Press Enter when all are done): uct3-edge3.mwt2.org Enter the first storage location (Press Enter when all are done)): /dcache Enter another storage location (Press Enter when all are done)): Enter another pool FQDN name(Press Enter when all are done): Created site-info.def file. install]

Running the VDT Installer (cont'd)‏ Copy vdt tarball, and site-info.def file to all nodes. Run './install.sh -d' for a dryrun. (This will be very verbose). If successful, run the actual install. Start dCache services in the following order:  pnfs node core services  other admin nodes core services  all dcache pool services

Verifying the install Check the status webpage: 

Verifying the install (cont'd)‏ Test the doors:  dCap: use dccp  gridFTP: use globus-url-copy  SRM: use srmcp globus-url-copy -dbg \ file:////tmp/test.file \ gsiftp://uct3-edge1.uchicago.edu/pnfs/uchicago.edu/data/ddm1/test.file dccp -d999 /tmp/test.file \ /pnfs/uchicago.edu/data/ddm1/test.file srmcp -debug file:////tmp/test.file \ srm://uct3-edge3.uchicago.edu:8443/pnfs/uchicago.edu/data/ddm1/test.file

Troubleshooting Always check the status page!

Troubleshooting (cont'd)‏ Check the logs:  Most dCache cells: /var/log/*Domain.log. Each cell, or service, will generate a log with the appropriate name  SRM: /opt/d-cache/libexec/apache-tomcat /logs/catalina.out  pnfs: /var/log/pnfsd.log, /var/log/pmountd.log, /var/log/dbserver.log dCache, The Book:  Ask for help:  dCache.org support:  OSG-storage:  OSG Storage Activities Meeting: Every Thursday utes

UC Tier3 Pool Usage

Wormholes pnfs provides a way to distribute configuration information to all directories in the pnfs filesystem via 'wormholes'. Accessed via the '.(config)()' subdirectory, which acts like a symlink to /pnfs/fs/admin/etc/config/ By default, reading files like this is disabled, so you'll need to enable access to these files. Without being able to read that file, pnfs won't be able to find the dcap door, and dccp's won't work! WARNING: enabling access to these files empties the file!

Wormholes (cont'd)‏ dcache-upgrade-v1.1.4]# cd \ /pnfs/fs/admin/etc/config/dCache dCache]# cat dcache.conf uct2-dc1.mwt2.org:22125 dCache] touch ".(fset)(dcache.conf)(io)(on)" dCache] echo "uct3-edge1.mwt2.org:22125" > \ /pnfs/fs/admin/etc/config/dCache/dcache.conf

Authorization with gPLAZMA Grid-aware PLuggable AuthoriZation Management Works in a manner similar to PAM Four available methods:  kpwd: 'legacy method'. flat file maps DN's to a local username, then username to uid, gid, and rootpath  grid-mapfile: uses a grid-mapfile, then a second file, storage-authzdb, to map username to uid, gid, and rootpath  gplazmalite-vorole-mapping: concatenate DN + Role, then provide uid, gid, and rootpath via storage-authzdb.  saml-vo-mapping: uses GUMS to map to username, may provide uid, gid, and rootpath or via storage- authzdb.

The Admin Interface dCache provides a shell-like interface accessed via ssh. config]# ssh -c blowfish -p dCache Admin (VII) (user=admin)‏ (local) admin > cd uct3-edge2_1 (uct3-edge2_1) admin > pnfs register (uct3-edge2_1) admin >.. (local) admin > cd uct3-edge3_1 (uct3-edge3_1) admin > pnfs register (uct3-edge3_1) admin >..

Autovacuuming Postgres Pfns database files could potentially reach a very large size and fill up your filesystem. To turn it on, uncomment all the entries in the AUTOVACUUM PARAMETERS section and change 'autovacuum = off' to 'on' in /var/lib/pgsql/data/postgresql.conf, then restart postgresql.

Useful URL's dCache homepage  VDT Installer homepage  VDT Installer HOWTO  OSG Storage Activities Meeting  OSG dCache overview  USATLAS dCache at BNL 