Introduction to OSG Storage Suchandra Thapa Computation Institute University of Chicago March 19, 20091GSAW 2009 Clemson.

Slides:



Advertisements
Similar presentations
30-31 Jan 2003J G Jensen, RAL/WP5 Storage Elephant Grid Access to Mass Storage.
Advertisements

OSG Public Storage Project Summary Ted Hesselroth October 5, 2010 Fermilab.
NGOP J.Fromm K.Genser T.Levshina M.Mengel V.Podstavkov.
A. Sim, CRD, L B N L 1 Data Management Foundations Workshop, Mar. 3, 2009 Storage in OSG and BeStMan Alex Sim Scientific Data Management Research Group.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
Understanding and Managing WebSphere V5
Hands-On Microsoft Windows Server 2008 Chapter 8 Managing Windows Server 2008 Network Services.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Open Science Grid Software Stack, Virtual Data Toolkit and Interoperability Activities D. Olson, LBNL for the OSG International.
Rsv-control Marco Mambelli – Site Coordination meeting October 1, 2009.
SRM at Clemson Michael Fenn. What is a Storage Element? Provides grid-accessible storage space. Is accessible to applications running on OSG through either.
OSG Services at Tier2 Centers Rob Gardner University of Chicago WLCG Tier2 Workshop CERN June 12-14, 2006.
A. Sim, CRD, L B N L 1 US CMS Workshop, Mar. 3, 2009 Berkeley Storage Manager (BeStMan) Alex Sim Scientific Data Management Research Group Computational.
OSG Site Provide one or more of the following capabilities: – access to local computational resources using a batch queue – interactive access to local.
Integration and Sites Rob Gardner Area Coordinators Meeting 12/4/08.
OSG Middleware Roadmap Rob Gardner University of Chicago OSG / EGEE Operations Workshop CERN June 19-20, 2006.
The Disk Resource Manager A Berkeley SRM for Disk Cache Implementation and Testing Experience on the OSG Jorge Luis Rodriguez IBP/Grid Mini-Workshop Apr.
Publication and Protection of Site Sensitive Information in Grids Shreyas Cholia NERSC Division, Lawrence Berkeley Lab Open Source Grid.
StoRM Some basics and a comparison with DPM Wahid Bhimji University of Edinburgh GridPP Storage Workshop 31-Mar-101Wahid Bhimji – StoRM.
May 8, 20071/15 VO Services Project – Status Report Gabriele Garzoglio VO Services Project – Status Report Overview and Plans May 8, 2007 Computing Division,
DCache at Tier3 Joe Urbanski University of Chicago US ATLAS Tier3/Tier2 Meeting, Bloomington June 20, 2007.
A. Sim, CRD, L B N L 1 OSG Applications Workshop 6/1/2005 OSG SRM/DRM Readiness and Plan Alex Sim / Jorge Rodriguez Scientific Data Management Group Computational.
G RID M IDDLEWARE AND S ECURITY Suchandra Thapa Computation Institute University of Chicago.
SLAC Experience on Bestman and Xrootd Storage Wei Yang Alex Sim US ATLAS Tier2/Tier3 meeting at Univ. of Chicago Aug 19-20,
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Author - Title- Date - n° 1 Partner Logo WP5 Summary Paris John Gordon WP5 6th March 2002.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
UMD TIER-3 EXPERIENCES Malina Kirn October 23, 2008 UMD T3 experiences 1.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
OSG Tier 3 support Marco Mambelli - OSG Tier 3 Dan Fraser - OSG Tier 3 liaison Tanya Levshina - OSG.
Overview of Privilege Project at Fermilab (compilation of multiple talks and documents written by various authors) Tanya Levshina.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
OSG AuthZ components Dane Skow Gabriele Carcassi.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
Configuring and Troubleshooting Identity and Access Solutions with Windows Server® 2008 Active Directory®
OSG Abhishek Rana Frank Würthwein UCSD.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
ALCF Argonne Leadership Computing Facility GridFTP Roadmap Bill Allcock (on behalf of the GridFTP team) Argonne National Laboratory.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Open Science Grid Build a Grid Session Siddhartha E.S University of Florida.
OSG Storage VDT Support and Troubleshooting Concerns Tanya Levshina.
April 25, 2006Parag Mhashilkar, Fermilab1 Resource Selection in OSG & SAM-On-The-Fly Parag Mhashilkar Fermi National Accelerator Laboratory Condor Week.
VOX Project Tanya Levshina. 05/17/2004 VOX Project2 Presentation overview Introduction VOX Project VOMRS Concepts Roles Registration flow EDG VOMS Open.
Bestman & Xrootd Storage System at SLAC Wei Yang Andy Hanushevsky Alex Sim Junmin Gu.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Site Authorization Service Local Resource Authorization Service (VOX Project) Vijay Sekhri Tanya Levshina Fermilab.
Tier 3 Support and the OSG US ATLAS Tier2/Tier3 Workshop at UChicago August 20, 2009 Marco Mambelli –
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Mario Reale – GARR NetJobs: Network Monitoring Using Grid Jobs.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
ITMT 1371 – Window 7 Configuration 1 ITMT Windows 7 Configuration Chapter 8 – Managing and Monitoring Windows 7 Performance.
Ted Hesselroth, OSG Site Administrators Meeting, December 13, 2007 Abhishek Singh Rana and Frank Wuerthwein UC San Diego dCache in OSG 1.0 and SRM 2.2.
OSG Status and Rob Gardner University of Chicago US ATLAS Tier2 Meeting Harvard University, August 17-18, 2006.
OSG STORAGE AND DATA MOVEMENT Tanya Levshina. Talk Outline  Data movement methods and limitation  Open Science Grid (OSG) Storage  Storage Resource.
A System for Monitoring and Management of Computational Grids Warren Smith Computer Sciences Corporation NASA Ames Research Center.
New Features of Xrootd SE Wei Yang US ATLAS Tier 2/Tier 3 meeting, University of Texas, Arlington,
Bologna, March 30, 2006 Riccardo Zappi / Luca Magnoni INFN-CNAF, Bologna.
OSG STORAGE OVERVIEW Tanya Levshina. Talk Outline  OSG Storage architecture  OSG Storage software  VDT cache  BeStMan  dCache  DFS:  SRM Clients.
A. Sim, CRD, L B N L 1 Production Data Management Workshop, Mar. 3, 2009 BeStMan and Xrootd Alex Sim Scientific Data Management Research Group Computational.
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
Open Science Grid Configuring RSV OSG Resource & Service Validation Thomas Wang Grid Operations Center (OSG-GOC) Indiana University.
Enabling Grids for E-sciencE Claudio Cherubino INFN DGAS (Distributed Grid Accounting System)
Grid Colombia Workshop with OSG Week 2 Startup Rob Gardner University of Chicago October 26, 2009.
BeStMan/DFS support in VDT OSG Site Administrators workshop Indianapolis August Tanya Levshina Fermilab.
a brief summary for users
Jean-Philippe Baud, IT-GD, CERN November 2007
Netscape Application Server
T-StoRM: a StoRM testing framework
Presentation transcript:

Introduction to OSG Storage Suchandra Thapa Computation Institute University of Chicago March 19, 20091GSAW 2009 Clemson

Overview Storage in OSG – Storage for VDT – Storage Extension – Support – Certification – Community tools – Documentation – Support – GIP & Storage Element Bestman – Overview – “full-mode”/“gateway” – “gateway”/Xrootd installation dCache – Overview – Monitoring tools – Opportunistic storage – Gratia Probes March 19, 20092GSAW 2009 Clemson

Why do we care about storage? Sites are growing to the point where putting a bunch of disks and nfs exporting it no longer works – Too much space needed to make it economical LHC projects will generate several petabytes of data a year Typical Tier 2 sites have on the order of 500TB – 1PB of storage space – I/O contention and loads will bring down a normal NFS server March 19, 2009GSAW 2009 Clemson3

Why care about storage (the sequel)? Would like additional information on storage – Let users easily get space information (available/used) without checking sites individually Advertise space availability and allow disk space to be reserved – If your site is sponsored by a VO like ATLAS, you don’t want to have another VO come in and use up all the space March 19, 2009GSAW 2009 Clemson4

March 19, 2009GSAW 2009 Clemson5 Solutions for OSG Storage OSG Storage for VDT is a well integrated distributed project between Wisconsin and Fermilab. Activities: Packaging storage software for VDT – Srm/dCache – BeStMan – BeStMan-gateway/Xrootd (NEW) Simplify configuration/installation for OSG Help VOs to use storage on OSG sites Develop and run validation tests Develop/maintain/package accounting and monitoring tools Support/test/package community tools Provide users and admins support Perform troubleshooting and debugging OSG liaison to storage developer groups Educate OSG community about storage, provide documentation Participate in grid schools organized by OSG

March 19, 2009GSAW 2009 Clemson6 OSG Storage Extension Project Tightly related to OSG Storage for VDT Goals for this year: – Develop software to match storage attributes of a job to a Storage Element – Develop web interface for SE discovery and authorization check for creating and managing space reservations – Work on consolidation of monitoring functionalities of Storage Elements into one interface. – Provide electronic means for sites to announce SE maintenance period

March 19, 2009GSAW 2009 Clemson7 Support Challenges Complicated, highly distributed services Huge variety of configuration options (software and hardware) Widely diverse utilization patterns dCache is known for poor error diagnostics, exception handling and propagation We do not enough experience with Xrootd Lack of monitoring/diagnostic tools Support team does not have access to the service. Support personnel – Often are not authorized to use the service as user – Can not access site logs and configuration – Often can not access storage monitoring pages on the site We would like to ask storage administrators for cooperation in: Notifying us about the reoccurring problems Provide us access to log files, configuration files

March 19, 2009GSAW 2009 Clemson8 Certification Maintain test stands – 6 nodes test stand for dCache – Planning to have 5 nodes test stand for BeStMan-gateway/xrootd Develop/run validation test suites before software is released to VDT – dCache test suite covers: all srm-fermi-client commands data replication space management load tests – BeSTMan testing provided by LBL Site registration, daily test results - Site could run tests with srm-tester-2 – instructions at – BeStMan-gateway/Xrootd covers: all supported srm-lbl-client /srm-fermi-client commands – work in progress at Fermilab

March 19, 2009GSAW 2009 Clemson9 Test Suites Results Fermilab Validation Test Suite Results LBL SRM Tester Results

March 19, 2009GSAW 2009 Clemson10 Community-contributed toolkit Goal is to provide a boost in effectiveness and efficiency ofoperating the deployed storage Download from du/toolkit Is packaged as rpms Collected/packaged by A. Rana

March 19, 2009GSAW 2009 Clemson11 Storage Documentation Revised documentation Main Page: ebHome Useful links under Storage Element administrators: – Opportunistic Storage/Space Reservation – Opportunistic Storage Model for USCMS – Gratia Storage Probes – Tools, Tips, FAQs – dCache Installation/references – BeStMan references

March 19, 2009GSAW 2009 Clemson12 Storage Element Discovery Glue Schema 1.3 – ControlProtocol SRM – AccessProtocol gsiftp – Storage Area Groups of Pools VOInfo – Path GIP/CeMon – Collects information on CE – Sends BDII/ReSS information – ReSS Information is used for job matching

March 19, 2009GSAW 2009 Clemson13 Storage Discovery Example You can find what sites support your VO or what information is advertised by a site about SE:

March 19, 2009GSAW 2009 Clemson14 Berkley Storage Management (BeStMan) What is BeStMan? – Developed in LBNL by Scientific Data Management Research Group – Full implementation of SRM v2.2 for disk based storage systems and mass storage systems – Supports transfer services: GsiFtp Ftp Http Https – Supports multiple transfer servers – GSI security with either grid-mapfile or GUMS server Who would benefit from BeStMan? – Sites with limited hardware resources – Sites with limited admin effort BeStMan could be used in two modes: – Full mode – Gateway mode

March 19, 2009GSAW 2009 Clemson15 BeStMan-gateway Generic SRM v2.2 load balancing frontend for GridFTP servers Light-weight implementation of SRM v2.2 for POSIX file systems – srmPing – srmLs – srmRm – srmMkdir – srmRmdir – srmPrepareToPut (Status, PutDone) – srmPrepareToGet(Status,ReleaseFiles) Designed to work with any Posix-like file systems – NFS, GPFS, GFS, NGFS, PNFS, HFS+, PVFS, AFS, Lustre, XrootdFS, Hadoop Doesn’t support queuing or disk space management

March 19, 2009GSAW 2009 Clemson16 BeStMan-gateway/Xrootd Xrootd - is designed to provide POSIX-like access to files and their enclosing directory namespace BeStMan-gateway needs the following additional components to work with Xrootd – FUSE - File System in User Space – XrootdFS - implements a Posix filesystem for an Xrootd storage cluster GridFtp needs Data Storage Interface (DSI) module in order to work with Xrootd storage

March 19, 2009GSAW 2009 Clemson17 BeStMan-gateway/Xrootd Architecture

March 19, 2009GSAW 2009 Clemson18 Before Installing BeStMan-gateway/Xrootd How many nodes could be used for storage? Minimum number of nodes is 3: BeStMan, XroodFS, fuse, GridFtp Xrootd redirector Xrootd data server node Chose authorization mechanism do you prefer? GUMS gridmap-file Do you need to support static space tokens? Select name for Mount Point for XrootdFS on BeStMan node Decide how to partition storage areas on Xrootd redirector and data server nodes

March 19, 2009GSAW 2009 Clemson19 XrootdFs Installation (I) Install fuse using “yum install” or rpms (e.g p?query=fuse) – fuse – fuse-libs – kernel-module-fuse EL Install and configure XrootdFS pacman -get $VDT_LOCATION/vdt/setup/configure_xrootdfs \ --user \ --cache \ --xrdr-host \ --xrdr-storage-path

March 19, 2009GSAW 2009 Clemson20 BeStMan-gateway Installation (II) Install and configure BestMan pacman -get OSG:Bestman OSG:Bestman $VDT_LOCATION/vdt/setup/configure_bestman --server y \ --user \ --cert \ --key \ --http-port \ --https-port \ --gums-host \ --gums-port \ --gums-dn \ --use-xrootd \ --with-tokens-list " [desc:<TOKEN_1_DESC][TOKEN_1_SIZE_GB];TOKEN_2_NAME[desc: TOKEN_2_DESC][TOKEN_2_SIZE]" \ --with-transfer-servers Modify /etc/sudoers Cmnd_Alias SRM_CMD = /bin/rm, /bin/mkdir, /bin/rmdir, /bin/mv, /bin/ls Runas_Alias SRM_USR = ALL, !root ALL=(SRM_USR) NOPASSWD: SRM_CMD

March 19, 2009GSAW 2009 Clemson21 GridFtp Installation (II) Install and configure GridFtp pacman -get OSG:Xrootd-GridFTPOSG:Xrootd-GridFTP $VDT_LOCATION/vdt/setup/configure_gridftp --use-xrootd \ --xrootd-host \ --xrootd-mount-point \ --xrootd-storage-path

March 19, 2009GSAW 2009 Clemson22 Xrootd-Redirector Installation (IV) Install and configure Xrootd redirector pacman -get OSG:Xrootd OSG:Xrootd $VDT_LOCATION/vdt/setup/configure_xrootd \ --server y \ --this-is-xrdr \ --user \ --xrdr-storage-path \ --xrdr-storage-cache \ --with-tokens-list " [desc:<TOKEN_1_DESC][TOKEN_1_SIZE_GB];TOKEN_2_NAME[desc :TOKEN_2_DESC][TOKEN_2_SIZE]" \ --public-cache-size

March 19, 2009GSAW 2009 Clemson23 Xrootd Data Server Installation (V) Install and configure Xrootd data server pacman -get OSG:rootdOSG:rootd $VDT_LOCATION/vdt/setup/configure_xrootd \ --server y \ --user \ --xrdr-host \ --xrdr-storage-path \ --xrdr-storage-cache \ --with-tokens-list " [desc:<TOKEN_1_DESC][TOKEN_1_SIZE_GB];TOKEN_2_NAME[desc :TOKEN_2_DESC][TOKEN_2_SIZE]" \ --public-cache-size

March 19, 2009GSAW 2009 Clemson24 Start/stop order for BeStMan-gateway/Xrootd Use vdt-control –on/--off mechanism Start severs in the following order (use reverse order to stop): – Xrootd redirector – Xrootd data server – GridFtp – XrootdFs/Bestman-gateway Test srm-ping srm:// :8443/srm/v2/server srm-copy file:////tmp/test srm://.gov:8443/srm/v2/server\?SFN= /test -spacetoken

March 19, 2009GSAW 2009 Clemson25 BeStMan is one of data storage solutions supported by OSG. BeStMan-gateway, BeStMan-gateway/Xrootd requested by ATLAS are just released in VDT Current version of software – BeStMan e1 – XrootdFS – GridFTP-Xrootd,xrootd-dsi – Prima – Xrootd VDT configuration script tailored to set up BeStMan “full mode”/gateway for Tier-2/Tier-3 BeStMan srm-clients are distributed as a part of VDT client cache – Fermi client – LBNL client – LCG-utils There are several installations of BeStMan on OSG sites BeStMan in VDT

March 19, 2009GSAW 2009 Clemson26 dCache Main Features nfs-mountable namespace Multiple access protocols – dcap (posix io), gsidcap – xrootd (posix io) – gsiftp (multiple channels) Replica Manager – Set min/max number of replicas Role-based authorization – Selection of authorization mechanisms Opportunistic storage Billing Admin interface – ssh, jython InformationProvider (not in production - yet) – SRM and gsiftp described in glue schema Platform, fs independent (Java) – 32 and 64-bit linux, solaris; ext3, xfs, zfs

March 19, 2009GSAW 2009 Clemson27 dCache OSG Tier-2 site Architecture Slide courtesy of Ted Hesselroth (from presentation: “Installing and Using SRM- dCache”)

March 19, 2009GSAW 2009 Clemson28 dCache Monitoring Numerous ways to monitor dCache Info Pages SrmWatch - 0/srmwatch/ Admin Interface ( UI or from command line) Scripts/web pages built by community – – src/plots.lzx src/plots.lzx

March 19, 2009GSAW 2009 Clemson29 SRMWatch Examples

March 19, 2009GSAW 2009 Clemson30 Gratia Service Gratia is the accounting service for OSG is provided by the Gratia external project. Main goal is to provide the stakeholders with a reliable and accurate set of views of the Grid resources usage. Job and other accounting information gathered by Gratia probes run on the compute element or other site nodes are reported to a Gratia collectors – Fermi collector: reportinghttp://gratia-fermi.fnal.gov:8886/gratia- reporting – OSG collector: reportinghttp://gratia.opensciencegrid.org:8886/gratia- reporting Accounting records collected by Gratia are forwarded to the EGEE accounting system, APEL: – iew.html iew.html

March 19, 2009GSAW 2009 Clemson31 dCache Gratia Probes – Storage Probe – Transfer Probe Storage Probe – Is responsible for reporting storage capacity and storage usage – Gets the pool information from the dCache admin server – Gets the SRM information from the SRM tables in the SRM Postgres database – Runs as a cron job on the host running the Postgres database server for SRM Transfer Probe – Reports the details of each file transfer into or out of a dCache file server – Gets this information from the dCache "billing" database. – Runs as a daemon process – For performance reasons, sites with large dCache billing databases are advised to alter the "billinginfo" table by indexing specific tables in order speed up the search for newly added records

March 19, 2009GSAW 2009 Clemson32 Gratia Transfer Probes Report

March 19, 2009GSAW 2009 Clemson33 dCache Information Service A snapshot of the current status of a dCache instance for external consumption – Works independently of the rest of dCache. – Pools information for various dCache components (configurable update period). Doesn't provide historic data Not an info-provider but GIP is a customer Access data: – Via the admin interface – Via XML Conduit - A TCP connection get complete state as XML. – Web front-end – Supports some advanced features: Derived data (re-)calculated as state changes. Multiple output formats and transports. If additional metrics, data formats or transports are needed, they can be added

March 19, 2009GSAW 2009 Clemson34 From Paul Millar’s presentation: “dCache seminar: introducing the info service “ Information Service Architecture

March 19, 2009GSAW 2009 Clemson35 Opportunistic Storage Opportunistic Storage in dCache 1.8 with SRM 2.2 – Provides a capability of specifying a portion of the total storage for opportunistic use – Allows particular VOs and Roles a privilege to use space other than that included in opportunistic storage – Files created through opportunistic use will not be permanently available in the storage system – A storage site administrator may configure the site for opportunistic use through space reservation. – Creation of space reservations is controlled by use of link groups – The administrator may assign storage pools to link groups – Certain pools are designated for opportunistic use. Numerous documents describing how to install and operate Opportunistic Storgae on Tier-2 sites – Setup Setup – lForCMS lForCMS

March 19, 2009GSAW 2009 Clemson36 dCache in VDT dCache is one of the data storage solutions supported by OSG dCache could be installed from VDT Current version is vdt (dcache p11) Distribution contains dCache-server, pnfs, postgress, gratia probes rpms and a configuration script tailored to set up dCache for Tier-2/Tier-3 Configuration script allows to do system setup, enable opportunistic storage, replication etc dCache-clients are distributed as a part of VDT client cache – Fermi client – LBNL client – LCG-utils There are multiple installations of dCache on OSG sites

March 19, 2009GSAW 2009 Clemson37 Summary We will continue to work on improving storage packaging in VDT – Feedback is welcome! We are trying to make support more efficient by providing FQA, debugging the most frequently occurred problems, working with developers on improving logging and error diagnostic – The quality of the support depends greatly on Storage Admins cooperation!!! We will add BeStMan test stand and will do BeStMan/Xrootd certification the same way it is done with dCache release We will be glad to accept/package more community tools We are trying to maintain documentation up-to-date, adding new interesting references and “how to do” tips – Please let us know if we are missing some important topics! As a liaison to software developers we will be happy to pass your requests/suggestions

Acknowledgements Tanya Levshina March 19, 2009GSAW 2009 Clemson38