OSG Public Storage Project Summary Ted Hesselroth October 5, 2010 Fermilab.

Slides:



Advertisements
Similar presentations
TeraGrid Deployment Test of Grid Software JP Navarro TeraGrid Software Integration University of Chicago OGF 21 October 19, 2007.
Advertisements

EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
CMS Applications Towards Requirements for Data Processing and Analysis on the Open Science Grid Greg Graham FNAL CD/CMS for OSG Deployment 16-Dec-2004.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
Understanding Active Directory
System Design/Implementation and Support for Build 2 PDS Management Council Face-to-Face Mountain View, CA Nov 30 - Dec 1, 2011 Sean Hardman.
Cyberinfrastructure Status July, NSF reverse site visit Refactoring and cleanup after review preparations Coordinating Node technology changes.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
1 Foundations V: Infrastructure and Architecture, Middleware Deborah McGuinness and Peter Fox CSCI Week 9, October 27, 2008.
Open Science Grid Software Stack, Virtual Data Toolkit and Interoperability Activities D. Olson, LBNL for the OSG International.
Rsv-control Marco Mambelli – Site Coordination meeting October 1, 2009.
OSG Public Storage and iRODS
SRM at Clemson Michael Fenn. What is a Storage Element? Provides grid-accessible storage space. Is accessible to applications running on OSG through either.
Scalable Systems Software Center Resource Management and Accounting Working Group Face-to-Face Meeting June 13-14, 2002.
Integration and Sites Rob Gardner Area Coordinators Meeting 12/4/08.
Usability Issues Documentation J. Apostolakis for Geant4 16 January 2009.
VOX Project Status T. Levshina. Talk Overview VOX Status –Registration –Globus callouts/Plug-ins –LRAS –SAZ Collaboration with VOMS EDG team Preparation.
1 Foundations V: Infrastructure and Architecture, Middleware Deborah McGuinness TA Weijing Chen Semantic eScience Week 10, November 7, 2011.
May 8, 20071/15 VO Services Project – Status Report Gabriele Garzoglio VO Services Project – Status Report Overview and Plans May 8, 2007 Computing Division,
Scalable Systems Software Center Resource Management and Accounting Working Group Face-to-Face Meeting October 10-11, 2002.
G RID M IDDLEWARE AND S ECURITY Suchandra Thapa Computation Institute University of Chicago.
1 Schema Registries Steven Hughes, Lou Reich, Dan Crichton NASA 21 October 2015.
ILDG Middleware Status Chip Watson ILDG-6 Workshop May 12, 2005.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
DIRAC Review (13 th December 2005)Stuart K. Paterson1 DIRAC Review Exposing DIRAC Functionality.
MAGDA Roger Jones UCL 16 th December RWL Jones, Lancaster University MAGDA  Main authors: Wensheng Deng, Torre Wenaus Wensheng DengTorre WenausWensheng.
1 Chapter Overview Preparing to Upgrade Performing a Version Upgrade from Microsoft SQL Server 7.0 Performing an Online Database Upgrade from SQL Server.
Getting started DIRAC Project. Outline  DIRAC information system  Documentation sources  DIRAC users and groups  Registration with DIRAC  Getting.
DDM Monitoring David Cameron Pedro Salgado Ricardo Rocha.
Metadata Mòrag Burgon-Lyon University of Glasgow.
Open Science Grid Open Science Grid: Beyond the Honeymoon Dane Skow Fermilab September 1, 2005.
David Adams ATLAS DIAL/ADA JDL and catalogs David Adams BNL December 4, 2003 ATLAS software workshop Production session CERN.
Overview of Privilege Project at Fermilab (compilation of multiple talks and documents written by various authors) Tanya Levshina.
INFSO-RI Enabling Grids for E-sciencE Enabling Grids for E-sciencE Pre-GDB Storage Classes summary of discussions Flavia Donno Pre-GDB.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
Cole David Ronnie Julio. Introduction Globus is A community of users and developers who collaborate on the use and development of open source software,
Jini Architecture Introduction System Overview An Example.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
AliEn AliEn at OSC The ALICE distributed computing environment by Bjørn S. Nilsen The Ohio State University.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
SAM Sensors & Tests Judit Novak CERN IT/GD SAM Review I. 21. May 2007, CERN.
David Adams ATLAS ATLAS distributed data management David Adams BNL February 22, 2005 Database working group ATLAS software workshop.
ESG-CET Meeting, Boulder, CO, April 2008 Gateway Implementation 4/30/2008.
JAliEn Java AliEn middleware A. Grigoras, C. Grigoras, M. Pedreira P Saiz, S. Schreiner ALICE Offline Week – June 2013.
State of Georgia Release Management Training
David Adams ATLAS ATLAS-ARDA strategy and priorities David Adams BNL October 21, 2004 ARDA Workshop.
April 25, 2006Parag Mhashilkar, Fermilab1 Resource Selection in OSG & SAM-On-The-Fly Parag Mhashilkar Fermi National Accelerator Laboratory Condor Week.
VOX Project Tanya Levshina. 05/17/2004 VOX Project2 Presentation overview Introduction VOX Project VOMRS Concepts Roles Registration flow EDG VOMS Open.
EGEE is a project funded by the European Union under contract IST Experiment Software Installation toolkit on LCG-2
Site Authorization Service Local Resource Authorization Service (VOX Project) Vijay Sekhri Tanya Levshina Fermilab.
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
OSG Security: Updates on OSG CA & Federated Identities Mine Altunay, PhD OSG Security Team OSG AHM March 24, 2015.
OSG STORAGE OVERVIEW Tanya Levshina. Talk Outline  OSG Storage architecture  OSG Storage software  VDT cache  BeStMan  dCache  DFS:  SRM Clients.
1 DIRAC Project Status A.Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille 10 March, DIRAC Developer meeting.
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Services for Distributed e-Infrastructure Access Tiziana Ferrari on behalf.
Open Science Grid Configuring RSV OSG Resource & Service Validation Thomas Wang Grid Operations Center (OSG-GOC) Indiana University.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Federating Data in the ALICE Experiment
Architecture Review 10/11/2004
Jean-Philippe Baud, IT-GD, CERN November 2007
Classic Storage Element
ALICE FAIR Meeting KVI, 2010 Kilian Schwarz GSI.
CUAHSI HIS Sharing hydrologic data
LCGAA nightlies infrastructure
Viet Tran Institute of Informatics Slovakia
DUCKS – Distributed User-mode Chirp-Knowledgeable Server
Leigh Grundhoefer Indiana University
Presentation transcript:

OSG Public Storage Project Summary Ted Hesselroth October 5, 2010 Fermilab

Provenance 2006: Acquire capability to allocate storage to VOs 2007: SRM 2.2 – space reservation – Intensely tested, debugged, documented by OSG Storage – Space reservation cleanup tool. 2008: Partial adoption – Used by Atlas, not used by CMS – Difficult to set up in dCache – Not supported in Bestman Gateway 2009: Increased use of opportunistic storage 2010: Blueprint meeting -renewed request/storage appliance – Requirements doc signed off – Design doc

Feedback from VOs Difficult to use – A large number of steps must be done by a user in order to run jobs using public storage. – Access to storage may not be available as advertised in the BDII. – There are difficulties in moving and tracking large numbers of files when they are treated as independent entities. Suggests requirements beyond those taken from the Blueprint meeting

Outcome of Requirements Process OSG Production Coordinator – Grant allocations to VOs. Resize, extend, rescind. – Clean up expired allocations VO Administrator – Request allocations. Resize, extend, rescind. – Make suballocations for users (and datasets). – Run access checker tool. – Clean up expired allocations. VO Member – Read, write, delete, (copy, and list) files. Includes registration update (and allocation enforcement). – Define datasets. – Replicate, delete datasets. Site Administrator – Help clean up allocation. – Set limits on number of files, concurrent connections.

Constraints on the Design No alteration of Storage Elements – Access continues to be through current clients. No centralized OSG service – Software to be operated by VO Accommodate usage outside the service – Use of traditional means will not have an adverse effect

Design Summary Database will store info on – Allocations – Replicas – Logical Namespace – Monitoring – Registered Users Database will have web-service front end – Invocation through wrapper scripts on the command line

Site registration/monitoring VO Administrator installs software (database, frontend, scripts). OSG Production Coordinator uses tool to register storage areas. – Discovery tool shows total and used space, VOs authorized. – Public storage areas are registered in Allocation database. VO Administrator runs monitoring tool – Discovers storage areas for which the VO is authorized – Checks access with probes – Results saved in Monitoring database

Making Allocations VOA makes a request to the PC asking for space – May optionally specify storage resource PC checks allocations and selects a storage resource. – Tool queries Allocation and Monitoring databases, shows resources that can meet the allocation parameters that the VO can access – New allocation object is made in database – All other VO’s Allocation databases are updated. PC informs VOA of new allocation

Using Allocations –write User invokes a script which does all of these Local file is specified as the source Logical full path is specified as the destination An allocation is selected – that has sufficient space – that has less than the maximum number of files – that is not expired – that is currently accessible to the VO The destination URL is composed VDT client tools are used to write the file The allocation is updated The file logical path is registered in the Namespace catalog The file replica is registered in the Replica catalog

Defining and using Datasets Allow operations on sets of files Especially useful for copying and deleting Could have hook into classads to trigger processing after upload In the Namespace catalog – A file or directory is tagged with a dataset handle – A file belongs to a dataset if it or one of its ancestors has a tag In the Replica catalog – There are dataset replica objects – Has list of member files which replicas that storage resource In transfer operations – List is composed from dataset replica – Resource selection done on the basis of total size

Defining Suballocations Similar procedure to defining allocations – But done by VOA, at the request of a user – Selection is made from the VOs allocations Can have suballocation for dataset – Similar to space reservation Space accounting tracks both the suballocation and its parent – Suballocation counted against parent when made – Writes/deletes update suballocation remaining space Users can clean up their own expired suballocations

Implementation Database – SQL for table creation, update, query – Postgres Web Service – Simple http-based – Message-level security – Access using curl in wrapper scripts – Possibly use Bestman for read, write, delete, copy, ls

Code Assets From the OSG Discovery Tool – Discovery of storage areas – Wrapper scripts for Java clients’ – Maven-build capability in the VDT. From the Pigeon Tool – Monitoring capability. From Bestman – SRM reference implementation – Close support. From MCAS – Restful web service – Development methodology – Fitnesse test suite. From OSG Storage – Postgres install script from the VDT installation package for dCache. – Java methods for WSS fast message-level security

Development Strategy Agile Methodology – Stakeholders involved from the beginning – Demonstrate new features every two weeks – Stakeholders test and give feedback – Frequent developers meetings for progress, short-term planning Continuous Integration – Test with every commit – Packaging is part of build – Nightly build available to stakeholders – Continuous documentation

Timeline October – Detailed implementation planning. – Infrastructure setup: twiki, issue tracker, code repository, build system, continuous integration and test system. – Kickoff meeting. November – SQL for database creation. Start SQL for updates and queries. Shell wrappers for SQL. – Pigeon integration. Installation scripts for testers. December – Finish SQL for updates and queries. Database performance testing/tuning. – Setup webservice code: stubs, queuing mechanism, Java SQL configuration. January – Web service installation script. Java SQL wrappers. Java methods for functions. SRM methods for transfer commands. February – GSI authentication for web service. Performance testing of web service. – End-to-end load testing. Reports capability March – VDT packaging. – ITB testing. – Post-facto registration and cleanup. RPMs for Operations Toolkit.

Unknowns Post-facto accounting. – We have reason to believe it can be done through gratia but have not tested this. Performance tuning – depending on the results of tests. Requirements creep – stakeholders to emphasize/deemphasize various elements, or request additional features.

Software not used iRODS SRM Space Reservation Existing transfer services Alien REDDNET

Extra slides

Database Schema

To provide storage space for non-owner VOs, sites generally allocate an untended common storage area authorized for several OSG VOs. While some VOs have availed themselves of those resources, the experience of the Engage and OSG Storage goups is that there are a number of barriers to its effective use. A large number of steps must be done by a user in order to run jobs using public storage. Access to storage may not be available as advertised in the BDII. There are difficulties in moving and tracking large numbers of files when they are treated as independent entities. A single VO may use up all the public storage on a site, preventing other VOs from having access. Without reportable information on the state and use of public storage, it is difficult to present its value to sites and other VOs.

The OSG Production Coordinator, OSG Storage, and the Engage and LIGO VOs have recognized a need for software to manage space allocations and data transfer for public storage on the Open Science Grid. OSG Storage produced and circulated a requirements document describing its capabilities, which was approved by the stakeholders after a comment period.

OSG Storage surveyed existing software and designed a service to meet the requirements. The service is to be deployed by VOs which make use of public storage, and allows VOs to track their use of allocations on public storage areas which are assigned to them by the Production Coordinator. The service also maintains a catalog of the VO's files, to allow cleanup of expired allocations and to support storage operations on sets of files. Finally, a monitoring component is included so that allocation and resource selection can be done for storage areas that are accessible to the VO.

We estimate it will take one FTE six months to write the service. This includes code/integration for the user interfaces and database operations, a VO test installation package, a VDT-compatible build and installation method, and documentation. We should have the participation of the OSG Production Coordinator and VO representatives throughout the development process, to exercise features as they become available and provide feedback on user experience, software performance, and documentation quality. This requires about 5 % per participant., and a deployment resource for one instance of the service per VO. We would not need participation of site administrators until near the end of the development process; an estimated one-half day would be asked of volunteers at that point. This assumes that the OSG- owned Storage Elements on Gridworks will be available as test endpoints. One non-virtual node should be provided for the developers' test instance of the new service.

On the specifics of what needs to be done, the design uses a database back end and anticipates wide area access through a GSI-authenticated web service front end. We would need to write SQL scripts to create, update, and query the database tables, and Java wrappers for access through the web service. For the front end we have the option of the Bestman reference implementation or a lightweight http endpoint with message-level security. Scripts for installation, startup, and command-line interface need to be written. The Pigeon access checker would be used for monitoring; it would need an add-on to write to the database. There is a good start on the documentation, as a twiki page ( has absorbed much of the requirements and design information.

We will build upon software assets acquired in the past. From the current work: vetted requirements and a thorough database schema and operations design. From the OSG Discovery Tool: discovery of storage areas, wrapper scripts for Java clients, and a maven-build capability in the VDT. From the Pigeon Tool, monitoring capability. From Bestman, the Bestman reference implementation and close support. From MCAS, a Restful web service, a development methodology, and the Fitnesse test suite. From OSG Storage: a Postgres install script from the VDT installation package for dCache. Also from the current work is Java methods WSS fast message-level security.

Unknowns are as follows. We have a requirement to do post-facto accounting. We have reason to believe it can be done through gratia but have not tested this. Performance tuning may be required, depending on the results of tests. While the requirements have been approved on paper, experience with the software may cause stakeholders to emphasize/deemphasize various elements, or request additional features.

Timeline October -Detailed implementation planning. Infrastructure setup: twiki, issue tracker, code repository, build system, continuous integration and test system. Kickoff meeting. November -SQL for database creation. Start SQL for updates and queries. Shell wrappers for SQL. Pigeon integration. Installation scripts for testers. December -Finish SQL for updates and queries. Database performance testing/tuning. Setup webservice code: stubs, queuing mechanism, Java SQL configuration. January -Web service installation script. Java SQL wrappers. Java methods for functions. SRM methods for transfer commands. February -GSI authentication for web service. Performance testing of web service. End-to-end load testing. Reports capability. March -VDT packaging. ITB testing. Post-facto registration and cleanup. RPMs for Operations Toolkit.