BaBar Data Distribution using the Storage Resource Broker Adil Hasan, Wilko Kroeger (SLAC Computing Services), Dominique Boutigny (LAPP), Cristina Bulfon.

Slides:



Advertisements
Similar presentations
RLS and DRS Roadmap Items Ann Chervenak Robert Schuler USC Information Sciences Institute.
Advertisements

OGF-23 iRODS Metadata Grid File System Reagan Moore San Diego Supercomputer Center.
Peter Berrisford RAL – Data Management Group SRB Services.
Data Management Expert Panel - WP2. WP2 Overview.
Database Architectures and the Web
FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
Plateforme de Calcul pour les Sciences du Vivant SRB & gLite V. Breton.
NATIONAL PARTNERSHIP FOR ADVANCED COMPUTATIONAL INFRASTRUCTURE SAN DIEGO SUPERCOMPUTER CENTER Particle Physics Data Grid PPDG Data Handling System Reagan.
Holding slide prior to starting show. Supporting Collaborative Working of Construction Industry Consortia via the Grid - P. Burnap, L. Joita, J.S. Pahwa,
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Chronopolis: Preserving Our Digital Heritage David Minor UC San Diego San Diego Supercomputer Center.
GRID DATA MANAGEMENT PILOT (GDMP) Asad Samar (Caltech) ACAT 2000, Fermilab October , 2000.
Data Grids: Globus vs SRB. Maturity SRB  Older code base  Widely accepted across multiple communities  Core components are tightly integrated Globus.
Jean-Yves Nief, CC-IN2P3 Wilko Kroeger, SCCS/SLAC Adil Hasan, CCLRC/RAL HEPiX, SLAC October 11th – 13th, 2005 BaBar data distribution using the Storage.
Robust Tools for Archiving and Preserving Digital Data Joseph JaJa, Mike Smorul, and Mike McGann Institute for Advanced Computer Studies Department of.
1 Use of the European Data Grid software in the framework of the BaBar distributed computing model T. Adye (1), R. Barlow (2), B. Bense (3), D. Boutigny.
UMIACS PAWN, LPE, and GRASP data grids Mike Smorul.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
Data Grid Web Services Chip Watson Jie Chen, Ying Chen, Bryan Hess, Walt Akers.
MongoDB Sharding and its Threats
Data Grid Interactions with Firewalls Michael Wan Reagan Moore SDSC/UCSD/NPACI.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
BaBar WEB job submission with Globus authentication and AFS access T. Adye, R. Barlow, A. Forti, A. McNab, S. Salih, D. H. Smith on behalf of the BaBar.
IRODS performance test and SRB system at KEK Yoshimi KEK Building data grids with iRODS 27 May 2008.
MCAT: A Metadata Catalog San Diego Supercomputing Center Part of the Storage Resource Broker (SRB)
Jan Storage Resource Broker Managing Distributed Data in a Grid A discussion of a paper published by a group of researchers at the San Diego Supercomputer.
Rule-Based Data Management Systems Reagan W. Moore Wayne Schroeder Mike Wan Arcot Rajasekar {moore, schroede, mwan, {moore, schroede, mwan,
INFSO-RI Enabling Grids for E-sciencE gLite Data Management Services - Overview Mike Mineter National e-Science Centre, Edinburgh.
1 School of Computer, National University of Defense Technology A Profile on the Grid Data Engine (GridDaEn) Xiao Nong
San Diego Supercomputer Center National Partnership for Advanced Computational Infrastructure San Diego Supercomputer Center National Partnership for Advanced.
MySQL and GRID Gabriele Carcassi STAR Collaboration 6 May Proposal.
28 April 2003Lee Lueking, PPDG Review1 BaBar and DØ Experiment Reports DOE Review of PPDG January 28-29, 2003 Lee Lueking Fermilab Computing Division D0.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
Production Data Grids SRB - iRODS Storage Resource Broker Reagan W. Moore
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
Data Distribution and Management Tim Adye Rutherford Appleton Laboratory BaBar Computing Review 9 th June 2003.
Globus Replica Management Bill Allcock, ANL PPDG Meeting at SLAC 20 Sep 2000.
The Replica Location Service The Globus Project™ And The DataGrid Project Copyright (c) 2002 University of Chicago and The University of Southern California.
User Management: Authentication & Authorization on the NorduGrid Balázs Kónya, AndersWäänänen 3 rd NorduGrid Workshop, 23 May, 2002 Helsinki.
The Global Land Cover Facility is sponsored by NASA and the University of Maryland.The GLCF is a founding member of the Federation of Earth Science Information.
GCRC Meeting 2004 BIRN Coordinating Center Software Development Vicky Rowley.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
26 September 2000Tim Adye1 Data Distribution Tim Adye Rutherford Appleton Laboratory BaBar Collaboration Meeting 26 th September 2000.
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
Introduction to The Storage Resource.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Data Management The European DataGrid Project Team
Rights Management for Shared Collections Storage Resource Broker Reagan W. Moore
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
The Storage Resource Broker and.
The National Grid Service User Accounting System Katie Weeks Science and Technology Facilities Council.
Directory Services CS5493/7493. Directory Services Directory services represent a technological breakthrough by integrating into a single management tool:
Collection-Based Persistent Archives Arcot Rajasekar, Richard Marciano, Reagan Moore San Diego Supercomputer Center Presented by: Preetham A Gowda.
A possible SRB to iRODS Migration (a path from here to there)‏ Adil Hasan RAL.
Enabling Grids for E-sciencE EGEE-II INFSO-RI Status of SRB/SRM interface development Fu-Ming Tsai Academia Sinica Grid Computing.
Jean-Philippe Baud, IT-GD, CERN November 2007
Performance measurement of transferring files on the federated SRB
BaBar-Grid Status and Prospects
Oxana Smirnova, Jakob Nielsen (Lund University/CERN)
The Data Grid: Towards an architecture for Distributed Management
U.S. ATLAS Grid Production Experience
GGF OGSA-WG, Data Use Cases Peter Kunszt Middleware Activity, Data Management Cluster EGEE is a project funded by the European.
Joseph JaJa, Mike Smorul, and Sangchul Song
Arcot Rajasekar Michael Wan Reagan Moore (sekar, mwan,
Integrating SRB with the GIGGLE framework
Gridifying the LHCb Monte Carlo production system
Presentation transcript:

BaBar Data Distribution using the Storage Resource Broker Adil Hasan, Wilko Kroeger (SLAC Computing Services), Dominique Boutigny (LAPP), Cristina Bulfon (INFN,Rome), Jean-Yves Nief (ccin2p3), Liliana Martin (Paris VI et VII), Andreas Petzold (TUD), Jim Cochran (ISU) (on behalf of the BaBar Computing Group) IX International Workshop on Advanced Computing and Analysis Techniques in Physics Research KEK Japan 1-5 December

BaBar – the parameters (computing-wise) ~80 institutions in Europe and North America. 5 Tier A computing centers: –SLAC (USA), ccin2p3 (France), RAL (UK), GridKA (Germany), Padova (Italy). Processing of data done in Padova. Bulk of simulation production done by remote institutions. BaBar computing is highly distributed. Reliable data distribution essential to BaBar. 2

The SRB The Storage Resource Broker (SRB) is developed by San Diego Supercomputing Center (SDSC). A client-server middleware for connecting heterogeneous data resources. Provides a uniform method to access the resources. Provides relational database backend to record file metadata (metadata catalog called MCAT) and for access control lists (acls). Can use Grid Security Infrastructure (GSI) authentication. Also provides Audit information. 3

The SRB SRB v3: –Define an SRB zone comprising of one MCAT and one or more SRB servers. –Provides applications to federate zones (synch MCATs, create users, data belonging to different zones). –Within one federation all SRB servers need to run on the same port. –Allows an SRB server at one site to belong to more than one zone. 4

The SRB in BaBar The SRB feature-set makes it a useful tool for data distribution. Particle Physics Data Grid (PPDG) effort initiated interest in SRB. PPDG and BaBar collaboration effort has gone into testing and deploying the SRB in BaBar. 5

The SRB in BaBar The BaBar system has 2 MCATs: one at SLAC and one at ccin2p3. Use SRB v3 to create and federate the two zones: SLAC and ccin2p3 zone. Advantage that client can connect to SLAC or ccin2p3 to see files at other site. 6

The SRB in BaBar MCAT enabled SRB server SRB server SRB clients Data replicate from/copied to SLAC Data copied from SLAC Data replicate from/copied to ccin2p3 Data copied From ccin2p3 SLAC ccin2p3 SLAC Zoneccin2p3 Zone 7

Data Distribution using SRB BaBar Data distribution with SRB consists of the following steps: –Publish files available for distribution in MCAT (publication I). –Locate files to distribute (location). –Distribute files (distribution). –Publish distributed files (publication II). Each of these steps requires the user to belong to some ACL (authorization). 8

Authorization BaBarGrid currently uses European Data Grid Virtual Organization (VO). –Consists of an Lightweight Directory Access Protocol (LDAP) database holding certificate Distinguished Name (DN) strings for all BaBar members. –Used to update Globus grid-mapfiles. SRB authentication akin to grid-mapfile: –Maps SRB username to DN string. –SRB username doesn’t have to map to UNIX username. Developing application to obtain user DN strings from VO. –App is experiment neutral. –Has ability to include information from Virtual Organization Management System. 9

Publication I The initial publication step (event store files) entails: –Publication of files into SRB MCAT once files have been produced and published in BaBar bookkeeping. –Files are grouped into collections based on run range, release, production type (SRB collection != BaBar collection). –Extra metadata information (such as file UUID, BaBar collection name) stored in MCAT. –SRB object name contains processing spec, etc that uniquely id the object. –~5K event files (or SRB objects) per SRB collection. 10

Publication I Detector conditions files are more complicated as files are constantly updated (ie not closed). As files are update in SRB need to prevent users from taking an inconsistent copy. Unfortunately SRB does not currently permit locking of collections. 11

Publication I Have devised a workaround: –Register conditions file objects under date-specified collection. –Register a ‘locator file’ object containing the conditions date-specified collection name. –Then, new conditions files registered under a new date- specified collection. –‘Locator file’ contents updated with new date-specified collection. This method prevents users from taking an inconsistent set of files. Only two sets kept at any one time. 12

Location & Distribution Location and distribution happen in one client application. User supplies BaBar collection name from BaBar bookkeeping. SRB searches MCAT for files that have that collection name as metadata. Files are then copied to target site. SRB allows simple checksum to be performed. –But checksum is not md5 or cksum. –Still can be useful. 13

Location & Distribution SRB allows 3 rd -party replication. –But, most likely we will always run distribution command from source or target site. Also have the ability to create a logical resource of more than 1 physical resource. –Can replicate to all resources with one command. –Useful if more than 1 site regularly needs the data. 14

Publication II Optionally can register copied file in MCAT (decision a matter of policy). Extra step for data distribution to ccin2p3: –Publication of files in ccin2p3 MCAT. –Required since current SRB v3 does not allow replication across zones. Extra step not a problem since need to integrity check data before publishing anyway. Important note: data can be published & accessed at ccin2p3 or SLAC since MCATs will be synch’d regularly. 15

SC2003 demonstration Demonstrated distribution of detector conditions files using scheme previously described to 5 sites: –SLAC, ccin2p3, Rome, Bristol, Iowa State. Data were distributed over 2 servers at SLAC and files were copied in a round-robin manner from each server. Files were continuously copied and deleted at target site. Demonstration ran 1 full week continuously without problems. 16

SC2003 Demonstration RomeCcin2p3... Etc SRB server MCAT Authenticate user Locate data Request to transfer data Request to transfer data Transfer data to target Transfer data to target 17

SC 2003 demonstration 18

Future work System currently being used, but not yet considered full production quality. Missing items: –SRB log file parser to automatically catch errors. –SRB server load monitor (cpu, memory). –Automatic generation of SRB.MdasEnv and.MdasAuth files. –Automatic generation of new and deletion of old users in MCAT. –Better packaging of client and server apps. –MCAT integrity checking scripts. –Better integration with BaBar Bookkeeping. 19

Future work Slightly longer term : –Inclusion of management system to manage SRB requests. If system heavily used will require system to queue requests. Currently looking at Stork to manage multiple requests ( –Interoperation with Replica Location Service (RLS) and Storage Resource Manager (SRM) (see Simon Metson’s talk). Allows integration with LCG tools. –Move to grid-services. 20

Summary Extensive testing and interaction with SRB developers has allowed BaBar to develop a data distribution system based on existing grid middle- ware. Used for distributing conditions files to 5 sites since October. Will be used to distribute: –Detector conditions files. –Event store files. –Random trigger files (used for simulation). BaBar’s data distribution system is sufficiently modular can be adapted to other environments. 21