Storage Management on the Grid Alasdair Earl University of Edinburgh.

Slides:



Advertisements
Similar presentations
30-31 Jan 2003J G Jensen, RAL/WP5 Storage Elephant Grid Access to Mass Storage.
Advertisements

S.L.LloydATSE e-Science Visit April 2004Slide 1 GridPP – A UK Computing Grid for Particle Physics GridPP 19 UK Universities, CCLRC (RAL & Daresbury) and.
The Quantum Chromodynamics Grid James Perry, Andrew Jackson, Matthew Egbert, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Tony Doyle - University of Glasgow GridPP EDG - UK Contributions Architecture Testbed-1 Network Monitoring Certificates & Security Storage Element R-GMA.
The Grid What is it? what is it for?. Your university or experiment logo here Web: information sharing Invented at CERN by Tim Berners-Lee No. of Internet.
Jens G Jensen Atlas Petabyte store Supporting Multiple Interfaces to Mass Storage Providing Tape and Mass Storage to Diverse Scientific Communities.
Level 1 Components of the Project. Level 0 Goal or Aim of GridPP. Level 2 Elements of the components. Level 2 Milestones for the elements.
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Peter Berrisford RAL – Data Management Group SRB Services.
NATIONAL PARTNERSHIP FOR ADVANCED COMPUTATIONAL INFRASTRUCTURE SAN DIEGO SUPERCOMPUTER CENTER Particle Physics Data Grid PPDG Data Handling System Reagan.
Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
Alain Romeyer - Dec Grid computing for CMS What is the Grid ? Let’s start with an analogy How it works ? (Some basic ideas) Grid for LHC and CMS.
DataGrid is a project funded by the European Union 22 September 2003 – n° 1 EDG WP4 Fabric Management: Fabric Monitoring and Fault Tolerance
Parallel Programming on the SGI Origin2000 With thanks to Moshe Goldberg, TCC and Igor Zacharov SGI Taub Computer Center Technion Mar 2005 Anne Weill-Zrahia.
Jean-Yves Nief, CC-IN2P3 Wilko Kroeger, SCCS/SLAC Adil Hasan, CCLRC/RAL HEPiX, SLAC October 11th – 13th, 2005 BaBar data distribution using the Storage.
POLITEHNICA University of Bucharest California Institute of Technology National Center for Information Technology Ciprian Mihai Dobre Corina Stratan MONARC.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
Edinburgh University Experimental Particle Physics Alasdair Earl PPARC eScience Summer School September 2002.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow GridPP Vendor Day 30 th April.
Copyright © 2010 Platform Computing Corporation. All Rights Reserved.1 The CERN Cloud Computing Project William Lu, Ph.D. Platform Computing.
Welcome e-Science in the UK Building Collaborative eResearch Environments Prof. Malcolm Atkinson Director 23 rd February 2004.
RomeWorkshop on eInfrastructures 9 December LCG Progress on Policies & Coming Challenges Ian Bird IT Division, CERN LCG and EGEE Rome 9 December.
GridPP9 – 5 February 2004 – Data Management DataGrid is a project funded by the European Union GridPP is funded by PPARC WP2+5: Data and Storage Management.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
The Grid Prof Steve Lloyd Queen Mary, University of London.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
Grid Infrastructure for the ILC Andreas Gellrich DESY European ILC Software and Physics Meeting Cambridge, UK,
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
File and Object Replication in Data Grids Chin-Yi Tsai.
Data Management The GSM-WG Perspective. Background SRM is the Storage Resource Manager A Control protocol for Mass Storage Systems Standard protocol:
Your university or experiment logo here Caitriana Nicholson University of Glasgow Dynamic Data Replication in LCG 2008.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
LHC Computing Plans Scale of the challenge Computing model Resource estimates Financial implications Plans in Canada.
The Grid System Design Liu Xiangrui Beijing Institute of Technology.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
Data Logistics in Particle Physics Ready or Not, Here it Comes… Prof. Paul Sheldon Vanderbilt University Prof. Paul Sheldon Vanderbilt University.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
San Diego Supercomputer Center National Partnership for Advanced Computational Infrastructure SRB + Web Services = Datagrid Management System (DGMS) Arcot.
Grid Glasgow Outline LHC Computing at a Glance Glasgow Starting Point LHC Computing Challenge CPU Intensive Applications Timeline ScotGRID.
GridPP Building a UK Computing Grid for Particle Physics Professor Steve Lloyd, Queen Mary, University of London Chair of the GridPP Collaboration Board.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
…building the next IT revolution From Web to Grid…
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
Eine Einführung ins Grid Andreas Gellrich IT Training DESY Hamburg
Grid Glasgow Outline LHC Computing at a Glance Glasgow Starting Point LHC Computing Challenge CPU Intensive Applications Timeline ScotGRID.
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
Introduction to The Storage Resource.
30 June Wide Area Networking Performance Challenges Olivier Martin, CERN UK DTI visit.
Computing for LHC Physics 7th March 2014 International Women's Day - CERN- GOOGLE Networking Event Maria Alandes Pradillo CERN IT Department.
Data Processing and the LHC Computing Grid (LCG) Jamie Shiers Database Group, IT Division CERN, Geneva, Switzerland
The Storage Resource Broker and.
1 GridFTP and SRB Guy Warner Training, Outreach and Education Team, Edinburgh e-Science.
GridPP2 Data Management work area J Jensen / RAL GridPP2 Data Management Work Area – Part 2 Mass storage & local storage mgmt J Jensen
Enabling Grids for E-sciencE EGEE-II INFSO-RI Status of SRB/SRM interface development Fu-Ming Tsai Academia Sinica Grid Computing.
Hall D Computing Facilities Ian Bird 16 March 2001.
] Open Science Grid Ben Clifford University of Chicago
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
18/12/03PPD Christmas Lectures 2003 Grid in the Department A Guide for the Uninvolved PPD Computing Group Christmas Lecture 2003 Chris Brew.
Dagmar Adamova, NPI AS CR Prague/Rez
UK GridPP Tier-1/A Centre at CLRC
LHC Collisions.
CS258 Spring 2002 Mark Whitney and Yitao Duan
LHCb(UK) Computing Status Glenn Patrick
Gridifying the LHCb Monte Carlo production system
LHCb thinking on Regional Centres and Related activities (GRIDs)
Presentation transcript:

Storage Management on the Grid Alasdair Earl University of Edinburgh

Talk outline Motivation Storage management issues Existing projects Current work Conclusion

CERN capacity requirements

Cost Initially record ~4PB / year 1 PB = 1 Tape silo 1 silo = £500k Media = £600k + overhead … Tape is the cheapest option but has issues Grid is the solution LHC computing expensive

LHC computing challenge Tier2 Centre ~1000 PCs Online System Offline Farm ~20,000 PCs CERN Computer Centre >20,000 PCs RAL Regional Centre US Regional Centre French Regional Centre Italian Regional Centre Institute Institute ~200 PCs Workstations ~100 Mbyte/sec ~100 MByte/sec Mbit/sec one bunch crossing per 25 ns 100 triggers per second each event is ~1 Mbyte physicists work on analysis “channels” each institute has ~10 physicists working on one or more channels data for these channels is cached by the institute server Physics data cache ~PByte/sec ~ Gbit/sec or Air Freight Tier2 Centre ~1000 PCs ~Gbit/sec Tier 0 Tier 1 Tier 3 Tier 4 assumes PC = ~ 25 SpecInt95 ScotGRID++ ~1000 PCs Tier 2

Competing requirements A durable, long term store for large amounts of data –Very long term data (centuries) –Very reliable storage (as good as paper) Metadata –Overview / administration information –Individual file information A security mechanism for access and modification Online access to data

Requesting data from a storage system user index Data storage system(s) Staging area Request manager

Storage Element Storage for European DataGrid Developed primarily by RAL under EDG Work Package 5 Metadata in XML –aiming for OGSA compliance Used by physics projects WP5 Logo

Storage Element Protocols: FTP, HTTP(s), GridFTP, scp … MSS: Mass Storage System - disk, tape, hierarchy … Old ModelNew Model

Storage Resource Broker Part of the NSF funded Data Intensive Computing Environments (DICE) project San Diego Supercomputing Center Centralised metadata system –Dublin Core In use with CMS and BaBar Large following in BioInformatics apps

Storage Resource Broker SRB MCAT DB SRB 1 2 3

Storage Resource Manager Aim: High performance data storage –Teragrid –Supercomputing apps (high data rate) Implemented by LBNL and FNAL 2 standards available –Process based –Web Services

Storage Resource Manager

Current work Grid storage is currently specialised –Each project has advantages and disadvantages Need for interoperation between existing projects UK interest –Interoperation of existing resources –National level –Local level

Current Work ScotGrid –Durham –Edinburgh –Glasgow Rollout of storage management software Discussions for joint project Strong push to solve interoperation issues Scottish Tier 2 Centre

Conclusions We have presented 3 major projects Storage management on the Grid is rapidly advancing The existing project teams are working together Edinburgh trying to solve some of the interoperation issues