Tier-1 Data Storage Challenges Extreme Data Workshop Andrew Sansum 20 th April 2012.

Slides:



Advertisements
Similar presentations
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
Advertisements

Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow GridPP Vendor Day 30 th April.
Tier-1 Overview Andrew Sansum 21 November Overview of Presentations Morning Presentations –Overview (Me) Not really overview – at request of Tony.
Copyright © 2010 Platform Computing Corporation. All Rights Reserved.1 The CERN Cloud Computing Project William Lu, Ph.D. Platform Computing.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
GridPP Steve Lloyd, Chair of the GridPP Collaboration Board.
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
The Grid Prof Steve Lloyd Queen Mary, University of London.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
Southgrid Technical Meeting Pete Gronbech: 16 th March 2006 Birmingham.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
Monitoring the Grid at local, national, and Global levels Pete Gronbech GridPP Project Manager ACAT - Brunel Sept 2011.
GridPP3 Project Management GridPP20 Sarah Pearce 11 March 2008.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
09/02 ID099-1 September 9, 2002Grid Technology Panel Patrick Dreher Technical Panel Discussion: Progress in Developing a Web Services Data Analysis Grid.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
Managing the CERN LHC Tier0/Tier1 centre Status and Plans March 27 th 2003 CERN.ch.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
UK Tier 1 Centre Glenn Patrick LHCb Software Week, 28 April 2006.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
Tier-1 Andrew Sansum Deployment Board 12 July 2007.
STATUS OF KISTI TIER1 Sang-Un Ahn On behalf of the GSDC Tier1 Team WLCG Management Board 18 November 2014.
The RAL Tier-1 and the 3D Deployment Andrew Sansum 3D Meeting 22 March 2006.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Development of a Tier-1 computing cluster at National Research Centre 'Kurchatov Institute' Igor Tkachenko on behalf of the NRC-KI Tier-1 team National.
IT-DSS Alberto Pace2 ? Detecting particles (experiments) Accelerating particle beams Large-scale computing (Analysis) Discovery We are here The mission.
A Computing Tier 2 Node Eric Fede – LAPP/IN2P3. 2 Eric Fede – 1st Chinese-French Workshop Plan What is a Tier 2 –Context and definition To be a Tier 2.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
RAL Plans for SC2 Andrew Sansum Service Challenge Meeting 24 February 2005.
Activities and Perspectives at Armenian Grid site The 6th International Conference "Distributed Computing and Grid- technologies in Science and Education"
UK Status and Plans Catalin Condurache – STFC RAL ALICE Tier-1/Tier-2 Workshop University of Torino, February 2015.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
RAL Site Report HEP SYSMAN June 2016 – RAL Gareth Smith, STFC-RAL With thanks to Martin Bly, STFC-RAL.
LHC collisions rate: Hz New PHYSICS rate: Hz Event selection: 1 in 10,000,000,000,000 Signal/Noise: Raw Data volumes produced.
CASTOR at RAL in 2016 Rob Appleyard. Contents Current Status Staffing Upgrade plans Questions Conclusion.
CERN IT Department CH-1211 Geneva 23 Switzerland t Service Reliability & Critical Services January 15 th 2008.
T0-T1 Networking Meeting 16th June Meeting
ITIL: Service Transition
WLCG IPv6 deployment strategy
WLCG Tier-2 Asia Workshop TIFR, Mumbai 1-3 December 2006
Experience of Lustre at QMUL
Ian Bird WLCG Workshop San Francisco, 8th October 2016
LCG Service Challenge: Planning and Milestones
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
Grid site as a tool for data processing and data analysis
HEPiX Spring 2014 Annecy-le Vieux May Martin Bly, STFC-RAL
Data Challenge with the Grid in ATLAS
Christos Markou Institute of Nuclear Physics NCSR ‘Demokritos’
Update on Plan for KISTI-GSDC
Experience of Lustre at a Tier-2 site
Dagmar Adamova, NPI AS CR Prague/Rez
UK GridPP Tier-1/A Centre at CLRC
Castor services at the Tier-0
Olof Bärring LCG-LHCC Review, 22nd September 2008
WLCG Service Interventions
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
UK Status and Plans Scientific Computing Forum 27th Oct 2017
Thoughts on Computing Upgrade Activities
Ákos Frohner EGEE'08 September 2008
LCG Operations Centres
Patrick Dreher Research Scientist & Associate Director
LHC Data Analysis using a worldwide computing grid
The Problem ~6,000 PCs Another ~1,000 boxes But! Affected by:
LHCb thinking on Regional Centres and Related activities (GRIDs)
Development of LHCb Computing Model F Harris
IPv6 update Duncan Rand Imperial College London
Presentation transcript:

Tier-1 Data Storage Challenges Extreme Data Workshop Andrew Sansum 20 th April 2012

RAL Tier-1 in a Nutshell Primary role – compute resources for LHC. Find Higgs etc Part of the global WLCG Grid (globally 386M jobs and 164PB disk in 2011) STFC grant to UK GRIDPP collaboration. Memorandum of Understanding between CERN and STFC High availability (>98%), annual hardware procurement cycle 11 July 2016 Tier-1 Status RAL – 10PB disk, 10PB tape Making global storage grid work is hard! Started “production” in Evolved through series of service challenges. First physics in 2009

Context: WLCG architecture 3 Institutes CERN computer centre RAL,UK ScotGridNorthGridSouthGridLondon FranceItalyGermanyUSA GlasgowEdinburghDurham Tier 0 Tier 1 National centres Tier 2 Regional groups Offline farm Online system Workstations Useful model for Particle Physics but not necessary for others RAL Tier-1

Storage Management Hierarchy 11 July 2016 Tier-1 Status VO data Management Generic Middleware FTS, LFC,MYPROXY Storage Elemement (site specific SRM/gridftp etc) Disk Pool Manager/HSM (eg CASTOR/dCache/DPM) Hardware

Challenge: Disk Hardware Experiments require: –High capacity (15PB by 2015) –High bandwidth (currently access disk at 50GBytes/second) Project requires low cost - £170 per TB CASTOR team prefer small allocation volumes (thin disk) Choose to buy many small disk servers Currently 490 servers 10PB disk capacity Only white box SATA RAID 6 16 or 24 disk drives

Challenges: Disk Drives Fail Tier-1 has about 10,000 disk drives deployed Observed drive failure rate 3-6% pa normally Expect a drive failure every day in normal operation Sometimes generations fail (see Vi-07-A since April) Weed out early life failures with aggressive load/acceptance test Drive failure exception monitoring Dedicated “Hardware technician” who replaces drives, repairs hardware etc. Failure rate monitoring to detect long term trends/departure from norm Standby capacity to allow emergency phase-out of a whole generation when necessary

Challenge: Disk Server Exceptions Experiments expect uninterrupted access to ALL their data, ALL of the time –Server failures can take small fractions of dataset offline for periods of hours to few days We observe a per server lifetime exception rate (excluding single drive failure) of about 1-2 (per server per 4 years) Leads to server interventions per month (not easy) Disk server exception monitoring and integrated, ticketing and callout system Require VO input regarding speed required and retention policy. Detailed intervention process to ensure successful hand off between teams Daily inter-team review meetings to track current exceptions

Challenge – Server Deployment Tracking state of ~490 disk servers is not easy. Servers need to progress smoothly through deployment system (don’t forget any, don’t miss critical step...) –Servers can be working, draining, waiting repair, repaired, certifying, ready to deploy, being deployed, deployed, etc etc Integrated management database to track state. Hardware tracking database integrated into deployment and monitoring systems. Detailed, documented, routine processes for deployment/removal Workflow managed through ticket system Regular deployment meetings to coordinate server moves

Challenge: Disk Procurement MoU commitments require timely hardware deployment Procurements search for disk solutions that are “High performance, high reliability, low cost”. –Target price currently £170/TB –Currently buy 16 or 24 drive white box disk servers Accept that purchased hardware does not always work as described/required (or even work at all). Seek companies strong on project management with good internal processes and links to manufacturers Procure from 2 suppliers with 2 different technical solutions Aggressive 4 week acceptance test Maintain operational buffer of 50% of 1 year procurement against problems

Challenge: Wide Area Network Require high bandwidth, low contention and resilience Team must be able to resolve complex performance problems on inter site transfers on deep middleware stack Access Tier-2s via production SJ5 Primary 2*10Gb/s to Reading Secondary 2*10Gb/s to London LHC Optical Private Network (OPN) for access to Tier-0 and Tier-1s Primary 10Gb/s to CERN Secondary 10Gb/s to CERN by alternative physical path Typically 3Gb/s (with 10Gb/s peaks)

Challenge: Local Area Network Need to accommodate: –500+ high bandwidth disk servers –500 worker nodes Must be low cost (hence commodity) Star topology 64 port Force10 C300 at centre 12 Nortel stacks with multi- 10Gb uplink Mix CPU nodes and worker nodes on same stacks to reduce inter stack traffic Routinely 20-40Gb/s with 100Gb/s peaks.

Challenge: Tape Robotics Experiments require 20PB long term “tape” storage by 2015 Capacity requirements evolve according to LHC schedule Experiment Bandwidth estimates/access patterns uncertain Migrating between generations can be slow (0.5PB/month) Keep planning flexible – respond to changes Avoid buying too early (experiments don’t always use what they say they will) “STK” SL8500 tape robot (2 nd non-Tier-1 robot) Currently using about 7500 tapes 3 generations of tape drives: 18 T10000A drives (LHCB+ALICE) 9 T10000B drives (CMS) 14 T10000C drives (ATLAS)

Challenge: CPU Hardware No real challenge! Very few deployment / operational problems with CPUs!! Usually just work!!! Mainly buy white box Typically 4 motherboards in chassis solution most cost effective at the moment.

Some last thoughts Moving data reliably between multiple sites for multiple experiments at high speeds is not easy! A global project needs a global community working together. Responsiveness and good communications are essential. Automation is vital. High capacity storage operation is not easy! It is possible to make commodity hardware work well in high volume but you need mature processes. The operational cost of managing commodity storage hardware faults is < 30% of hardware price. Spending more is unlikely to be cost effective for us. A storage system that is resilient against any single hardware or service failure makes operational life less stressful – we have a system that is only partly so. 11 July 2016