ATLAS Tier-2 Storage Status AGLT2 OSG Storage Forum – U Chicago – Sep 22 2010 Shawn McKee / University of Michigan OSG Storage ForumShawn McKee1.

Slides:



Advertisements
Similar presentations
NAGIOS AND CACTI NETWORK MANAGEMENT AND MONITORING SYSTEMS.
Advertisements

Status of BESIII Distributed Computing BESIII Workshop, Mar 2015 Xianghu Zhao On Behalf of the BESIII Distributed Computing Group.
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
Chapter 5: Server Hardware and Availability. Hardware Reliability and LAN The more reliable a component, the more expensive it is. Server hardware is.
Lesson 12 – NETWORK SERVERS Distinguish between servers and workstations. Choose servers for Windows NT and Netware. Maintain and troubleshoot servers.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment Chapter 8: Implementing and Managing Printers.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment, Enhanced Chapter 8: Implementing and Managing Printers.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment Chapter 8: Implementing and Managing Printers.
70-270, MCSE/MCSA Guide to Installing and Managing Microsoft Windows XP Professional and Windows Server 2003 Chapter Five Managing Disks and Data.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
Utilizing Condor and HTC to address archiving online courses at Clemson on a weekly basis Sam Hoover 1 Project Blackbird Computing,
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Windows Server MIS 424 Professor Sandvig. Overview Role of servers Performance Requirements Server Hardware Software Windows Server IIS.
1 The Google File System Reporter: You-Wei Zhang.
LAN / WAN Business Proposal. What is a LAN or WAN? A LAN is a Local Area Network it usually connects all computers in one building or several building.
GeoVision Solutions Storage Management & Backup. ๏ RAID - Redundant Array of Independent (or Inexpensive) Disks ๏ Combines multiple disk drives into a.
Section 11.1 Identify customer requirements Recommend appropriate network topologies Gather data about existing equipment and software Section 11.2 Demonstrate.
1 Guide to Novell NetWare 6.0 Network Administration Chapter 13.
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Hadoop Hardware Infrastructure considerations ©2013 OpalSoft Big Data.
StoRM Some basics and a comparison with DPM Wahid Bhimji University of Edinburgh GridPP Storage Workshop 31-Mar-101Wahid Bhimji – StoRM.
1 Computer and Network Bottlenecks Author: Rodger Burgess 27th October 2008 © Copyright reserved.
Multi-Tiered Storage with Xrootd at ATLAS Western Tier 2 Andrew Hanushevsky Wei Yang SLAC National Accelerator Laboratory 1CHEP2012, New York
08/30/05GDM Project Presentation Lower Storage Summary of activity on 8/30/2005.
Week #3 Objectives Partition Disks in Windows® 7 Manage Disk Volumes Maintain Disks in Windows 7 Install and Configure Device Drivers.
UMD TIER-3 EXPERIENCES Malina Kirn October 23, 2008 UMD T3 experiences 1.
Using NAS as a Gateway to SAN Dave Rosenberg Hewlett-Packard Company th Street SW Loveland, CO 80537
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Server Performance, Scaling, Reliability and Configuration Norman White.
BNL Facility Status and Service Challenge 3 HEPiX Karlsruhe, Germany May 9~13, 2005 Zhenping Liu, Razvan Popescu, and Dantong Yu USATLAS/RHIC Computing.
BNL Tier 1 Service Planning & Monitoring Bruce G. Gibbard GDB 5-6 August 2006.
ATLAS Great Lakes Tier-2 (AGL-Tier2) Shawn McKee (for the AGL Tier2) University of Michigan US ATLAS Tier-2 Meeting at Harvard Boston, MA, August 17 th,
Factors affecting ANALY_MWT2 performance MWT2 team August 28, 2012.
Shawn McKee/University of Michigan
VMware vSphere Configuration and Management v6
Cluster Consistency Monitor. Why use a cluster consistency monitoring tool? A Cluster is by definition a setup of configurations to maintain the operation.
Status SC3 SARA/Nikhef 20 juli Status & results SC3 throughput phase SARA/Nikhef Mark van de Sanden.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Florida Tier2 Site Report USCMS Tier2 Workshop Livingston, LA March 3, 2009 Presented by Yu Fu for the University of Florida Tier2 Team (Paul Avery, Bourilkov.
Experiments in Utility Computing: Hadoop and Condor Sameer Paranjpye Y! Web Search.
AGLT2 Site Report Shawn McKee University of Michigan March / OSG-AHM.
Tackling I/O Issues 1 David Race 16 March 2010.
Database CNAF Barbara Martelli Rome, April 4 st 2006.
1 CEG 2400 Fall 2012 Network Servers. 2 Network Servers Critical Network servers – Contain redundant components Power supplies Fans Memory CPU Hard Drives.
AGLT2 Site Report Shawn McKee/University of Michigan HEPiX Fall 2012 IHEP, Beijing, China October 14 th, 2012.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Hands-On Microsoft Windows Server 2008 Chapter 7 Configuring and Managing Data Storage.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
9/22/10 OSG Storage Forum 1 CMS Florida T2 Storage Status Bockjoo Kim for the CMS Florida T2.
AGLT2 Site Report Shawn McKee/University of Michigan Bob Ball, Chip Brock, Philippe Laurens, Ben Meekhof, Ryan Sylvester, Richard Drake HEPiX Spring 2016.
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
Storage at the ATLAS Great Lakes Tier-2 Tier-2 Storage Administrator Talks Shawn McKee / University of Michigan OSG Storage ForumShawn McKee1.
Atlas Tier 3 Overview Doug Benjamin Duke University.
STORAGE EXPERIENCES AT MWT2 (US ATLAS MIDWEST TIER2 CENTER) Aaron van Meerten University of Chicago Sarah Williams Indiana University OSG Storage Forum,
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
Status of BESIII Distributed Computing
High Availability Planning for Tier-2 Services
Bob Ball/University of Michigan
The Beijing Tier 2: status and plans
Joint AGLT2-MWT2 Networking meeting
AGLT2 Site Report Shawn McKee/University of Michigan
Experience of Lustre at a Tier-2 site
AGLT2 Site Report Shawn McKee/University of Michigan
Presentation transcript:

ATLAS Tier-2 Storage Status AGLT2 OSG Storage Forum – U Chicago – Sep Shawn McKee / University of Michigan OSG Storage ForumShawn McKee1

AGLT2 Storage: Outline TOverview of AGLT2 T Description, hardware and configuration TStorage Implementation T Original options: AFS, NFS, Lustre, dCache T Choose dCache (broad support in USATLAS/OSG) TIssues T Monitoring needed T Manpower to maintain (including maintaining custom scripts) T Reliability and performance TFuture storage systems T dCache or NFSv4(.x) or Lustre (1.8.x+) or Xrootd or ??? OSG Storage Forum2Shawn McKee

The ATLAS Great Lakes Tier-2 TThe ATLAS Great Lakes Tier-2 (AGLT2) is hosted at the University of Michigan and Michigan State University. T Currently about 2800 job slots, 1.16 PB of Tier-2 storage TWe have a flat model with a single “site” spanning two locations OSG Storage Forum3Shawn McKee

AGLT2 Design Components TBuild upon “commodity” Dell storage and compute components q Inexpensive building blocks q Deploy inexpensive high-capacity storage systems with large partitions using the XFS filesystem --- Still must address SRM q Storage node is a headnode RAID6 shelves (detailed later) q Incorporate 10GE networks for high-performance data transfers q Initially SR fiber (2006/7) q Then CX4 (copper) (2008/9) q Next SFP+ (copper, very inexpensive) (2010) q Evolving toward more 10GE ports as they become more inexpensive q Utilize 2.6 custom kernels (UltraLight) and SL5 x86_64 OSG Storage Forum4Shawn McKee

AGLT2 Storage Configuration TAGLT2 has 18 dCache storage nodes distributed between MSU and UM. Node storage sizes 40TB, 52TB and 104TB. TTotal dCache production storage is 1157TB TWe use space-tokens to manage space allocations for ATLAS and currently have 11 space-token areas. TWe use XFS for the file system on our 77 pools. Each pool varies from TB in size. TdCache version is using Chimera TWe also use AFS (1.4.12), NFSv3 and Lustre OSG Storage ForumShawn McKee5

Current AGLT2 Storage Node 1/2 UMFS05 OSG Storage Forum6Shawn McKee Current Node 2xE5450 processors 2xE5450 processors 32GB, 10GE Myricom CX4 1TB disks (15 disks/shelf) 4xRAID6 array (13TB each)

Current AGLT2 Storage Node 2/2 OSG Storage Forum7Shawn McKee Current Node 2xE5560 processors 2xE5560 processors 32GB, 10GEx2 Myricom CX4 2TB disks (15 disks/shelf) 4xRAID6 array (26TB each)

New AGLT2 Storage Node OSG Storage Forum8Shawn McKee New Node 2xE5620 processors 2xE5620 processors 48GB, 10GEx2 Intel X520 SFP+ 2TB disks (12 disks/shelf) 6xRAID6 array (20TB each) New node is 12-14U TB Uses redundant cabling Improves performance NL SAS disks (6 Gbps) instead of SATA Adding 700TB (Sep 2010)

AGLT2 Network Architecture Example TGood pricing from Dell for access layer switches: TOriginal PC6248/M6220 q Managed with 10GE ports, lots of 10/100/1000 copper ports q QoS and layer 3 capabilities, redundant power supplies qNew PC8024F q 24 10GE ports SFP+ q 4 “combo” ports (choose) q About $125/port! OSG Storage Forum9Shawn McKee

New UM Storage Network Backbone OSG Storage ForumShawn McKee10 New network will build upon Two 8024F switches. Uses “twinax” SFP+ cabling Four SR optics “uplinks” to Cisco. Four “twinax” links Between 8024F units. Each node has an SFP+ link to each 8024F Designed to inexpensively give us significantly increased inter-node bandwidth. MSU migrating toward a dual- uplink “star” network as well

dCache at AGLT2 TChoose dCache for a number of reasons q We wanted a way to tie multiple storage locations into a single user accessible name-space providing unified storage. q Broad support; BNL and OSG both support it. q Robust --- no major data loss AGLT2 in our few years of use q The price is right! TBut lots of manpower and attention has been required q Significant effort required to “watch for” and debug problems early q For a while we felt like the little Dutch boy…every time we fixed a problem, another popped up! q 1.9.5/Chimera has improved the situation. Much less labor intensive during the last year. q Transition from “hard” space-tokens by pool groups to a single large pool group was also a huge improvement for manageability. OSG Storage ForumShawn McKee11

AGLT2 Lustre Storage TWe have been exploring options for alternative storage systems. POSIX filesystems are interesting q “Direct” access to data using the filesystem q Broadly used and supported q Potentially good performance q BestMan can provide the needed SRM interface TLustre was originally tested by us in 2007: Too early! q In 2009 we setup a 200TB (4 server) Lustre configuration with dual MDT+MGS headnodes configured with heartbeat using V1.6.6 z Performance very good (faster for clients than /tmp) z Some “issues” with huge load on one storage server q In 2010 we have installed Lustre for our Tier-3 users q Recently upgraded to which fixed some issues we had q Uses 4 servers (each with 32GB RAM or more) and 10GE network q Needs more testing. Will soon add 2 more servers (OSS’s) q Total space 170TB, soon to added 144TB more OSG Storage ForumShawn McKee12

Monitoring Storage TMonitoring our Tier-2 is a critical part of our work and storage monitoring is one of the most important aspects TWe use a number of tools to track status: q Cacti q Ganglia q Custom web pages q alerting (low space, failed auto-tests, specific errors noted) TSee or OSG Storage ForumShawn McKee13

Scripts to Manage Storage TAs we have found problems we sometimes needed to automate the solution q Rephot – Automated replica creation developed by Wenjing Wu scans for “hot” files and automatically adds/removes replicas q Pool-balancer – Rebalances pools within a group TNumerous scripts (cron) used for maintenance q Directory ownership q Consistency checking, including Adler32 checksum verification q Repairing inconsistencies q Tracking usage (real space and space-token) q Auto-loading PNFSID into Chimera DB for faster access OSG Storage ForumShawn McKee14

Diagnosing AGLT2 Storage Issues TInformation is scattered across server nodes and log files TJava based components have verbose output which makes it hard to find the real problem TWith dCache we find the large number of components have a large phase-space for problems TErrors are not always indicative of the real problem… TIs syslog-ng a part of the solution? TCan error messages be significantly improved? OSG Storage ForumShawn McKee15

Inter-site Network Use OSG Storage ForumShawn McKee16 An important consideration for AGLT2 is the needed inter-site bandwidth. Storage and compute are roughly equally split between sites interconnected via a 10GE triangle. How long will this be sufficient given the increase in computing and storage located at each site?

Future AGLT2 Storage Alternatives TWe are interested in comparing our current dCache to BestMan+X where ‘X’ is q NFSv4 q Lustre q Xrootd TWe are interested in these characteristics for each option: q Expertise required to install/configure q Manpower required to maintain q Robustness over time q Performance (single and multiple read/writer) TOur plan is to test these during the coming year…how do they compare to dCache for us? OSG Storage Forum Shawn McKee17

Summary, Issues and Considerations TAGLT2 has a large storage component that is working well TStorage (and compute) is evolving/growing: q Concerns about power, space and cooling q …but also about needed “bandwidth” (network and I/O) keeping up TFor any hardware/technology/software always concerned about “manpower” costs. TData access model is important. Default of stage-in to local disk may not be the best model for much of our data, especially some kinds of analysis jobs. Testing direct access to dCache didn’t show any difference at AGLT2 but we didn’t use libdCap++. Ongoing work here TScaling the network, especially between sites, as the resources grow is an ongoing challenge. OSG Storage Forum Shawn McKee18

?Questions? OSG Storage ForumShawn McKee19