ATLAS Great Lakes Tier-2 Site Report Ben Meekhof USATLAS Tier2/Tier3 Workshop SLAC, Menlo Park, CA November 30 th, 2007.

Slides:



Advertisements
Similar presentations
Report of Liverpool HEP Computing during 2007 Executive Summary. Substantial and significant improvements in the local computing facilities during the.
Advertisements

S.Chechelnitskiy / SFU Simon Fraser Running CE and SE in a XEN virtualized environment S.Chechelnitskiy Simon Fraser University CHEP 2007 September 6 th.
DOSAR Workshop VI April 17, 2008 Louisiana Tech Site Report Michael Bryant Louisiana Tech University.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
Site Report US CMS T2 Workshop Samir Cury on behalf of T2_BR_UERJ Team.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
Virtual Desktop Infrastructure Solution Stack Cam Merrett – Demonstrator User device Connection Bandwidth Virtualisation Hardware Centralised desktops.
Condor at Brookhaven Xin Zhao, Antonio Chan Brookhaven National Lab CondorWeek 2009 Tuesday, April 21.
1 INDIACMS-TIFR TIER-2 Grid Status Report IndiaCMS Meeting, Sep 27-28, 2007 Delhi University, India.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Windows Server MIS 424 Professor Sandvig. Overview Role of servers Performance Requirements Server Hardware Software Windows Server IIS.
Tier 3g Infrastructure Doug Benjamin Duke University.
Data oriented job submission scheme for the PHENIX user analysis in CCJ Tomoaki Nakamura, Hideto En’yo, Takashi Ichihara, Yasushi Watanabe and Satoshi.
UCL Site Report Ben Waugh HepSysMan, 22 May 2007.
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Southgrid Technical Meeting Pete Gronbech: 16 th March 2006 Birmingham.
Site Lightning Report: MWT2 Mark Neubauer University of Illinois at Urbana-Champaign US ATLAS Facilities UC Santa Cruz Nov 14, 2012.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
David Hutchcroft on behalf of John Bland Rob Fay Steve Jones And Mike Houlden [ret.] * /.\ /..‘\ /'.‘\ /.''.'\ /.'.'.\ /'.''.'.\ ^^^[_]^^^ * /.\ /..‘\
BINP/GCF Status Report BINP LCG Site Registration Oct 2009
GridKa SC4 Tier2 Workshop – Sep , Warsaw Tier2 Site.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
INDIACMS-TIFR Tier 2 Grid Status Report I IndiaCMS Meeting, April 05-06, 2007.
Sandor Acs 05/07/
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
Support in setting up a non-grid Atlas Tier 3 Doug Benjamin Duke University.
26SEP03 2 nd SAR Workshop Oklahoma University Dick Greenwood Louisiana Tech University LaTech IAC Site Report.
São Paulo Regional Analysis Center SPRACE Status Report 22/Aug/2006 SPRACE Status Report 22/Aug/2006.
ALMA Archive Operations Impact on the ARC Facilities.
Brookhaven Analysis Facility Michael Ernst Brookhaven National Laboratory U.S. ATLAS Facility Meeting University of Chicago, Chicago 19 – 20 August, 2009.
Rob Allan Daresbury Laboratory NW-GRID Training Event 25 th January 2007 Introduction to NW-GRID R.J. Allan CCLRC Daresbury Laboratory.
ATLAS Great Lakes Tier-2 (AGL-Tier2) Shawn McKee (for the AGL Tier2) University of Michigan US ATLAS Tier-2 Meeting at Harvard Boston, MA, August 17 th,
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
Virtualization Supplemental Material beyond the textbook.
UKI-SouthGrid Overview and Oxford Status Report Pete Gronbech SouthGrid Technical Coordinator HEPSYSMAN – RAL 10 th June 2010.
US ATLAS Western Tier 2 Status Report Wei Yang Nov. 30, 2007 US ATLAS Tier 2 and Tier 3 workshop at SLAC.
PERFORMANCE AND ANALYSIS WORKFLOW ISSUES US ATLAS Distributed Facility Workshop November 2012, Santa Cruz.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Florida Tier2 Site Report USCMS Tier2 Workshop Livingston, LA March 3, 2009 Presented by Yu Fu for the University of Florida Tier2 Team (Paul Avery, Bourilkov.
Final Implementation of a High Performance Computing Cluster at Florida Tech P. FORD, X. FAVE, K. GNANVO, R. HOCH, M. HOHLMANN, D. MITRA Physics and Space.
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
PROOF tests at BNL Sergey Panitkin, Robert Petkus, Ofer Rind BNL May 28, 2008 Ann Arbor, MI.
BNL Oracle database services status and future plans Carlos Fernando Gamboa, John DeStefano, Dantong Yu Grid Group, RACF Facility Brookhaven National Lab,
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
AGLT2 Site Report Shawn McKee University of Michigan March / OSG-AHM.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
Storage at the ATLAS Great Lakes Tier-2 Tier-2 Storage Administrator Talks Shawn McKee / University of Michigan OSG Storage ForumShawn McKee1.
STORAGE EXPERIENCES AT MWT2 (US ATLAS MIDWEST TIER2 CENTER) Aaron van Meerten University of Chicago Sarah Williams Indiana University OSG Storage Forum,
Bob Ball/University of Michigan
The Beijing Tier 2: status and plans
Luca dell’Agnello INFN-CNAF
AGLT2 Site Report Shawn McKee/University of Michigan
HC Hyper-V Module GUI Portal VPS Templates Web Console
Chapter 13: I/O Systems “The two main jobs of a computer are I/O and [CPU] processing. In many cases, the main job is I/O, and the [CPU] processing is.
Presentation transcript:

ATLAS Great Lakes Tier-2 Site Report Ben Meekhof USATLAS Tier2/Tier3 Workshop SLAC, Menlo Park, CA November 30 th, 2007

AGLT2 Capacity Overview Dedicated hardware for AGLT2 ran first jobs in April, 2007Dedicated hardware for AGLT2 ran first jobs in April, 2007 Cycles from mix of dual quad-core Xeons and dual dual-core OpteronsCycles from mix of dual quad-core Xeons and dual dual-core Opterons Currently have ~683 available job slots, each with 2GB of RAM (UM site all T3/T2) Aglt2.org address space includes MSU and UM sites - MSU is nearly online with their recent purchase:Aglt2.org address space includes MSU and UM sites - MSU is nearly online with their recent purchase: Processing of AGLT2 jobs at MSU site ready by mid-December. Panda submission will transparently distribute across both physical sites. Fault-tolerant disk storage capacity: 86TB online now, +160TB ready.Fault-tolerant disk storage capacity: 86TB online now, +160TB ready. 120 TB raw disk now + 54 TB soon at MSU (utilized via dCache).120 TB raw disk now + 54 TB soon at MSU (utilized via dCache). Network connectivity via dual 10Gb links over MiLRNetwork connectivity via dual 10Gb links over MiLR With MSU site online, total Si2K for AGLT2 about 2,355,132. Total storage dedicated to Tier2 nearly 500TB.With MSU site online, total Si2K for AGLT2 about 2,355,132. Total storage dedicated to Tier2 nearly 500TB. Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 2

UM Cluster Composition CPU Details Compute nodesCompute nodes 79 dual quad-core 2.66Ghz Xeons in Dell PE1950, 16GB RAM 48 configured with 2 x 750GB SATA disks – striped into 1.2TB RAID0 volume, used as dCache pool. 51 configured with 2 x 750GB SATA disks – JBOD, each disk a dCache pool. 10 dual dual-core Opteron 285, 8GB RAM, 4 250GB disks Configured with 3 disks for dCache pools, remainder as system/tmp space. All with Tier-3 priority, but…. Per MOU, Tier-2 uses half of this resource Prioritized access handled by condor configuration 1,412,236 possible Si2K dedicated to Tier2 Dual Gb NICs (1 private, 1 public) New compute nodes: Good pricing - Dell 1950 dual quad-core Intel systems about $3,600. Purchased 52 at UM, 54 at MSU. Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 3

UM Cluster Composition Storage Details Dedicated Tier2 Disk StorageDedicated Tier2 Disk Storage ~120TB raw dCache + 40TB RAID50 dCache write pools. 46TB served via NFS, RAID5|6 volumes. 160TB RAID space ready for allocation as needed (using for tests while we can). 7.5TB for muon calibration files Dual NICs (1Gb private, 10Gb public) New storage node: Dell PE2950 with four MD1000 shelves (see supplemental diagram). Delivers 40TB RAID50 for just under $25,000. Purchased 6 at UM, 5 at MSU. Also purchased 3 PE2950 for ~$7000 with no MD1000 for use as new headnodes. Tier3 StorageTier3 Storage One 40TB PE2950 RAID50 storage node, NFS exported. Various smaller RAID6 volumes, 43TB. 83TB total available via NFS mounts. Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 4

New Storage Node Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 5

UM Rack Layout ~ November 2007 May 2007 Oct Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop

UM Rack Layout ~ November 2007 Oct Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop

UM Installation Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 8

MSU Cluster Composition UM and MSU submit bid request jointly, obtained same pricing.UM and MSU submit bid request jointly, obtained same pricing. Compute NodesCompute Nodes 54 PE1950 compute nodes 16GB RAM, Dual quad-core Xeon 5355 (2.67Ghz), 2 x 750GB SATA HD. No RAID, drives allocated with 2 dCache partitions and job /tmp space. Total SPECint_2000 at MSU = 940,896 Si2k.Total SPECint_2000 at MSU = 940,896 Si2k. Storage NodesStorage Nodes 5 Dell PE2950 storage nodes with MD1000 disk units and 225TB of raw space for allocation in AGLT2. Details of allocation, dCache vs NFS, not yet determined but will match UM Physical infrastructure done, 10G external network connected to world now. Compute nodes allocated (Rocks cluster tools used for allocation, as at UM).Physical infrastructure done, 10G external network connected to world now. Compute nodes allocated (Rocks cluster tools used for allocation, as at UM). Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 9

MSU Site Power: 150 KW Liebert PDUPower: 150 KW Liebert PDU Cooling: 30 Ton Liebert AirhandlerCooling: 30 Ton Liebert Airhandler additional unit to be added early next year Safety: Vesda smoke systemSafety: Vesda smoke system Size: 400 Sq. Feet, 20 Racks.Size: 400 Sq. Feet, 20 Racks. Despite enormous amounts of frustration and investment of time/money, everything has come through and MSU has a very nice facility. Many congratulations to MSU people on overcoming the challenges renovating the space - we are installed, networked, and ready to turn on!Despite enormous amounts of frustration and investment of time/money, everything has come through and MSU has a very nice facility. Many congratulations to MSU people on overcoming the challenges renovating the space - we are installed, networked, and ready to turn on! Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 10

MSU Rack Layout – November 2007 Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 11

MSU Installation Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 12

MSU Installation Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 13

Network Equipment Network switches: Dell PowerConnect stacking modules (48Gbps) 48 copper 10/100/1000 ports and two 10 gig module slots. UM has 8 switches in stacks of 2 and 6 units. MSU purchased 6 units, 5 stacked and 1 spare. Pricing under $3000 per unit, including stacking/uplink modules and CX cables. Reliability problems at UM site. Switch has many software bugs exacerbated by large stacks and complex VLAN arrangements. Some hardware failures (2 of 8 units at UM since April 07). Dell support not prepared for complex configurations, and often unreceptive and/or unwilling to acknowledge that the switch is not behaving as documented. Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 14

Network Architecture Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 15

Remote Site Access Raritan KVM for console access to non-Dell machines.Raritan KVM for console access to non-Dell machines. DRAC access to Dell consoles – unfortunately IE only.DRAC access to Dell consoles – unfortunately IE only. DRAC Virtual Media – remote installations, works simply and reliably.DRAC Virtual Media – remote installations, works simply and reliably. Serial Console IP servers for access to switches.Serial Console IP servers for access to switches. APC power infrastructure, remotely switch outlets.APC power infrastructure, remotely switch outlets. Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 16

FY08 Plans for AGLT2 September 2007 UM/MSU combined spent around on storage/compute:September 2007 UM/MSU combined spent around $625K on storage/compute: 11 storage nodes ~ 460 TB, 106 compute nodes for ~ 1635 kSI2K additional CPU UM/MSU - tentatively planning to purchase in Spring, aroundUM/MSU - tentatively planning to purchase in Spring, around $400K-$500K If we spend the equipment money for the whole next fiscal year AND use some of our local leveraged funds Anticipate being able to acquire similar capability to the last round of purchases MSU – capable of significant expansion because of augmented cooling and space beyond original plansMSU – capable of significant expansion because of augmented cooling and space beyond original plans Support for up to 13 more racks of similar equipment with planned cooling expansions Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop

Software Access pacman kit installation to afs spacepacman kit installation to afs space Mirror for Tier-3 access elsewhere OSG installed to AFS where read-only access requiredOSG installed to AFS where read-only access required OSG with VDT 1.6.1i Soft links back to local disk for configs and needed write access Gatekeepers upgraded to OSG 0.8.0, VDT 1.8.1b via local installation. Gatekeeper is gate01.aglt2.orgGatekeeper is gate01.aglt2.org Condor v6.9.4 collector located on cluster head node umopt1 gate02.aglt2.org is backup gatekeeper OS is SLC4.4 or 4.5OS is SLC4.4 or 4.5 Kernel is customized for AGLT2 and UltraLight UL3smp 64-bit kernel with 32-bit compatibility libraries for code Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 18

Resource Allocation Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop We use Condor job scheduler. Policy on Tier-2 is:We use Condor job scheduler. Policy on Tier-2 is: Highest priority for usatlas2 (admin, software install) Next highest priority for usatlas1 (ATLAS production) Next highest priority for usatlas3 (US ATLAS users) Lowest priority for OSG (including usatlas4) The Tier-3 is “integrated” with the Tier-2 provisioning (ROCKS) but has different priorities reflecting 50% contribution to the Tier-2.The Tier-3 is “integrated” with the Tier-2 provisioning (ROCKS) but has different priorities reflecting 50% contribution to the Tier-2. If either Tier-2 or Tier-3 is idle, the other expands to fillIf either Tier-2 or Tier-3 is idle, the other expands to fill Equilibrium when all is busy at about 80% of job slots to Tier-2 OSG/usatlas4 limited to, at most, 16 coresOSG/usatlas4 limited to, at most, 16 cores This only applies to UM – MSU is still planning Tier3 structureThis only applies to UM – MSU is still planning Tier3 structure 19

AGLT2 Analysis Queues Analysis queue is setup and tested, generally idle.Analysis queue is setup and tested, generally idle. One dedicated 8-CPU Dell node. Bob Ball is working with other T2 sites to help setup. OU (Condor) and BU (PBS) volunteered first. All analysis queue sites are waiting for full set of AOD and NTUPLES needed for analysis jobs.All analysis queue sites are waiting for full set of AOD and NTUPLES needed for analysis jobs. Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 20

Local Monitors EnvironmentalEnvironmental APC for power, temp, humidity Ganglia for overall, fast view of performanceGanglia for overall, fast view of performance Perl script for details of running processes and immediate log access Hardware health and performance via snmp/cactiHardware health and performance via snmp/cacti OS logs to central, searchable syslog-ng serverOS logs to central, searchable syslog-ng server notification of most “out of range” violations notification of most “out of range” violations Following is a sampling of available plotsFollowing is a sampling of available plots Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 21

Ganglia and Perl Job Monitors Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 22

Cacti Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 23

Syslog-ng Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 24

Performance Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 25

Problems Non ATLAS-wide AGLT2 outages from variety of sourcesNon ATLAS-wide AGLT2 outages from variety of sources Gate-keeper, head node crashes under heavy load - planning to move both systems to more reliable equipment Network outages – equipment issues. NFS server crashes - Disruptive, but not fatal. We seem to have no or few of these anymore on our newer systems with the latest kernel and nfs utilities. Software configurationSoftware configuration Recently upgraded OSG/VDT on our gatekeeper, site operations disrupted most of a day. Upgrade was successful, and we were back online the same day. Accounting issues due to broken Gratia, fixed October 2007 onwards.Accounting issues due to broken Gratia, fixed October 2007 onwards. Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 26

ATLAS Memory Limits We have seen jobs allocating 2.5GB or more of RAM.We have seen jobs allocating 2.5GB or more of RAM. Jobs like this cause nodes to use swap RAM, which can hugely slow the node down and cause loads to rise excessivelyJobs like this cause nodes to use swap RAM, which can hugely slow the node down and cause loads to rise excessively Some sites configure with no swap – nodes crash if a job allocates more than was planned for each job slot.Some sites configure with no swap – nodes crash if a job allocates more than was planned for each job slot. It is possible to have the scheduler (condor) kill jobs that exceed memory limitsIt is possible to have the scheduler (condor) kill jobs that exceed memory limits An ATLAS-wide policy regarding this would be a good idea.An ATLAS-wide policy regarding this would be a good idea. Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 27

Summary AGLT2 is in good shape and will only get better.AGLT2 is in good shape and will only get better. Accounting issues resolved as of Oct 2007.Accounting issues resolved as of Oct Dell switch problems responsible for some outages. Servers generally very reliable, though.Dell switch problems responsible for some outages. Servers generally very reliable, though. OSG/VDT is up to date on the gatekeepers.OSG/VDT is up to date on the gatekeepers. Lots of monitors allow quick response to problems. Flexible, cross- platform/browser remote access tools are essential.Lots of monitors allow quick response to problems. Flexible, cross- platform/browser remote access tools are essential. Expansion scheduled at MSU site is very close to being finished and processing jobs. Lots of additional expansion planned.Expansion scheduled at MSU site is very close to being finished and processing jobs. Lots of additional expansion planned. Review presentation by S. Mckee for more info about I/O performance tests performed at AGLT2. Review presentation by S. Mckee for more info about I/O performance tests performed at AGLT2. Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 28

Supplementary Slides Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 29

Performance – AGLT2 Ben Meekhof - AGLT2 - USATLAS Tier2/Tier3 Workshop 30