ATLAS Great Lakes Tier-2 (AGL-Tier2) Shawn McKee (for the AGL Tier2) University of Michigan US ATLAS Tier-2 Meeting at Harvard Boston, MA, August 17 th,

Slides:



Advertisements
Similar presentations
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH Home server AFS using openafs 3 DB servers. Web server AFS Mail Server.
Advertisements

CHEPREO Tier-3 Center Achievements. FIU Tier-3 Center Tier-3 Centers in the CMS computing model –Primarily employed in support of local CMS physics community.
Chris Brew RAL PPD Site Report Chris Brew SciTech/PPD.
GridKa SC4 Tier2 Workshop – Sep , Warsaw Tier2 Site Adam Padee ( ) Ryszard Gokieli ( ) Krzysztof.
Southwest Tier 2 Center Status Report U.S. ATLAS Tier 2 Workshop - Harvard Mark Sosebee for the SWT2 Center August 17, 2006.
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
Distributed IT Infrastructure for U.S. ATLAS Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
Computing Resources Joachim Wagner Overview CNGL Cluster MT Group Cluster School Cluster Desktop PCs.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
Site Report US CMS T2 Workshop Samir Cury on behalf of T2_BR_UERJ Team.
Site Report: ATLAS Great Lakes Tier-2 HEPiX 2011 Vancouver, Canada October 24 th, 2011.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
05/18/03Maurizio Davini Hepix2003 Department of Physics University of Pisa Site Report Maurizio Davini Department of Physics and INFN Pisa.
Computing/Tier 3 Status at Panjab S. Gautam, V. Bhatnagar India-CMS Meeting, Sept 27-28, 2007 Delhi University, Delhi Centre of Advanced Study in Physics,
Michigan Grid Testbed Report Shawn McKee University of Michigan UTA US ATLAS Testbed Meeting April 4, 2002.
27/04/05Sabah Salih Particle Physics Group The School of Physics and Astronomy The University of Manchester
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
 For Oracle Based Products  Windows XP Professional-SP3 / Window7-SP1 (32 Bit)  Min. Dual Processor or above  RAM 2GB or above  Internet explorer.
RAL PPD Site Update and other odds and ends Chris Brew.
Site Lightning Report: MWT2 Mark Neubauer University of Illinois at Urbana-Champaign US ATLAS Facilities UC Santa Cruz Nov 14, 2012.
Profiling Grid Data Transfer Protocols and Servers George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison USA.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
GridKa SC4 Tier2 Workshop – Sep , Warsaw Tier2 Site.
ATLAS Great Lakes Tier-2 Site Report Ben Meekhof USATLAS Tier2/Tier3 Workshop SLAC, Menlo Park, CA November 30 th, 2007.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
Location: BU Center for Computational Science facility, Physics Research Building, 3 Cummington Street.
PDSF at NERSC Site Report HEPiX April 2010 Jay Srinivasan (w/contributions from I. Sakrejda, C. Whitney, and B. Draney) (Presented by Sandy.
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH OS X Home server AFS using openafs 3 DB servers Kerberos 4 we will move.
Spending Plans and Schedule Jae Yu July 26, 2002.
Support in setting up a non-grid Atlas Tier 3 Doug Benjamin Duke University.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed.
Rob Allan Daresbury Laboratory NW-GRID Training Event 25 th January 2007 Introduction to NW-GRID R.J. Allan CCLRC Daresbury Laboratory.
KIT – The cooperation of Forschungszentrum Karlsruhe GmbH und Universität Karlsruhe (TH) Hadoop on HEPiX storage test bed at FZK Artem Trunov.
KOLKATA Grid Site Name :- IN-DAE-VECC-02Monalisa Name:- Kolkata-Cream VO :- ALICECity:- KOLKATACountry :- INDIA Shown many data transfers.
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
HEP Computing Status Sheffield University Matt Robinson Paul Hodgson Andrew Beresford.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Weekly Report By: Devin Trejo Week of June 21, 2015-> June 28, 2015.
Computational Research in the Battelle Center for Mathmatical medicine.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
Enabling Technologies for Distributed Computing Dr. Sanjay P. Ahuja, Ph.D. Fidelity National Financial Distinguished Professor of CIS School of Computing,
US ATLAS Western Tier 2 Status Report Wei Yang Nov. 30, 2007 US ATLAS Tier 2 and Tier 3 workshop at SLAC.
ATLAS Midwest Tier2 University of Chicago Indiana University Rob Gardner Computation and Enrico Fermi Institutes University of Chicago WLCG Collaboration.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Florida Tier2 Site Report USCMS Tier2 Workshop Livingston, LA March 3, 2009 Presented by Yu Fu for the University of Florida Tier2 Team (Paul Avery, Bourilkov.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
Final Implementation of a High Performance Computing Cluster at Florida Tech P. FORD, X. FAVE, K. GNANVO, R. HOCH, M. HOHLMANN, D. MITRA Physics and Space.
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
BaBar Cluster Had been unstable mainly because of failing disks Very few (
Ole’ Miss DOSAR Grid Michael D. Joy Institutional Analysis Center.
AGLT2 Site Report Shawn McKee University of Michigan March / OSG-AHM.
Southwest Tier 2 (UTA). Current Inventory Dedidcated Resources  UTA_SWT2 320 cores - 2GB/core Xeon EM64T (3.2GHz) Several Headnodes 20TB/16TB in IBRIX/DDN.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
Storage at the ATLAS Great Lakes Tier-2 Tier-2 Storage Administrator Talks Shawn McKee / University of Michigan OSG Storage ForumShawn McKee1.
ATLAS Tier-2 Storage Status AGLT2 OSG Storage Forum – U Chicago – Sep Shawn McKee / University of Michigan OSG Storage ForumShawn McKee1.
6th DOSAR Workshop University Mississippi Apr. 17 – 18, 2008
NL Service Challenge Plans
Cluster / Grid Status Update
LCG 3D Distributed Deployment of Databases
Lattice QCD Computing Project Review
ATLAS Great Lakes Tier-2 Site Report
Southwest Tier 2 Center Status Report
5th DOSAR Workshop Louisiana Tech University Sept. 27 – 28, 2007
AGLT2 Site Report Shawn McKee/University of Michigan
Presentation transcript:

ATLAS Great Lakes Tier-2 (AGL-Tier2) Shawn McKee (for the AGL Tier2) University of Michigan US ATLAS Tier-2 Meeting at Harvard Boston, MA, August 17 th, 2006

Shawn McKee - AGL-Tier2 (MSU/UM) 2 The AGL Tier2 The ATLAS Great Lakes Tier-2 (AGL-Tier2) is a joint effort of Michigan and Michigan State Universities.

Shawn McKee - AGL-Tier2 (MSU/UM) 3 AGL Personnel Involved We have a large group of faculty between MSU and UM who will be involved in related aspects of the AGL-Tier2We have a large group of faculty between MSU and UM who will be involved in related aspects of the AGL-Tier2 Table at the right shows the expected contribution by name and yearTable at the right shows the expected contribution by name and year

Shawn McKee - AGL-Tier2 (MSU/UM) 4 10GE Protected Network We will have a single “/23” network for the AGL-Tier2We will have a single “/23” network for the AGL-Tier2 Internally each site (UM/MSU) will have a /24 Our network will have 3 10GE wavelengths on MiLR in a “triangle”Our network will have 3 10GE wavelengths on MiLR in a “triangle” Loss of any of the 3 waves doesn’t impact connectivity for both sites

Shawn McKee - AGL-Tier2 (MSU/UM) 5 Existing MSU Cluster

Shawn McKee - AGL-Tier2 (MSU/UM) 6 Current UMROCKS Cluster We have a 5 rack AMD Athlon cluster with 70 operational nodes (2000/2400/2600 dual processor, 2 GB RAM)We have a 5 rack AMD Athlon cluster with 70 operational nodes (2000/2400/2600 dual processor, 2 GB RAM) Two 100+GB disksTwo 100+GB disks Plan to have ~100 nodes operationalPlan to have ~100 nodes operational ROCKS V4.1ROCKS V4.1

Shawn McKee - AGL-Tier2 (MSU/UM) 7 UMROCKS Installation ROCKS V4.1 with SLC V4.3 (32 bit) OSROCKS V4.1 with SLC V4.3 (32 bit) OS Significant additional software: TWiki, Cacti, Inventory, Syslog, etc.Significant additional software: TWiki, Cacti, Inventory, Syslog, etc. Cluster running dCache as wellCluster running dCache as well

Shawn McKee - AGL-Tier2 (MSU/UM) 8 Ganglia Info on Athlon Cluster Currently ~70 operational nodes – plan for ~100 from partsCurrently ~70 operational nodes – plan for ~100 from parts

Shawn McKee - AGL-Tier2 (MSU/UM) 9 AGL-Tier2 TWiki

Shawn McKee - AGL-Tier2 (MSU/UM) 10 Cacti Graphing/Monitoring

Shawn McKee - AGL-Tier2 (MSU/UM) 11 AGL-Tier2 Inventory Software

Shawn McKee - AGL-Tier2 (MSU/UM) 12 Existing Servers/Services In addition to the UMROCKS cluster we have a number of servers/services operationalIn addition to the UMROCKS cluster we have a number of servers/services operational Two gatekeepers: Dual Xeon 3.6 (2MB cache), 4GB RAM, Intel SE7520AF2 motherboards, IMM card (IPMI) called gate01/gate02.grid.umich.eduTwo gatekeepers: Dual Xeon 3.6 (2MB cache), 4GB RAM, Intel SE7520AF2 motherboards, IMM card (IPMI) called gate01/gate02.grid.umich.edu AFS Cell atlas.umich.edu hosted on linat02/linat03/linat04.grid.umich.edu with file servers attera/linat06/linat07/linat08/atums1/atums2 (about 6TB)AFS Cell atlas.umich.edu hosted on linat02/linat03/linat04.grid.umich.edu with file servers attera/linat06/linat07/linat08/atums1/atums2 (about 6TB) NFS data servers umfs01/umfs02/linat09/linat10/linat11 hosting about 22TB totalNFS data servers umfs01/umfs02/linat09/linat10/linat11 hosting about 22TB total Hypnos.grid.umich.edu is dCache headnode for UMROCKSHypnos.grid.umich.edu is dCache headnode for UMROCKS Have MonALISA node at ml-um.ultralight.org and other monitoring servicesHave MonALISA node at ml-um.ultralight.org and other monitoring services Oracle server on one of the “prototype” systems for Calibration/Alignment DB replicationOracle server on one of the “prototype” systems for Calibration/Alignment DB replication Planned servers: NDT node, GridFTP, DQ2Planned servers: NDT node, GridFTP, DQ2

Shawn McKee - AGL-Tier2 (MSU/UM) 13 OSG & ATLAS Software Status/Plans As shown we have both AFS and NFS storage at our Tier-2As shown we have both AFS and NFS storage at our Tier-2 We plan to install software on AFS (good for readonly type data). OSG (0.4.1) and ATLAS software already in AFS (/afs/atlas.umich.edu)We plan to install software on AFS (good for readonly type data). OSG (0.4.1) and ATLAS software already in AFS (/afs/atlas.umich.edu) ATLAS software is mirrored via Pacman on our AFS cell at: software is mirrored via Pacman on our AFS cell at: All users have their home space in AFS. Our system is setup to get Kereberos TGT (and AFS Tokens) at login via gssklog (instructions on TWiki)All users have their home space in AFS. Our system is setup to get Kereberos TGT (and AFS Tokens) at login via gssklog (instructions on TWiki) All OSG accounts created with “uniqname” IDsAll OSG accounts created with “uniqname” IDs

Shawn McKee - AGL-Tier2 (MSU/UM) 14 Prototype Opteron Cluster Testbed for dual dual-core systems (Opteron 280s, 4GB ram)Testbed for dual dual-core systems (Opteron 280s, 4GB ram)

Shawn McKee - AGL-Tier2 (MSU/UM) 15 Prototype Building Block Details We have purchased 5 dual dual-core Opteron 280 systems and an NFS storage server to test with.We have purchased 5 dual dual-core Opteron 280 systems and an NFS storage server to test with. Worker nodes are using Supermicro H8DAR-T(1U) motherboards (AMD 8132 chipset), 4GB of RAM, dual dual-core Opteron 280, three 250GB SATA-II hot-swappable drives, CDROM (4 cores/1U)Worker nodes are using Supermicro H8DAR-T(1U) motherboards (AMD 8132 chipset), 4GB of RAM, dual dual-core Opteron 280, three 250GB SATA-II hot-swappable drives, CDROM (4 cores/1U) Disk server is a dual dual-core Opteron 280, 5U, 24 SATA-II (500GB) drives, dual 250GB system disks, 8GB of RAM, dual 1GE NICs, Areca 1170 RAID6 controller (11TB)Disk server is a dual dual-core Opteron 280, 5U, 24 SATA-II (500GB) drives, dual 250GB system disks, 8GB of RAM, dual 1GE NICs, Areca 1170 RAID6 controller (11TB) Need to test I/O scaling using NFS with ~20 clients / storage serverNeed to test I/O scaling using NFS with ~20 clients / storage server Possible to use 10GE NIC on storage server if network impacts the performance.Possible to use 10GE NIC on storage server if network impacts the performance.

Shawn McKee - AGL-Tier2 (MSU/UM) 16 Michigan State Space Planning

Shawn McKee - AGL-Tier2 (MSU/UM) 17 Proposal Numbers for AGL-Tier2 The table above is from our proposal. The FY06 numbers assumed full funding in FY06.The table above is from our proposal. The FY06 numbers assumed full funding in FY06. We need to discuss the US ATLAS needs and timescales to deploy resources for the AGL-Tier2.We need to discuss the US ATLAS needs and timescales to deploy resources for the AGL-Tier2. Consideration is the availability of the long-term computer space at both MSU and UMConsideration is the availability of the long-term computer space at both MSU and UM We have 59 CPU-years of Opteron time from our Center for Advanced Computing which we will dedicate to the AGL-Tier2 (gate02/torque)We have 59 CPU-years of Opteron time from our Center for Advanced Computing which we will dedicate to the AGL-Tier2 (gate02/torque)

Shawn McKee - AGL-Tier2 (MSU/UM) 18 Planning for AGL-Tier2 Profile MSU and UM are both working on high-quality server spaces.MSU and UM are both working on high-quality server spaces. Michigan will share the Internet2/Merit space (MITC)Michigan will share the Internet2/Merit space (MITC) 4MW of power for 2MW of cooling and 2MW of equipment (Flywheels and generators) ~280 racks of space Lots of fiber access The MITC space is scheduled to be ready March 2007The MITC space is scheduled to be ready March 2007 We have “interim” space in our colleges server room IF we need to use it (up to 6 racks worth), but this would require two moves.We have “interim” space in our colleges server room IF we need to use it (up to 6 racks worth), but this would require two moves. MSU space will be ready in the same timescale (spring 2007)MSU space will be ready in the same timescale (spring 2007)

Shawn McKee - AGL-Tier2 (MSU/UM) 19 Summary The AGL-Tier2 is very close to being able to provide cycles for US ATLAS.The AGL-Tier2 is very close to being able to provide cycles for US ATLAS. The AGL-Tier2 should truly be a single “site” from US ATLAS’s point of view, even though our equipment and services are distributed between two campuses 60 miles apart. This is because of MiLR (10GE) and a common network address block.The AGL-Tier2 should truly be a single “site” from US ATLAS’s point of view, even though our equipment and services are distributed between two campuses 60 miles apart. This is because of MiLR (10GE) and a common network address block. We have some useful services running to help monitor and manage our equipment. If others are interested we are happy to share…We have some useful services running to help monitor and manage our equipment. If others are interested we are happy to share… The amount of cycles and storage required should be discussed as well as the time profile so we can optimize our plans.The amount of cycles and storage required should be discussed as well as the time profile so we can optimize our plans.