National Energy Research Scientific Computing Center (NERSC) NERSC Site Report Shane Canon NERSC Center Division, LBNL 10/15/2004.

Slides:



Advertisements
Similar presentations
Die Kooperation von Forschungszentrum Karlsruhe GmbH und Universität Karlsruhe (TH) StorNextFS, a fast global Filesysteme in a heterogeneous Cluster Environment.
Advertisements

WestGrid Collaboration and Visualization Brian Corrie Collaboration and Visualization Coordinator WestGrid/SFU.
Copyright © 2014 EMC Corporation. All Rights Reserved. Linux Host Installation and Integration for Block Upon completion of this module, you should be.
Engenio 7900 HPC Storage System. 2 LSI Confidential LSI In HPC LSI (Engenio Storage Group) has a rich, successful history of deploying storage solutions.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Trends in Storage Subsystem Technologies Michael Joyce, Senior Director Mylex & OEM Storage Subsystems IBM.
National Energy Research Scientific Computing Center (NERSC) The GUPFS Project at NERSC GUPFS Team NERSC Center Division, LBNL November 2003.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
TeraGrid Information Services December 1, 2006 JP Navarro GIG Software Integration.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
Module 9: Configuring Storage
CASPUR SAN News Andrei Maslennikov Orsay, April 2001.
InfiniSwitch Company Confidential. 2 InfiniSwitch Agenda InfiniBand Overview Company Overview Product Strategy Q&A.
MURI Hardware Resources Ray Garcia Erik Olson Space Science and Engineering Center at the University of WI - Madison.
MDC417 Follow me on Working as Practice Manager for Insight, he is a subject matter expert in cloud, virtualization and management.
Scientific Computing Experimental Physics Lattice QCD Sandy Philpott May 20, 2011 IT Internal Review 12GeV Readiness.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
Tier 1 Facility Status and Current Activities Rich Baker Brookhaven National Laboratory NSF/DOE Review of ATLAS Computing June 20, 2002.
Copyright © 2014 EMC Corporation. All Rights Reserved. Windows Host Installation and Integration for Block Upon completion of this module, you should be.
SoCal Infrastructure OptIPuter Southern California Network Infrastructure Philip Papadopoulos OptIPuter Co-PI University of California, San Diego Program.
Laboratório de Instrumentação e Física Experimental de Partículas GRID Activities at LIP Jorge Gomes - (LIP Computer Centre)
PDSF at NERSC Site Report HEPiX April 2010 Jay Srinivasan (w/contributions from I. Sakrejda, C. Whitney, and B. Draney) (Presented by Sandy.
JLab Scientific Computing: Theory HPC & Experimental Physics Thomas Jefferson National Accelerator Facility Newport News, VA Sandy Philpott.
EGEE is a project funded by the European Union under contract IST HellasGrid Hardware Tender Christos Aposkitis GRNET EGEE 3 rd parties Advanced.
INFSO-RI Enabling Grids for E-sciencE Hellas Grid infrastructure update Kostas Koumantaros, Christos Aposkitis EGEE-HellasGrid Coordination.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Probe Plans and Status SciDAC Kickoff July, 2001 Dan Million Randy Burris ORNL, Center for.
Storage and Storage Access 1 Rainer Többicke CERN/IT.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
KIT – The cooperation of Forschungszentrum Karlsruhe GmbH und Universität Karlsruhe (TH) Hadoop on HEPiX storage test bed at FZK Artem Trunov.
HEPiX FNAL ‘02 25 th Oct 2002 Alan Silverman HEPiX Large Cluster SIG Report Alan Silverman 25 th October 2002 HEPiX 2002, FNAL.
1 The NERSC Global File System NERSC June 12th, 2006.
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
Ultimate Integration Joseph Lappa Pittsburgh Supercomputing Center ESCC/Internet2 Joint Techs Workshop.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
CASPUR Site Report Andrei Maslennikov Lead - Systems Rome, April 2006.
Queensland University of Technology CRICOS No J VMware as implemented by the ITS department, QUT Scott Brewster 7 December 2006.
National Energy Research Scientific Computing Center (NERSC) NERSC View of the Greenbook Bill Saphir Chief Architect NERSC Center Division, LBNL 6/23/2004.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
1 Accomplishments. 2 Overview of Accomplishments  Sustaining the Production Earth System Grid Serving the current needs of the climate modeling community.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
ISCSI. iSCSI Terms An iSCSI initiator is something that requests disk blocks, aka a client An iSCSI target is something that provides disk blocks, aka.
ClinicalSoftwareSolutions Patient focused.Business minded. Slide 1 Opus Server Architecture Fritz Feltner Sept 7, 2007 Director, IT and Systems Integration.
National Energy Research Scientific Computing Center (NERSC) CHOS - CHROOT OS Shane Canon NERSC Center Division, LBNL SC 2004 November 2004.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
Barriers to IB adoption (Storage Perspective) Ashish Batwara Software Solution Architect May 01, 2007.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Florida Tier2 Site Report USCMS Tier2 Workshop Livingston, LA March 3, 2009 Presented by Yu Fu for the University of Florida Tier2 Team (Paul Avery, Bourilkov.
Office of Science U.S. Department of Energy NERSC Site Report HEPiX October 20, 2003 TRIUMF.
KIT – University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association STEINBUCH CENTRE FOR COMPUTING - SCC
SA1 operational policy training, Athens 20-21/01/05 Presentation of the HG Node “Isabella” and operational experience Antonis Zissimos Member of ICCS administration.
PDSF and the Alvarez Clusters Presented by Shane Canon, NERSC/PDSF
Portuguese Grid Infrastruture(s) Gonçalo Borges Jornadas LIP 2010 Braga, Janeiro 2010.
LBNL/NERSC Site Report Cary Whitney NERSC
Southern California Infrastructure Philip Papadopoulos Greg Hidley.
LBNL/NERSC/PDSF Site Report for HEPiX Catania, Italy April 17, 2002 by Cary Whitney
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
Farming Andrea Chierici CNAF Review Current situation.
Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb
Jefferson Lab Site Report Sandy Philpott HEPiX Fall 07 Genome Sequencing Center Washington University at St. Louis.
ORNL is managed by UT-Battelle for the US Department of Energy OLCF HPSS Performance Then and Now Jason Hill HPC Operations Storage Team Lead
The NERSC Global File System and PDSF Tom Langley PDSF Support Group NERSC at Lawrence Berkeley National Laboratory Fall HEPiX October 2006.
INFN Site Report R.Gomezel October 9-13, 2006 Jefferson Lab, Newport News.
The Beijing Tier 2: status and plans
Статус ГРИД-кластера ИЯФ СО РАН.
Vladimir Sapunenko On behalf of INFN-T1 staff HEPiX Spring 2017
GGF15 – Grids and Network Virtualization
NERSC Reliability Data
National Energy Research Scientific Computing Center (NERSC)
A New Storage Test Bed and the Research Projects on it
Presentation transcript:

National Energy Research Scientific Computing Center (NERSC) NERSC Site Report Shane Canon NERSC Center Division, LBNL 10/15/2004

NERSC Outline PDSF Other Computational Systems Networking Storage GUPFS Security

PDSF – New Hardware 49 Dual Xeon Systems 10 Dual Opteron Systems All nodes are using native SATA controller (SI 3112 and SI 3114) All nodes are gigE Upgraded hard drives on 14 nodes (Added ~14 TB formatted Foundry FES48 – 2 10G, 48 1G ports

PDSF – Other Changes New hardware will run SL (3.03) CHOS already installed and will help ease transition to SL for users New nodes will run under Sun GridEngine – PDSF did not renew LSF maintenance – LSF nodes will slowly be transitioned over to SGE

PDSF Projects Exploratory work has been hampered by involvement with NCS procurement, GUPFS project (and bike accidents) Recent focus has been –CHOS –Deployment of new hardware –SL –Lustre

PDSF - Lustre Still not tested with users Newer versions seem much more robust Good at spot lighting flakey hardware Older hardware is being reconfigured for use as a Lustre pool. Roughly 10 TB of total space.

NERSC - IBM SP Upgraded to 5.2 –Serious problems at first –IBM dispatched team to diagnose and fix problems Added FibreChannel disk –~13 TB –FAStT 700 based

NERSC Systems - NCS Award has been made No formal announcement until acceptance is completed

NERSC Systems - NVS New Visualization System Small Altix System (4 nodes) Some early issues –Channel bonded Ethernet Jumbo not supported Using a Apple Xserve raid on it until O3k is decommissioned

Networking – 10G NERSC is building up a 10G infrastructure Two MG8s provide core switching and routing for 10G network Jumbo frames Initially focused on core, mass storage, and visualization system. Exploring ways to extend to Seaborg. PDSF provided its own 10G Layer 3 switch.

NERSC - WAN 10 G upgrade to WAN is in the works Waiting on Bay Area Metropolitan Area Network deployment by ES Net. Procurement is already under way

Mass Storage Latest Hardware –New Movers will have 10G links (testing is starting) –LSI based storage Other projects –DMAPI work –Portals and other web interfaces into HPSS

Security - OTP Project on hold while funding is explored To date various tokens have been evaluated Focus is on products that are extensible and can be integrated fully in to NERSC and DOE infrastructures Testing of cross RADIUS delegation Should integrate into Grid using MyProxy or KCA approach

Bro Lite DOE Funded Simplify Bro –Configuration (GUI) –Output filters Available: Soon Beta slots available Contact:

GUPFS Planned deployment late 2005 Unified filesystem spanning all NERSC systems (NCS, Seaborg, PDSF) Possible candidates –GPFS, ADIC, Lustre, Panasas, Storage Tank Results: Contact:

GUPFS Tested File Systems – Sistina GFS 4.2, 5.0, 5.1, and 5.2 Beta – ADIC StorNext File System 2.0 and 2.2 – Lustre 0.6 (1.0 Beta 1), 0.9.2, 1.0, 1.0.{1,2,3,4}, – IBM GPFS for Linux, 1.3 and 2.2. Beta 2.3. – SANFS starting soon – Panasas Fabric – FC (1Gb/s and 2Gb/s): Brocade SilkWorm, Qlogic SANbox2, Cisco MDS 9509, SANDial Shadow – Ethernet (iSCSI): Cisco SN 5428, Intel & Adaptec iSCSI HBA, Adaptec TOE, Cisco MDS 9509 – Infiniband (1x and 4x): InfiniCon and Topspin IB to GE/FC bridges (SRP over IB, iSCSI over IB), – Inter-connect: Myrinnet 2000 (Rev D) Storage – Traditional Storage: Dot Hill, Silicon Gear, Chaparral – New Storage: Yotta Yotta GSX 2400, EMC CX 600, 3PAR, DDN S2A 8500

Procurements Several Procurements are starting up GUPFS –Global Filesystem for NERSC –Deployment targeted for Spring 2005 NERSC5 – –Follow on to Seaborg –Likely target is 2005/2006 NCSe –Second year of funding for new capability at NERSC (NCS was first block) –Target Workload still being determined

PDSF - Utilization STAR has steadily picked up production over past months primary reason Continued to encourage use of SGE pool for smaller groups and Grid projects