Office of Science U.S. Department of Energy NERSC Site Report HEPiX October 20, 2003 TRIUMF.

Slides:



Advertisements
Similar presentations
Tony Doyle - University of Glasgow GridPP EDG - UK Contributions Architecture Testbed-1 Network Monitoring Certificates & Security Storage Element R-GMA.
Advertisements

The RHIC-ATLAS Computing Facility at BNL HEPIX – Edinburgh May 24-28, 2004 Tony Chan RHIC Computing Facility Brookhaven National Laboratory.
IBM 1350 Cluster Expansion Doug Johnson Senior Systems Developer.
Information Technology Center Introduction to High Performance Computing at KFUPM.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
IBM RS6000/SP Overview Advanced IBM Unix computers series Multiple different configurations Available from entry level to high-end machines. POWER (1,2,3,4)
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Grid Services at NERSC Shreyas Cholia Open Software and Programming Group, NERSC NERSC User Group Meeting September 17, 2007.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
National Energy Research Scientific Computing Center (NERSC) The GUPFS Project at NERSC GUPFS Team NERSC Center Division, LBNL November 2003.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Computing/Tier 3 Status at Panjab S. Gautam, V. Bhatnagar India-CMS Meeting, Sept 27-28, 2007 Delhi University, Delhi Centre of Advanced Study in Physics,
Online Systems Status Review of requirements System configuration Current acquisitions Next steps... Upgrade Meeting 4-Sep-1997 Stu Fuess.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
Paul Scherrer Institut 5232 Villigen PSI HEPIX_AMST / / BJ95 PAUL SCHERRER INSTITUT THE PAUL SCHERRER INSTITUTE Swiss Light Source (SLS) Particle accelerator.
MURI Hardware Resources Ray Garcia Erik Olson Space Science and Engineering Center at the University of WI - Madison.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
23 Oct 2002HEPiX FNALJohn Gordon CLRC-RAL Site Report John Gordon CLRC eScience Centre.
Scientific Computing Experimental Physics Lattice QCD Sandy Philpott May 20, 2011 IT Internal Review 12GeV Readiness.
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
Laboratório de Instrumentação e Física Experimental de Partículas GRID Activities at LIP Jorge Gomes - (LIP Computer Centre)
PDSF at NERSC Site Report HEPiX April 2010 Jay Srinivasan (w/contributions from I. Sakrejda, C. Whitney, and B. Draney) (Presented by Sandy.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
ARGONNE NATIONAL LABORATORY Climate Modeling on the Jazz Linux Cluster at ANL John Taylor Mathematics and Computer Science & Environmental Research Divisions.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Probe Plans and Status SciDAC Kickoff July, 2001 Dan Million Randy Burris ORNL, Center for.
STAR Off-line Computing Capabilities at LBNL/NERSC Doug Olson, LBNL STAR Collaboration Meeting 2 August 1999, BNL.
KIT – The cooperation of Forschungszentrum Karlsruhe GmbH und Universität Karlsruhe (TH) Hadoop on HEPiX storage test bed at FZK Artem Trunov.
CASPUR Site Report Andrei Maslennikov Lead - Systems Amsterdam, May 2003.
CASPUR Site Report Andrei Maslennikov Lead - Systems Rome, April 2006.
HEP Computing Status Sheffield University Matt Robinson Paul Hodgson Andrew Beresford.
11 January 2005 High Performance Computing at NCAR Tom Bettge Deputy Director Scientific Computing Division National Center for Atmospheric Research Boulder,
TeraGrid Quarterly Meeting Arlington, VA Sep 6-7, 2007 NCSA RP Status Report.
Gareth Smith RAL PPD RAL PPD Site Report. Gareth Smith RAL PPD RAL Particle Physics Department Overview About 90 staff (plus ~25 visitors) Desktops mainly.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
1 Overall Architectural Design of the Earth System Grid.
ClinicalSoftwareSolutions Patient focused.Business minded. Slide 1 Opus Server Architecture Fritz Feltner Sept 7, 2007 Director, IT and Systems Integration.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
Office of Science U.S. Department of Energy Grid Security at NERSC/LBL Presented by Steve Chan Network, Security and Servers
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
ATLAS Midwest Tier2 University of Chicago Indiana University Rob Gardner Computation and Enrico Fermi Institutes University of Chicago WLCG Collaboration.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
SA1 operational policy training, Athens 20-21/01/05 Presentation of the HG Node “Isabella” and operational experience Antonis Zissimos Member of ICCS administration.
PDSF and the Alvarez Clusters Presented by Shane Canon, NERSC/PDSF
National Energy Research Scientific Computing Center (NERSC) NERSC Site Report Shane Canon NERSC Center Division, LBNL 10/15/2004.
LBNL/NERSC Site Report Cary Whitney NERSC
Ole’ Miss DOSAR Grid Michael D. Joy Institutional Analysis Center.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
LBNL/NERSC/PDSF Site Report for HEPiX Catania, Italy April 17, 2002 by Cary Whitney
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
Data Infrastructure in the TeraGrid Chris Jordan Campus Champions Presentation May 6, 2009.
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Jefferson Lab Site Report Sandy Philpott HEPiX Fall 07 Genome Sequencing Center Washington University at St. Louis.
Report from US ALICE Yves Schutz WLCG 24/01/2007.
INFN Site Report R.Gomezel October 9-13, 2006 Jefferson Lab, Newport News.
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
Cluster / Grid Status Update
NERSC Reliability Data
National Energy Research Scientific Computing Center (NERSC)
The National Grid Service Mike Mineter NeSC-TOE
Presentation transcript:

Office of Science U.S. Department of Energy NERSC Site Report HEPiX October 20, 2003 TRIUMF

Office of Science U.S. Department of Energy LBL, NERSC, and PDSF LBL manages the NERSC Center for DOE PDSF is the production Linux cluster at NERSC used primarily for HEP science Site report will touch on activities of interest to HEPiX community at each of these levels

Office of Science U.S. Department of Energy PDSF - New Hardware 96 Dual Athlon Systems 8 Storage Nodes - ~18 TB formatted All gigabit attached (Dell switches) Purchased two Opteron systems for testing

Office of Science U.S. Department of Energy PDSF Projects HostDB - Presentation later Sun GridEngine Evaluation – Met all requirements (long list) – Putting in semi-production on retired nodes Grid certificate DN kernel module 1-wire based monitoring and control network High Availability Server – Uses heartbeat code – IDE based Fibre-Channel array

Office of Science U.S. Department of Energy PDSF - Other news Aztera – Zambeel folded – StorAd is making best effort to support the system New User Groups – KamLAND – e896 – ALICE

Office of Science U.S. Department of Energy IBM SP Upgraded – 208 nodes added - 16 way Nighthawk II – Additional 20 TB of disk Total System – 10 Tflops/s peak – 7.8 TB memory – 44 TB of GPFS storage

Office of Science U.S. Department of Energy Mass Storage Hardware – New DataDirect disk cache – New tape drives allow high capacity cartridges (200 GB) Software – Currently running HPSS 4.3 – Testing 5.1 Testing – DMAPI – htar command

Office of Science U.S. Department of Energy Grid Activities GridFTP and gatekeeper deployed on all productions system (except gatekeeper on Seaborg which is coming soon) Integrating account management system with grid certificates Testing myproxy based system Portal Web interface to HPSS

Office of Science U.S. Department of Energy Networking Jumbo support to ESNET – Looking for other sites to test Jumbo across WAN New production router (Juniper)

Office of Science U.S. Department of Energy GUPFS Hardware testbed: – 3Par Data – Yotta Yotta – Dell EMC – Dot Hill – Data Direct (Soon) – Panasas Interconnect hardware: – Topspin (IB) – Infinicon (IB) – Cisco (ISCSI) – Qlogic (ISCSI) – Adaptec (ISCSI) – Myrinet 2000 – Various FC Filesystems: – ADIC license – GPFS license – GFS 5.2 license – Lustre Test clients: Dual processor 2.2GHz Xeons 2GB memory 2 PCI-X Local HD for OS

Office of Science U.S. Department of Energy Distributed System Dept. Net100 ( - Built on Web100 (PSC, NCAR, NCSA) and NetLogger (LBNL), Net100 modifies operating systems to respond dynamically to network conditions and make adjustments in network transfers, sending data as fast as the network will allow. Self Configuring Network Monitor (SCNM) - ( provide accurate, comprehensive, and on-demand, application-to-application monitoring capabilities throughout the interior of the interconnecting network domains.

Office of Science U.S. Department of Energy Distributed Systems (cont’d) Netlogger ( pyGlobus ( Python interface to the Globus Toolkit. LIGO gravity wave experiment is using it to replicate TB/day data around the US with the LIGO Data Replicator ( group.phys.uwm.edu/LDR/) DOEGrids.org PKI for the DOE science community, part of federation supporting international scientific collaborations

Office of Science U.S. Department of Energy Repaired Hardware System from 2000 wide spread failure (half of 90 systems) Had broken systems inspected by LBL Electronics Shop Discovered 4 bad capictors (~$2) Prep’d systems can be repaired for ~$20/board 16 systems repaired so far Plan to eventually repair all system from batch