Canada’s National Laboratory for Particle and Nuclear Physics Laboratoire national canadien pour la recherche en physique nucléaire et en physique des.

Slides:



Advertisements
Similar presentations
Owned and operated as a joint venture by a consortium of Canadian universities via a contribution through the National Research Council Canada Propriété
Advertisements

1 Isabel Trigger ACOT March 13 th, 2009 ATLAS at the LHC …and at TRIUMF CANADA’S NATIONAL LABORATORY FOR PARTICLE AND NUCLEAR PHYSICS Owned and operated.
Owned and operated as a joint venture by a consortium of Canadian universities via a contribution through the National Research Council Canada Propriété.
Proposal for ½ MW photo-fission driver based on TESLA 1.3 GHz SCRF technology (Shane Koscielniak, 09 Nov 2007) Electron Linac CANADA ’ S NATIONAL LABORATORY.
S.Chechelnitskiy / SFU Simon Fraser Running CE and SE in a XEN virtualized environment S.Chechelnitskiy Simon Fraser University CHEP 2007 September 6 th.
Owned and operated as a joint venture by a consortium of Canadian universities via a contribution through the National Research Council Canada Propriété.
The CDCE BNL HEPIX – LBL October 28, 2009 Tony Chan - BNL.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
Proposal for ½ MW photo-fission driver based on TESLA 1.3 GHz SCRF technology (Shane Koscielniak, 09 Nov 2007) Electron Linac CANADA ’ S NATIONAL LABORATORY.
Owned and operated as a joint venture by a consortium of Canadian universities via a contribution through the National Research Council Canada Propriété.
Owned and operated as a joint venture by a consortium of Canadian universities via a contribution through the National Research Council Canada Propriété.
Storage Virtualisation – Why? A Case Study Steve O’Donnell Global Head of Data Centres and Bridge Operations BT plc.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
CERN IT Department CH-1211 Genève 23 Switzerland t Next generation of virtual infrastructure with Hyper-V Michal Kwiatek, Juraj Sucik, Rafal.
James Inkster, Msc., SFU/TRIUMF Simplifying the Preparation of 18 F-based Biomolecular Imaging Agents Through ‘Click’ Chemistry CANADA’S NATIONAL LABORATORY.
SRF Activities at TRIUMF & PAVAC
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
Steven McDonald May TRIUMF Site Report CANADA’S NATIONAL LABORATORY FOR PARTICLE AND NUCLEAR PHYSICS Owned and operated as a joint.
CC - IN2P3 Site Report Hepix Fall meeting 2009 – Berkeley
October, Scientific Linux INFN/Trieste B.Gobbo – Compass R.Gomezel - T.Macorini - L.Strizzolo INFN - Trieste.
Presented by: Sanketh Beerabbi University of Central Florida COP Cloud Computing.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
CANADA’S NATIONAL LABORATORY FOR PARTICLE AND NUCLEAR PHYSICS Owned and operated as a joint venture by a consortium of Canadian universities via a contribution.
Databases at TRIUMF Andrew Wong CANADA’S NATIONAL LABORATORY FOR PARTICLE AND NUCLEAR PHYSICS Owned and operated as a joint venture by a consortium of.
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
Configuration Management with Cobbler and Puppet Kashif Mohammad University of Oxford.
Owned and operated as a joint venture by a consortium of Canadian universities via a contribution through the National Research Council Canada Propriété.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
JLab Scientific Computing: Theory HPC & Experimental Physics Thomas Jefferson National Accelerator Facility Newport News, VA Sandy Philpott.
IST Storage & Backup Group 2011 Jack Shnell Supervisor Joe Silva Senior Storage Administrator Dennis Leong.
Owned and operated as a joint venture by a consortium of Canadian universities via a contribution through the National Research Council Canada Propriété.
Owned and operated as a joint venture by a consortium of Canadian universities via a contribution through the National Research Council Canada Propriété.
January 18, 2013 All Hands January ARIEL Completion and the next 5 Year Plan.
Owned and operated as a joint venture by a consortium of Canadian universities via a contribution through the National Research Council Canada Propriété.
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
High Performance Computing (HPC) Data Center Proposal Imran Latif, Facility Project Manager Scientific & Enterprise Computing Data Centers at BNL 10/14/2015.
2012 Objectives for CernVM. PH/SFT Technical Group Meeting CernVM/Subprojects The R&D phase of the project has finished and we continue to work as part.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
To provide the world with a next generation storage platform for unstructured data, enabling deployment of mobile applications, virtualization solutions,
R. Krempaska, October, 2013 Wir schaffen Wissen – heute für morgen Controls Security at PSI Current Status R. Krempaska, A. Bertrand, C. Higgs, R. Kapeller,
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
IT-INFN-CNAF Status Update LHC-OPN Meeting INFN CNAF, December 2009 Stefano Zani 10/11/2009Stefano Zani INFN CNAF (TIER1 Staff)1.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
Introduction to Exadata X5 and X6 New Features
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
TRIUMF Site Report for HEPiX, JLAB, October 9-13, 2006 – Corrie Kost TRIUMF SITE REPORT Corrie Kost & Steve McDonald Update since Hepix Spring 2006.
Owned and operated as a joint venture by a consortium of Canadian universities via a contribution through the National Research Council Canada Propriété.
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
November 28, 2007 Dominique Boutigny – CC-IN2P3 CC-IN2P3 Update Status.
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Owned and operated as a joint venture by a consortium of Canadian universities via a contribution through the National Research Council Canada Propriété.
Scientific Linux Connie Sieh CSAM Meeting May 2, 2006.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Owned and operated as a joint venture by a consortium of Canadian universities via a contribution through the National Research Council Canada Propriété.
Owned and operated as a joint venture by a consortium of Canadian universities via a contribution through the National Research Council Canada Propriété.
Supercool Neutrons (Ultracold Neutrons) Jeff Martin University of Winnipeg Skywalk 2007 research supported by Natural Sciences and Engineering Research.
Administering the SOWN Network David R Newman & Chris Malton.
Extreme Scale Infrastructure
Lenovo – DataCore Deployment Ready Offerings
Subatomic Physics Detector Lab
CERN Data Centre ‘Building 513 on the Meyrin Site’
tdct – a Capfast replacement
Update on Plan for KISTI-GSDC
Christof Hanke, HEPIX Spring Meeting 2008, CERN
OPS235: Lab 2 Virtual Machines – Part I
R.Keitel Epics Collaboration Meeting, Padova, Oct. 2008
Virtualization Dr. S. R. Ahmed.
Productive + Hybrid + Intelligent + Trusted
Presentation transcript:

Canada’s National Laboratory for Particle and Nuclear Physics Laboratoire national canadien pour la recherche en physique nucléaire et en physique des particules Owned and operated as a joint venture by a consortium of Canadian universities via a contribution through the National Research Council Canada Propriété d’un consortium d’universités canadiennes, géré en co-entreprise à partir d’une contribution administrée par le Conseil national de recherches Canada TRIUMF Site Report HEPix Fall 2009, NERSC/LBNL Kelvin Raywood TRIUMF, Vancouver, Canada

York, Guelph and Queens U's joined as full members –11 full-members + 4 associate-members 40 th anniversary SRF e-linac project funded –2 nd driver in 40 years –Photo-fission for nuclear physics and medical-isotope production 99 Mo –Contribute to LHC-SPL, ILC & CLS upgrades External HPC facilities: WestGrid –Bugaboo: SL5, 1280 cores (E5450), DDR/IB, 2GB/core, 350TB Lustre –Orcinus: CentOS-5, 3072 cores(E5440) DDR/IB, 2GB/core –Checkers: SL5, 1280 cores(E5440), DDR/IB, 2GB/core –Storage: 1.5PB GPFS + 800TB (LTO 2,3,4) : Tivoli HSM Lab update HEPiX Fall 2009TRIUMF Site Report2

Charged to review operations of leader and three of four TRIUMF computing-groups –MIS (3), GSC (5), CCN (6) : +1 since 1990 –WLCG/ATLAS Tier-1 Centre (9) – not reviewed Triggered by increased demand for app. devel by MIS –modern admin systems, identity management, online work-flow,... and by requirements of experiments supported by GSC –DAQ and analysis software, FPGA programming Senior management were largely unaware of issues with CCN –documentation, response time, high availability, disaster recovery, modernisation,... Plans for resolution –app-devel stack, change control, virtualisation/HA, web-site,... –People requirements: MIS +2, GSC +2, CCN +2 External Review of TRIUMF Computing HEPiX Fall 2009TRIUMF Site Report3

Recommendations of Review HEPiX Fall 2009TRIUMF Site Report4 Strong endorsement of computing leader –Increase power for resource allocation, approve user-requests,... Core Computing and Networking –Off-load commodity computing to University partners, commercial vendors –+1 “short-term staff”, and +1 FTE General Scientific Computing –Reduce scope of activities –Require experiments to fund DAQ development and support Management Information Systems –Hire external consultant to validate / augment the long-term plan –Explore alternatives to in-house developed apps

RPMS for modifying configs HEPiX Fall 2009TRIUMF Site Report5 Special extensions:.ADD.APPEND.ED.REPLACE.SED.SYMLINK # // # rpm: triumf-sshd-protect_root # Allow only ssh-key auth for root PermitRootLogin without-password # // # // # rpm: triumf-sshd-protect_root # Allow only ssh-key auth for root PermitRootLogin without-password # // # $Id: Makefile :26:54Z YUM_REPO = triumf-server include../Makefile.rpmbuild # $Id: Makefile :26:54Z YUM_REPO = triumf-server include../Makefile.rpmbuild # triumf-sshd-protect_root # $Id: Specfile :47:17Z # Version: 1.1 Release: 1 PreReq: openssh-server %description Disables ssh to root via password %post service sshd condrestart > /dev/null || : %postun [ "$1" = 0 ] && \ service sshd condrestart >/dev/null|| : # triumf-sshd-protect_root # $Id: Specfile :47:17Z # Version: 1.1 Release: 1 PreReq: openssh-server %description Disables ssh to root via password %post service sshd condrestart > /dev/null || : %postun [ "$1" = 0 ] && \ service sshd condrestart >/dev/null|| : Special vars: YUM_REPO, PKG_ARCH, DIST_RELEASE,... make install Builds rpm Installs in repo

Linux configuration management HEPiX Fall 2009TRIUMF Site Report6 Deploy single-purpose servers, preferably virtual Start with minimal installation and record extra packages –Small number of kickstarts and VM gold-masters Use well behaved rpms for common configurations –Nagios client, RAID monitoring, ssh keys, yum, syslog, ntp,... –Config changes undone on removal Keep everything in a version-control system (svn)‏ Separate software, configs, data Push custom configs to server – Makefile checks svn status – Reloads service # $Id: Makefile :33:11Z # Install nagios configuration files SERVER = trnetmon.triumf.ca CONFIGS = $(wildcard *.cfg)‏ INSTALL_ROOT = /etc/nagios/conf.d RELOAD_CMD = /sbin/service nagios reload CONFIG_OWNER = 82 CONFIG_GROUP = 82 include../../Makefile.server # $Id: Makefile :33:11Z # Install nagios configuration files SERVER = trnetmon.triumf.ca CONFIGS = $(wildcard *.cfg)‏ INSTALL_ROOT = /etc/nagios/conf.d RELOAD_CMD = /sbin/service nagios reload CONFIG_OWNER = 82 CONFIG_GROUP = 82 include../../Makefile.server

SUN Unified Storage Server HEPiX Fall 2009TRIUMF Site Report7 Acquired in spring –Based on ZFS, RAID-Z DP –Provides iSCSI, NFS, in kernel CIFS –Hybrid storage, DRAM, SSD, SATA –File & block level snapshots –Scales to 500TB Locked up on system-disk failure –Known bug –fixed in system update that we had not yet applied Q3 update contains new iSCSI stack –Required rebuild of iSCSI config –client config needed modification

ATLAS Tier-1 Upgrades HEPiX Fall 2009TRIUMF Site Report 8 * 8% 1TB drives, 92% 2TB drives (Nov.)‏ New capacity is installed and being commissioned Will provide ~7% of world-wide ATLAS resources Upgrade Oracle RAC, ~30TB, +2 nodes

ATLAS Tier-1 Infrastructure HEPiX Fall 2009TRIUMF Site Report9 Limited floor space: 43' x 22' No false floor Rack optimization: - high density solution - hot & cold configuration Power estimate: ~0.4 MW (up to 2012)‏ (including cooling)‏ Cooling solution: Liebert XD system (very flexible) 340 kW total capacity (~1/4 used)‏ UPS: 225 kVA (in the future CPU racks on regular power or expand UPS capacity) (~1/3 used). (no diesel backup except for core network)‏ Expansion beyond 2012: new infrastructure will be required (TRIUMF next 5YP)‏ (higher density / LHC delay)‏

ATLAS Tier-1 HSM HEPiX Fall 2009TRIUMF Site Report10 High performance HSM CHEP09 (Denice Deatrich, Simon Liu, Reda Tafirout)‏ –In production –interfaced with dCache storage developed at TRIUMF –No proprietry code for tape-drivers or changer –TCP socket based server daemon –backend MySQL Design goal: Efficient reprocessing –Reorder and prioritise requests –Control over file and tape grouping Minimses tape mounts Maximises reads per mount No prestaging ~0.2TB / h Prestaging ~1TB / h

LHCOPN Meeting at TRIUMF HEPiX Fall 2009TRIUMF Site Report11 1 st LHCOPN meeting at T1 outside Europe Dante perSONAR multi-domain monitoring installed at all sites Monitoring T0-T1 traffic Considered extending LHCOPN to T1-T1 and T1-T2 Traffic patterns unknown Need traffic-pattern specs from WLCG before designing a network topology

Network Status – the last km HEPiX Fall 2009TRIUMF Site Report12 All TRIUMF external network connections and ATLAS lightpaths pass through this shack. February 2009 October 2009

Lac TRIUMF Lake HEPiX Fall 2009TRIUMF Site Report13

Thank You! Merci! 4004 Wesbrook Mall | Vancouver BC | Canada V6T 2A3 | Tel | Fax |