1 Computing & Networking User Group Meeting Roy Whitney Andy Kowalski Sandy Philpott Chip Watson 17 June 2008.

Slides:



Advertisements
Similar presentations
XenData SX-520 LTO Archive Servers A series of archive servers based on IT standards, designed for the demanding requirements of the media and entertainment.
Advertisements

XenData SXL-3000 LTO Archive System Turnkey video archive system with near-line LTO capacities scaling from 150 TB to 750 TB, designed for the demanding.
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
IHEP Site Status Jingyan Shi, Computing Center, IHEP 2015 Spring HEPiX Workshop.
Desktop Computing Strategic Project Sandia National Labs May 2, 2009 Jeremy Allison Andy Ambabo James Mcdonald Sandia is a multiprogram laboratory operated.
12. March 2003Bernd Panzer-Steindel, CERN/IT1 LCG Fabric status
Technology Steering Group January 31, 2007 Academic Affairs Technology Steering Group February 13, 2008.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Technology Steering Group January 31, 2007 Academic Affairs Technology Steering Group February 13, 2008.
12 GeV Era Computing (CNI) Andy Kowalski May 20, 2011.
Edinburgh Site Report 1 July 2004 Steve Thorn Particle Physics Experiments Group.
Windows Server MIS 424 Professor Sandvig. Overview Role of servers Performance Requirements Server Hardware Software Windows Server IIS.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
UCL Site Report Ben Waugh HepSysMan, 22 May 2007.
The D0 Monte Carlo Challenge Gregory E. Graham University of Maryland (for the D0 Collaboration) February 8, 2000 CHEP 2000.
Status of WLCG Tier-0 Maite Barroso, CERN-IT With input from T0 service managers Grid Deployment Board 9 April Apr-2014 Maite Barroso Lopez (at)
IT in the 12 GeV Era Roy Whitney, CIO May 31, 2013 Jefferson Lab User Group Annual Meeting.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
Outline IT Organization SciComp Update CNI Update
Design & Management of the JLAB Farms Ian Bird, Jefferson Lab May 24, 2001 FNAL LCCWS.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Paul Scherrer Institut 5232 Villigen PSI HEPIX_AMST / / BJ95 PAUL SCHERRER INSTITUT THE PAUL SCHERRER INSTITUTE Swiss Light Source (SLS) Particle accelerator.
Nov 1, 2000Site report DESY1 DESY Site Report Wolfgang Friebel DESY Nov 1, 2000 HEPiX Fall
Computing and IT Update Jefferson Lab User Group Roy Whitney, CIO & CTO 10 June 2009.
NLIT May 26, 2010 Page 1 Computing Jefferson Lab Users Group Meeting 8 June 2010 Roy Whitney CIO & CTO.
Scientific Computing Experimental Physics Lattice QCD Sandy Philpott May 20, 2011 IT Internal Review 12GeV Readiness.
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility HEPiX – Fall, 2005.
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Developing & Managing A Large Linux Farm – The Brookhaven Experience CHEP2004 – Interlaken September 27, 2004 Tomasz Wlodek - BNL.
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
JLab Scientific Computing: Theory HPC & Experimental Physics Thomas Jefferson National Accelerator Facility Newport News, VA Sandy Philpott.
Integrating JASMine and Auger Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
2  Supervisor : MENG Sreymom  SNA 2012_Group4  Group Member  CHAN SaratYUN Sinot  PRING SithaPOV Sopheap  CHUT MattaTHAN Vibol  LON SichoeumBEN.
1 Week #10Business Continuity Backing Up Data Configuring Shadow Copies Providing Server and Service Availability.
IST Storage & Backup Group 2011 Jack Shnell Supervisor Joe Silva Senior Storage Administrator Dennis Leong.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National Laboratory.
Operating Systems & Information Services CERN IT Department CH-1211 Geneva 23 Switzerland t OIS Update on Windows 7 at CERN & Remote Desktop.
11 January 2005 High Performance Computing at NCAR Tom Bettge Deputy Director Scientific Computing Division National Center for Atmospheric Research Boulder,
Outline: Status: Report after one month of Plans for the future (Preparing Summer -Fall 2003) (CNAF): Update A. Sidoti, INFN Pisa and.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Status of GSDC, KISTI Sang-Un Ahn, for the GSDC Tier-1 Team
Jefferson Lab Site Report Sandy Philpott HEPiX Fall 07 Genome Sequencing Center Washington University at St. Louis.
Scientific Linux Inventory Project (SLIP) Troy Dawson Connie Sieh.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
1 P. Murat, Mini-review of the CDF Computing Plan 2006, 2005/10/18 An Update to the CDF Offline Plan and FY2006 Budget ● Outline: – CDF computing model.
Working with Windows 7 at CERN
Compute and Storage For the Farm at Jlab
LQCD Computing Project Overview
Chapter 7: Using Windows Servers
Update on Plan for KISTI-GSDC
Windows 7 deployment at CERN
Bernd Panzer-Steindel, CERN/IT
LQCD Computing Operations
Scientific Computing At Jefferson Lab
Outline IT Division News Other News
PLANNING A SECURE BASELINE INSTALLATION
Lee Lueking D0RACE January 17, 2002
Jefferson Lab Scientific Computing Update
Presentation transcript:

1 Computing & Networking User Group Meeting Roy Whitney Andy Kowalski Sandy Philpott Chip Watson 17 June 2008

2 Users and JLab IT Ed Brash is User Group Board of Directors’ representative on the IT Steering Committee. Physics Computing Committee (Sandy Philpott) Helpdesk and CCPR requests and activities Challenges –Constrained budget Staffing Aging infrastructure –Cyber Security

3 Computing and Networking Infrastructure Andy Kowalski

4 CNI Outline Helpdesk Computing Wide Area Network Cyber Security Networking and Asset Management

5 Helpdesk Hour 8am-12pm M-F –Submit a CCPR via –Dial x7155 –Send to Windows XP, Vista and RHEL5 Supported Desktops –Migrating older desktops Mac Support?

6 Computing Servers Upgraded –Dovecot IMAP Server (Indexing) –New File Server and IMAP Servers (Farm Nodes) Servers Migrating to Virtual Machines Printing –Centralized Access via jlabprt.jlab.org –Accounting Coming Soon Video Conferencing (working on EVO)

7 Wide Area Network Bandwidth –10Gbps WAN and LAN backbone –Offsite Data Transfer Servers scigw.jlab.org(bbftp) qcdgw.jlab.org(bbcp)

8 Cyber Security Challenge The threat: sophistication and volume of attacks continue to increase. –Phishing Attacks Spear Phishing/Whaling are now being observed at JLab. Federal, including DOE, requirements to meet the cyber security challenges require additional measures. JLab uses a risk based approach that incorporates achieving the mission while at the same time dealing with the threat.

9 Cyber Security Managed Desktops –Skype Allowed From Managed Desktops On Certain Enclaves Network Scanning Intrusion Detection PII/SUI (CUI) Management

10 Networking and IT Asset Management Network Segmentation/Enclaves –Firewalls Computer Registration – Managing IP Addresses –DHCP Assigns all IP addresses (most static) Integrated with registration Automatic Port Configuration –Rolling out now –Uses registration database

11 Scientific Computing Chip Watson & Sandy Philpott

12 SciComp Outline Upgrading the farm Expanding disk cache and /work Migrating to a new tape library (silo) Planning for 12 GeV LQCD

13 Farm Evolution Motivation Capacity upgrades –Re-use of HPC clusters Movement to Open Source –O/S upgrade –Change from LSF to PBS

14 Farm Evolution Timetable Nov 07: Auger/PBS available – RHEL nodes Jan 08: Fedora 8 (F8) available – 50 nodes May 08: Friendly-user mode; IFARML4,5 Jun 08: Production –F8 only; IFARML nodes from LSF IFARML alias Jul 08: IFARML nodes from LSF Aug 08: IFARML nodes from LSF Sep 08: RHEL3/LSF->F8/PBS Migration complete –No renewal of LSF or RHEL for cluster nodes

15 Farm F8/PBS Differences Code must be recompiled –2.6 kernel –gcc 4 Software installed locally via yum –cernlib –Mysql Time limits: 1 day default, 3 days max stdout/stderr to ~/farm_out notification

16 Farm Future Plans Additional nodes –from HPC clusters CY08: ~120 4g nodes CY09-10: ~60 6n nodes –Purchase as budgets allow Support for 64 bit systems when feasible & needed

17 Storage Evolution Deployment of Sun x4500 “thumpers” Decommissioning of Panasas (old /work server) Planned replacement of old cache nodes

18 Tape Library Current STK “Powderhorn” silo is nearing end-of-life –Reaching capacity & running out of blank tapes –Doesn’t support upgrade to higher density cartridges –Is officially end-of-life December 2010 Market trends –LTO (Linear Tape Open) Standard has proliferated since 2000 –LTO-4 is 4x density, capacity/$, and bandwidth of 9940b: 800 GB/tape, $100/TB, 120 MB/s –LTO-5, out next year, will double capacity, 1.5x bandwidth: 1600 GB/tape, 180 MB/s –LTO-6 will be out prior to the 12 GeV era 3200 GB/tape, 270 MB/s

19 Tape Library Replacement Competitive procurement now in progress –Replace old system, support 10x growth over 5 years Phase 1 in August –System integration, software evolution –Begin data transfers, re-use 9940b tapes Tape swap through January 2 PB capacity by November DAQ to LTO-4 in January 2009 Old silo gone in March 2009 End result: breakeven on cost by the end of 2009!

20 Long Term Planning Continue to increase compute & storage capacity in most cost effective manner Improve processes & planning –PAC submission process –12 GeV Planning…

E.g.: Hall B Requirements Event Simulation SPECint_rate2006 sec/event1.8 Number of events1.00E+12 Event size (KB)20 % Stored Long Term10%25% Total CPU ( SPECint_rate2006 )5.7E+04 Petabytes / year (PB)25555 Data Acquisition Average event size (KB)20 Max sustained event rate (kHz) Average event rate (kHz)0010 Average 24-hour duty factor (%)0% 50%60%65% Weeks of operation / year00030 Network (n*10gigE)11111 Petabytes / year st Pass Analysis SPECint_rate2006 sec/event 1.5 Number of analysis passes001.5 Event size out / event size in22222 Total CPU ( SPECint_rate2006 )0.0E E-038.4E-03 Silo Bandwidth (MB/s) Petabytes / year Total SPECint_rate E+04 SPECint_rate2006 / node # nodes needed (current year) Petabytes / year25512

22 LQCD Computing JLab operates 3 clusters with nearly 1100 nodes, primarily for LQCD plus some accelerator modeling National LQCD Computing Project ( : BNL, FNAL, JLab; USQCD Collaboration) LQCD II proposal would double the hardware budget to enable key calculations JLab Experimental Physics & LQCD computing share staff (operations & software development) & tape silo, providing efficiencies for both