Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA 757-269-7770.

Slides:



Advertisements
Similar presentations
The RHIC-ATLAS Computing Facility at BNL HEPIX – Edinburgh May 24-28, 2004 Tony Chan RHIC Computing Facility Brookhaven National Laboratory.
Advertisements

Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
1 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Mass Storage For BaBar at SLAC Andrew Hanushevsky Stanford.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
Week 6 Operating Systems.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
UCL Site Report Ben Waugh HepSysMan, 22 May 2007.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
Design & Management of the JLAB Farms Ian Bird, Jefferson Lab May 24, 2001 FNAL LCCWS.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
1 Computing & Networking User Group Meeting Roy Whitney Andy Kowalski Sandy Philpott Chip Watson 17 June 2008.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
An Overview of PHENIX Computing Ju Hwan Kang (Yonsei Univ.) and Jysoo Lee (KISTI) International HEP DataGrid Workshop November 8 ~ 9, 2002 Kyungpook National.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
NLIT May 26, 2010 Page 1 Computing Jefferson Lab Users Group Meeting 8 June 2010 Roy Whitney CIO & CTO.
Scientific Computing Experimental Physics Lattice QCD Sandy Philpott May 20, 2011 IT Internal Review 12GeV Readiness.
Module 9: Preparing to Administer a Server. Overview Introduction to Administering a Server Configuring Remote Desktop to Administer a Server Managing.
HEPiX/HEPNT TRIUMF,Vancouver 1 October 18, 2003 NIKHEF Site Report Paul Kuipers
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility HEPiX – Fall, 2005.
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
Linux Servers with JASMine K. Edwards, A. Kowalski, S. Philpott HEPiX May 21, 2003.
28 April 2003Imperial College1 Imperial College Site Report HEP Sysman meeting 28 April 2003.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
JLab Scientific Computing: Theory HPC & Experimental Physics Thomas Jefferson National Accelerator Facility Newport News, VA Sandy Philpott.
Integrating JASMine and Auger Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
Fermi National Accelerator Laboratory SC2006 Fermilab Data Movement & Storage Multi-Petabyte tertiary automated tape store for world- wide HEP and other.
Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski.
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility (formerly CEBAF - The Continuous Electron Beam Accelerator Facility)
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Disk Farms at Jefferson Lab Bryan Hess
26/4/2001LAL Site Report - HEPix - LAL 2001 LAL Site Report HEPix – LAL Apr Michel Jouvin
The KLOE computing environment Nuclear Science Symposium Portland, Oregon, USA 20 October 2003 M. Moulson – INFN/Frascati for the KLOE Collaboration.
OSG Abhishek Rana Frank Würthwein UCSD.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
Outline: Status: Report after one month of Plans for the future (Preparing Summer -Fall 2003) (CNAF): Update A. Sidoti, INFN Pisa and.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Batch Software at JLAB Ian Bird Jefferson Lab CHEP February, 2000.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
Oct. 6, 1999PHENIX Comp. Mtg.1 CC-J: Progress, Prospects and PBS Shin’ya Sawada (KEK) For CCJ-WG.
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
Patrick Gartung 1 CMS 101 Mar 2007 Introduction to the User Analysis Facility (UAF) Patrick Gartung - Fermilab.
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Jefferson Lab Site Report Sandy Philpott HEPiX Fall 07 Genome Sequencing Center Washington University at St. Louis.
Systems Software / The Operating System CSC October 14, 2010.
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
Compute and Storage For the Farm at Jlab
Module 9: Preparing to Administer a Server
Ákos Frohner EGEE'08 September 2008
TYPES OFF OPERATING SYSTEM
A Web-Based Data Grid Chip Watson, Ian Bird, Jie Chen,
Chapter 2: System Structures
Module 9: Preparing to Administer a Server
Presentation transcript:

Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA HEPiX – October, 2004

Central Computing –Distracted by SPAM problem –Evaluated and purchased MXLogic Offsite solution Filters virus/spam before getting to Lab –Upgraded our hardware Windows builds –Purchased MS Enterprise Agreement –Developed an automatic build process –Upgrading all of our systems to Windows XP –Still evaluating SP2, problems with CAD, etc.

File Server Storage Adaptec 2200S Raid and Linux XFS –Linux kernel 2.6 and Adaptec firmware (build 7244) It doesn’t work (I/O errors, etc.) –RedHat EL3 WS kernel works fine, but no XFS support –Tested ext3 performance unacceptable (20MB/s read, 34MB/s write) XFS performance (approx 100MB/s read/write) –Dropped back to prior Adaptec BIOS and 2.6 kernel works fine

File Server Storage (cont) Purchased 2 StorageTek B280 systems –14 TB of disk space –4 Sun V210 head units –Stable, but slow, NFS performance Aggregate -- 6MB/s write, 63MB/s read Each node MB/s write, 1.4MB/s read average

File Server Storage (cont) Evaluating 10TB Panasas system –Tested 2 protocols (directFLOW and NFS) –No directFLOW problems –NFS finally stable at version 2.1.4c –Good performance with either Aggregate MB/s write, MB/s read Each node – MB/s write, MB/s read

Jasmine Changes Jasmine is Jlab’s mass storage system (disk+tape) stores ~1PB and can routinely move 20TB/day. Disk cache system recently rewritten for performance and reliability –I/O load spread out over pool of many disk servers –Files belong to file groups (per experiment) with quotas –Quotas may be exceeded if there is enough disk space; allows more flexible use of disk –Files deleted from servers in a modified LRU fashion –Files may be pinned until used by the batch farm

Jasmine changes (2) New programmatic interfaces for –Batch Farm (Auger) –Other services that need to move files (SRM, DAQ, LQCD disk cache) More reliance on MySQL database; concurrency and load are challenging Writing 9940B tapes Experiment data rates now ~30MB/sec

Auger Changes Auger is Jlab’s Batch farm management system. Uses LSF to run jobs, keeps accounting in a database for web or command line presentation. Users can submit thousands of jobs using a compact job description that includes file retrieval and storage. Interfaces with Jasmine to stage files to disk before the job runs on the farm to keep CPUs busy

Jasmine & Auger Web Interface Java Server Pages

Projects upgrade –Still evaluating software/hardware Desktop systems –MacOS-X –Linux, Unix –Windows Power/Cooling issues –Reached limit of current Computer Room –New Computer Center to open in Jan 2006 –Increased power requirements for 800 MHz FSB systems 1.3A to 2.1A (single CPU) 1.6A to 2.8A (dual CPU) –Shutdown problems with non-ACPI enabled systems