HEPiX Meeting Summary Autumn 2000 Jefferson Laboratory, Newport News, Virginia, USA.

Slides:



Advertisements
Similar presentations
The RHIC-ATLAS Computing Facility at BNL HEPIX – Edinburgh May 24-28, 2004 Tony Chan RHIC Computing Facility Brookhaven National Laboratory.
Advertisements

Parallel Session B2 - CPU and Resource Allocation " Panelists: – Charles Young (BaBar) – David Bigagli " Seed Questions: – Batch queuing system in use?
4/2/2002HEP Globus Testing Request - Jae Yu x Participating in Globus Test-bed Activity for DØGrid UTA HEP group is playing a leading role in establishing.
CERN LCG Overview & Scaling challenges David Smith For LCG Deployment Group CERN HEPiX 2003, Vancouver.
HEPiX Meeting Wrap Up Fall 2000 JLab. Meeting Highlights Monitoring –Several projects underway –Collaboration of ideas occurred –Communication earlier.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Jean-Yves Nief, CC-IN2P3 Wilko Kroeger, SCCS/SLAC Adil Hasan, CCLRC/RAL HEPiX, SLAC October 11th – 13th, 2005 BaBar data distribution using the Storage.
GRID Workload Management System Massimo Sgaravatto INFN Padova.
Site report: CERN Helge Meinhard (at) cern ch HEPiX fall SLAC.
1 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Mass Storage For BaBar at SLAC Andrew Hanushevsky Stanford.
Server Operating Systems Last Update Copyright Kenneth M. Chipps Ph.D. 1.
Backup Rationalisation Reorganisation of the CERN Computer Centre Backups David Asbury IT/DS Friday 6 December 2002.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
HEPIX 3 November 2000 Current Mass Storage Status/Plans at CERN 1 HEPIX 3 November 2000 H.Renshall PDP/IT.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
HEPiX Catania 19 th April 2002 Alan Silverman HEPiX Large Cluster SIG Report Alan Silverman 19 th April 2002 HEPiX 2002, Catania.
HEPiX Orsay 27 th April 2001 Alan Silverman HEPiX Large Cluster SIG Report Alan Silverman 27 th April 2001 HEPiX 2001, Orsay.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
CC - IN2P3 Site Report Hepix Fall meeting 2009 – Berkeley
CASPUR Site Report Andrei Maslennikov Sector Leader - Systems Catania, April 2001.
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
October, Scientific Linux INFN/Trieste B.Gobbo – Compass R.Gomezel - T.Macorini - L.Strizzolo INFN - Trieste.
CERN IT Department CH-1211 Genève 23 Switzerland t Experience with Windows Vista at CERN Rafal Otto Internet Services Group IT Department.
LAL Site Report Michel Jouvin LAL / IN2P3
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Nov 1, 2000Site report DESY1 DESY Site Report Wolfgang Friebel DESY Nov 1, 2000 HEPiX Fall
23 Oct 2002HEPiX FNALJohn Gordon CLRC-RAL Site Report John Gordon CLRC eScience Centre.
Overview of day-to-day operations Suzanne Poulat.
20-22 September 1999 HPSS User Forum, Santa Fe CERN IT/PDP 1 History  Test system HPSS 3.2 installation in Oct 1997 IBM AIX machines with IBM 3590 drives.
Using Virtual Servers for the CERN Windows infrastructure Emmanuel Ormancey, Alberto Pace CERN, Information Technology Department.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Developing & Managing A Large Linux Farm – The Brookhaven Experience CHEP2004 – Interlaken September 27, 2004 Tomasz Wlodek - BNL.
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
Integrating JASMine and Auger Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
April 25, 2001HEPiX/HEPNT FERMI SITE REPORT Lisa Giacchetti.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
4-8 th October 1999CERN Site Report, HEPiX SLAC. A.Silverman CERN Site Report HEPNT/HEPiX October 1999 SLAC Alan Silverman CERN/IT/DIS.
Condor Usage at Brookhaven National Lab Alexander Withers (talk given by Tony Chan) RHIC Computing Facility Condor Week - March 15, 2005.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility (formerly CEBAF - The Continuous Electron Beam Accelerator Facility)
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National Laboratory.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
HEPiX FNAL ‘02 25 th Oct 2002 Alan Silverman HEPiX Large Cluster SIG Report Alan Silverman 25 th October 2002 HEPiX 2002, FNAL.
26/4/2001LAL Site Report - HEPix - LAL 2001 LAL Site Report HEPix – LAL Apr Michel Jouvin
Cluster Configuration Update Including LSF Status Thorsten Kleinwort for CERN IT/PDP-IS HEPiX I/2001 LAL Orsay Tuesday, December 08, 2015.
TiBS Fermilab – HEPiX-HEPNT Ray Pasetes October 22, 2003.
14 th April 1999CERN Site Report, HEPiX RAL. A.Silverman CERN Site Report HEPiX April 1999 RAL Alan Silverman CERN/IT/DIS.
HEPiX 2 nd Nov 2000 Alan Silverman Proposal to form a Large Cluster SIG Alan Silverman 2 nd Nov 2000 HEPiX – Jefferson Lab.
November 1, 2000HEPiX/HEPNT FERMI SITE REPORT Lisa Giacchetti.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Lisa Giacchetti AFS: What is everyone doing? LISA GIACCHETTI Operating Systems Support.
HEPiX/HEPNT report Helge Meinhard, Alberto Pace, Denise Heagerty / CERN-IT Computing Seminar 05 November 2003.
CERN - European Organization for Nuclear Research Windows 2000 Update FOCUS June 13 th, 2002.
W.A.Wojcik/CCIN2P3, Nov 1, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
1/11/2000LAL Site Report - HEPix - JLab 2000 LAL Site Report HEPix – Jlab Nov Michel Jouvin
Batch Software at JLAB Ian Bird Jefferson Lab CHEP February, 2000.
A UK Computing Facility John Gordon RAL October ‘99HEPiX Fall ‘99 Data Size Event Rate 10 9 events/year Storage Requirements (real & simulated data)
Status of W2K at INFN Gian Piero Siroli, Dept. of Physics, Univ. of Bologna and INFN HEPiX-HEPNT 2000, Jefferson Lab.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Jefferson Lab Site Report Sandy Philpott HEPiX Fall 07 Genome Sequencing Center Washington University at St. Louis.
Scientific Linux Connie Sieh CSAM Meeting May 2, 2006.
CCIN2P3 Site Report - BNL, Oct 18, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center.
CC - IN2P3 Site Report Hepix Spring meeting 2011 Darmstadt May 3rd
Presentation transcript:

HEPiX Meeting Summary Autumn 2000 Jefferson Laboratory, Newport News, Virginia, USA

Site Reports - 1 Jefferson lab (host): 4-6 GEV continuous beam electron accelerator. 600 staff, 70M$/year budget. STK silo with 8 Redwood (replace by soon) drives. From 2005 double energy and run at MB/sec to accumulate 3 PB/year. Also do lattice QCD and propose to buy 256 node Compaq Alpha cluster this year. Have negotiated good price for LSF. IN2P3: French national and Babar regional centre. 6 STK silos but limited Redwood. Have reserved 40 Mbits/sec line to Slac for Babar data. Bought an 96 node IBM Linux netfinity cluster in 1 AU high units in single frame. Working on 64-bit version of RFIO. FNAL: Started certifying Solaris 8 and Linux RH 7.ENSTORE (cf CASTOR) in ‘limited’ production use. Working on migrating to Kerberos 5 (MIT) with heavy mods for AFS. Have bought Linux PCs in a racked solution.

Site Reports - 2 RAL: Have closed HP farm. Looking for robotics upgrade – STK, IBM or ADIC (ex Grau-Abba). Problem that Babar want Linux RH 6.2 while CDF want FNAL Linux. BNL: Have 1200 node home made QCD-SP supercomputer. Have VA-Linux rack mounted PCs and will use HP Open-View for site security management LAL: physicists use Windows NT4 for desktop and pressure for Linux is lowering. They find VMWARE (Linux under NT) useful. DESY: Working on a centrally managed and fully transparent user disk cache (ex-Eurostore 2 in my opinion). Use CODINE for batch - recently bought by SUN and made open source – and LSF but will drop LSF (too expensive). Have installed Kerberos 5 with AFS support (Swedish, not MIT version) but not yet in production. 6 man months work and worker is leaving !

Site Reports - 3 INFN: Heavy use of CONDOR – 200 machines since 2 years. Building national GRID to be synchronised with CERN GRID. No plans to move central facilities to Linux because of lack of management tools. SACLAY: Running Veritas netbackup for all machines. SLAC: Going to rack mounted Linux. Largest HPSS site in volume (200 TB cf 25 at CERN). Will add second instance of HPSS to support general staging. Plan to add drive to existing but currently getting 1/3 write media failures in field test. LBL Parallel Distributed Systems Facility: Support Atlas and Rich (BNL) using AFS knfs gateway to CERN and BNL. Use HPSS at NERSC. Local farm of rack mounted Linux PCs.

AFS discussion with IBM-Transarc IBM Transarc lab suport AFS, DFS and DCE. Standard IBM support pricing using local support (IBM Suisse) as first level. Local support will be trained by IBM Transarc. Currently AFS 3.6 (supporting RH Linux kernels 2.2.x) and no plans for 3.7. Admit AFS does not generate a lot of revenue. Interested in Kerberos 5 support but think this should be done by Open-AFS in which case may take it back into product (not convincing !). Will release Solaris 8 and W2K clients soon (last Friday!). W2K is only tolerant, not MSI certified. Essentially a WNT port using no special W2K features. Includes some cache corruption bug fixes (hopefully ours !). Plan future NAS/SAN Enterprise-wide file system to replace AFS and DCE- 2/3 years and driven by IBM San Jose. AFS development team same size but partly moved to India. Open AFS source tree not maintained by IBM (probably CMU). Official end of 3.6 support end 2002 (changed to ‘after next major release!) HEPIX concludes AFS is now in maintenance mode from IBM so want their customers to either go to Open AFS or Enterprise system in a few years.

Cluster Monitoring and Control FNAL developing Next Generation Operations for farm management. Has concept of fast and slow monitoring streams unlike CERN. CERN developing PEM. Trying to follow standards. May collaborate with Sun Java competence centre. SLAC have enhanced RANGER tool - based on perl scripts and rule-sets but no system overview features. IN2P3 have just begun to develop Global Monitoring system. They think NGOP and PEM are too complicated. Conclusion –Several projects underway –Collaboration of ideas occurred –Communication earlier in the process may have resulted in more collaboration

Batch systems Jefferson - migrating from LSF to free Public Batch System (PBS) from Nasa with good experiences. Wisconsin - home of Condor, good overview. A live product we should try. Not a batch system however (desktop spare cycles). Platform - a sales talk for LSF ! Good directions for farming architecture. IN2P3 - continuing to enhance their home written BQS including adding a Java GUI interface both for users and administrators. FNAL - have rewritten FBS as FBSNG (farms batch system next generation) to be now independent of LSF. Conclusions –as before no common approach from HEP but gives a choice for the future if we want to stop LSF.

Mass Storage CERN - wants to migrate off Redwoods to (probably) STK 9940 –deploying CASTOR phase 1 now for all new experiment data –will keep HPSS for now for “user tapes” data –will tender at end of 2001 for next 4-5 years robotics/drives Jefferson - have been using OSM but now a dead product so developing their own Jasmine tape/disk data mover system. They already have a mature disk pool management system. Conclusion –Various projects for Mass storage interfaces: Castor, JASMine, Enstore –Continued problems with STK Redwood drives (all sites in discussion) –Reported problems with STK 9940 (SLAC, not yet seen at Cern)

Grid Projects at all major labs related to data grid - INFN internal grid to ‘synchronise’ with CERN - US GriPhyN internal physics grid - RAL developing a Globus infrastructure between it and FNAL for CDF collaboration work (4 UK universities + RAL + FNAL) HEP distributed computing model can exploit the concepts of the Grid work Technical and management solutions need to be developed It’s clear that coordination between sites/labs will be needed - danger of incompatible European and US Grids

Linux Usage continues to grow. Most sites using rack mounted either vendor integrated or just pizza boxes in standard racks. Expertise in management also continues to grow Commodity computing analysis farms are in our future Human and $ resources are in short supply everywhere

Work To Be Done - 1 Joint OS certification –Cern, Fermi and Slac starting on project to jointly certify Linux and Solaris OS’s –Aim to limit duplication of effort –Are there others who would like to work on this? HEPiX mail list clean up –Lists are moving to listserv.fnal.gov (web interface) Default setup requires list owner permission to subscribe but this process is automated Anyone on list can post to list Detailed info on listserv usage online HEPiX web pages need to be updated with new info

Work To Be Done - 2 Large Cluster SIG –Alan Silverman organizing this special interest group –Proposed to meet separately from HEPiX/HEPNT Would also report status at HEPiX/HEPNT Primary goals: –Keep sites aware of what relevant work is in progress or even planned –Be aware and promote collaboration

Future Meetings LAL and Cern have volunteered to host spring ’01 meeting –Since LAL offered first we will pursue this option to start –Plan for meeting in April (Easter is 4/15) Joint meeting with HEPNT will continue Volunteers for Fall ’01 North American meeting?