Update on NIKHEF D0 farm June 2002. June 6, 2002D0RACE2 Outline Status of D0 farm –Upgrades –Move into grid network Use of EDG testbed and DAS-2 cluster.

Slides:



Advertisements
Similar presentations
Andrew McNab - Manchester HEP - 17 September 2002 Putting Existing Farms on the Testbed Manchester DZero/Atlas and BaBar farms are available via the Testbed.
Advertisements

Voodoo D0 on EDG. Get certificate certificate.nikhef.nl Organization nikhef ? Open ssl necessary on dors?
Upgrade D0 farm. Reasons for upgrade RedHat 7 needed for D0 software New versions of –ups/upd v4_6 –fbsng v1_3f+p2_1 –sam Use of farm for MC and analysis.
ICHEP visit to NIKHEF. D0 Monte Carlo farm hoeve (farm) MC request schuur (barn) SAM MHz 2-CPU nodes 50 * 40 GB 1.2 TB.
Installation mcc from minitar mc_runjob from upd –Copied from /d0dist to /d0/mcc/mc_runjob –Changes in /d0 –Undo mod of Rod Walker.
University of Southampton Electronics and Computer Science M-grid: Using Ubiquitous Web Technologies to create a Computational Grid Robert John Walters.
D0 NIKHEF. March 22, Nijmegen2 Amsterdam D0 farm (SAM station hoeve, node schuur) –2 servers, 50 dual cpu nodes –Used for MC production.
LNL CMS M.Biasotto, Bologna, 29 aprile LNL Analysis Farm Massimo Biasotto - LNL.
NIKHEF Testbed 1 Plans for the coming three months.
McFarm: first attempt to build a practical, large scale distributed HEP computing cluster using Globus technology Anand Balasubramanian Karthik Gopalratnam.
INFN - Ferrara BaBarGrid Meeting SPGrid Efforts in Italy BaBar Collaboration Meeting - SLAC December 11, 2002 Enrica Antonioli - Paolo Veronesi.
OpenVMS System Management A different perspective by Andy Park TrueBit b.v.
GRID workload management system and CMS fall production Massimo Sgaravatto INFN Padova.
Israel Cluster Structure. Outline The local cluster Local analysis on the cluster –Program location –Storage –Interactive analysis & batch analysis –PBS.
EU funding for DataGrid under contract IST is gratefully acknowledged GridPP Tier-1A Centre CCLRC provides the GRIDPP collaboration (funded.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Chapter 7 Microsoft Windows XP. Windows XP Versions XP Home XP Home XP Professional XP Professional XP Professional 64-Bit XP Professional 64-Bit XP Media.
Dave Newbold, University of Bristol24/6/2003 CMS MC production tools A lot of work in this area recently! Context: PCP03 (100TB+) just started Short-term.
27/04/05Sabah Salih Particle Physics Group The School of Physics and Astronomy The University of Manchester
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
HEP Experiment Integration within GriPhyN/PPDG/iVDGL Rick Cavanaugh University of Florida DataTAG/WP4 Meeting 23 May, 2002.
1 Data Management D0 Monte Carlo needs The NIKHEF D0 farm The data we produce The SAM data base The network Conclusions Kors Bos, NIKHEF, Amsterdam Fermilab,
October, Scientific Linux INFN/Trieste B.Gobbo – Compass R.Gomezel - T.Macorini - L.Strizzolo INFN - Trieste.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
BaBar MC production BaBar MC production software VU (Amsterdam University) A lot of computers EDG testbed (NIKHEF) Jobs Results The simple question:
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Alain Romeyer - 15/06/20041 CMS farm Mons Final goal : included in the GRID CMS framework To be involved in the CMS data processing scheme.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina, L.Lueking,
NIKHEF Test Bed Status David Groep
CHEP 2003Stefan Stonjek1 Physics with SAM-Grid Stefan Stonjek University of Oxford CHEP th March 2003 San Diego.
SAMGrid as a Stakeholder of FermiGrid Valeria Bartsch Computing Division Fermilab.
LHCb and DataGRID - the workplan for 2001 Eric van Herwijnen Wednesday, 28 march 2001.
1 The D0 NIKHEF Farm Kors Bos Ton Damen Willem van Leeuwen Fermilab, May
Partner Logo German Cancio – WP4-install LCFG HOW-TO - n° 1 LCFGng configuration examples Updated 10/2002
July 28' 2011INDIA-CMS_meeting_BARC1 Tier-3 TIFR Makrand Siddhabhatti DHEP, TIFR Mumbai July 291INDIA-CMS_meeting_BARC.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH OS X Home server AFS using openafs 3 DB servers Kerberos 4 we will move.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
26SEP03 2 nd SAR Workshop Oklahoma University Dick Greenwood Louisiana Tech University LaTech IAC Site Report.
G. Cancio, L. Cons, Ph. Defert - n°1 October 2002 Software Packages Management System for the EU DataGrid G. Cancio Melia, L. Cons, Ph. Defert. CERN/IT.
DataGRID WPMM, Geneve, 17th June 2002 Testbed Software Test Group work status for 1.2 release Andrea Formica on behalf of Test Group.
London Tier 2 Status Report GridPP 11, Liverpool, 15 September 2004 Ben Waugh on behalf of Owen Maroney.
NIKHEF Test Bed Status David Groep
Southgrid Technical Meeting Pete Gronbech: May 2005 Birmingham.
GridPP11 Liverpool Sept04 SAMGrid GridPP11 Liverpool Sept 2004 Gavin Davies Imperial College London.
August 30, 2002Jerry Gieraltowski Launching ATLAS Jobs to either the US-ATLAS or EDG Grids using GRAPPA Goal: Use GRAPPA to launch a job to one or more.
GRID activities in Wuppertal D0RACE Workshop Fermilab 02/14/2002 Christian Schmitt Wuppertal University Taking advantage of GRID software now.
LCG LCG-1 Deployment and usage experience Lev Shamardin SINP MSU, Moscow
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Outline: Status: Report after one month of Plans for the future (Preparing Summer -Fall 2003) (CNAF): Update A. Sidoti, INFN Pisa and.
Re-Reconstruction Of Generated Monte Carlo In a McFarm Context 2003/09/26 Joel Snow, Langston U.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
International Workshop on HEP Data Grid Aug 23, 2003, KNU Status of Data Storage, Network, Clustering in SKKU CDF group Intae Yu*, Joong Seok Chae Department.
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
Separate distribution of the analysis code (and more) P. Hristov 19/03/2014.
Installing the Galaxie Driver The 335 DAD Galaxie driver is installed in two steps: 1.The 335 DAD driver files are copied to the Galaxie driver directory.
15-Feb-02Steve Traylen, RAL WP6 Test Bed Report1 RAL/UK WP6 Test Bed Report Steve Traylen, WP6 PPGRID/RAL, UK
Mass Storage at SARA Peter Michielse (NCF) Mark van de Sanden, Ron Trompert (SARA) GDB – CERN – January 12, 2005.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
Tests at Saclay D. Calvet, A. Formica, Z. Georgette, I. Mandjavidze, P. Micout DAPNIA/SEDI, CEA Saclay Gif-sur-Yvette Cedex.
How to run MC on the D0 farm. Steps On hoeve (user fbsuser) –Get MC request –Create macro –Submit jobs On schuur (user willem) –Store files into SAM –Clear.
Status report NIKHEF Willem van Leeuwen February 11, 2002 DØRACE.
NL Service Challenge Plans
The Linux Operating System
Willem van Leeuwen SAM Users Meeting June 27, 2001
NIKHEF/Nijmegen analysis farm status
UM D0RACE STATION Status Report Chunhui Han June 20, 2002
User interaction and workflow management in Grid enabled e-VLBI experiments Dominik Stokłosa Poznań Supercomputing and Networking Center, Supercomputing.
Status report NIKHEF Willem van Leeuwen February 11, 2002 DØRACE.
Presentation transcript:

Update on NIKHEF D0 farm June 2002

June 6, 2002D0RACE2 Outline Status of D0 farm –Upgrades –Move into grid network Use of EDG testbed and DAS-2 cluster –Modifications of mc_runjob Some conclusions

June 6, 2002D0RACE3 Configuration of D0 farm All software on server –Easy modification of code Nodes booted from server –Easy replacement of nodes Designed for D0 MCC but usable for others –Antares –L3/cosmics

June 6, 2002D0RACE4 Upgrades Farm server and nodes upgraded to RH 7.2 File server still RH 6.2 (no driver for RAID) SAM fbsng mc_runjob p

June 6, 2002D0RACE5 Experiences after upgrade Performance of SAM improved –Almost no resubmits MCC problems –Filenames too long Accepted by SAM, rejected by Enstore –Complex division by zero in pythia

June 6, 2002D0RACE6 How to get more cpu’s for D0 Use idling nodes of –EDG testbed (now 20 cpu’s) –DAS-2 cluster (now 64 cpu’s) Therefore: –Move D0 farm into grid network –Adapt mc_runjob

June 6, 2002D0RACE7 Old network layout hoeve schuur switch router hefnet surfnet fbs sam

June 6, 2002D0RACE8 New network layout farmrouter D0 hefnet lambda hoeve schuur das-2 sam fbs EDG testbed

June 6, 2002D0RACE9 New configuration All nodes behind router Almost all servers behind router –DAS-2 must be reachable from outside Software now installed on nodes (LCFG) SAM station on schuur

June 6, 2002D0RACE10 Modification of mc_runjob remove calls to batch from RunJob.py separate code and data package code in rpm –1.2 GB – takes hours to create RPM install rpm on nodes with LCFG

June 6, 2002D0RACE11 Grid job #!/bin/sh macro=$1 pwd=`pwd` cd /opt/fnal/d0/mcc/mcc-dist. mcc_dist_setup.sh cd $pwd dir=/opt/fnal/d0/mcc/mc_runjob/py_script python $dir/Linker.py script=$macro willem]$ cat test.pbs #PBS -l nodes=1 # Changing to directory as requested by user cd /home/willem # Executing job as requested by user./submit minbias.macro PBS jobsubmit

June 6, 2002D0RACE12 RunJob class for grid class RunJob_farm(RunJob_batch) : def __init__(self,name=None) : RunJob_batch.__init__(self,name) self.myType="runjob_farm" def Run(self) : self.jobname = self.linker.CurrentJob() self.jobnaam = string.splitfields(self.jobname,'/')[-1] comm = 'chmod +x ' + self.jobname commands.getoutput(comm) if self.tdconf['RunOption'] == 'RunInBackground' : RunJob_batch.Run(self) else : bq = self.tdconf['BatchQueue'] dirn = os.path.dirname(self.jobname) comm = 'cd ' + dirn + '; sh ' + self.jobnaam + ' `pwd` >& stdout' runcommand(comm)

June 6, 2002D0RACE13 First experiences Jobs submitted on EDG and DAS-2 with PBS Wait between the submission of jobs! Don’t forget the time limit –Specs different on EDG and DAS-2 –We need dg-job-submit

June 6, 2002D0RACE14 Some conclusions The EDG setup is more scalable but (yet) less flexible We have to rethink MCC –Smaller tar balls –Location of card files –Location of minimum bias files We started the irreversible road to the Grid