D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina, L.Lueking,

Slides:



Advertisements
Similar presentations
Categories of I/O Devices
Advertisements

Buffers & Spoolers J L Martin Think about it… All I/O is relatively slow. For most of us, input by typing is painfully slow. From the CPUs point.
Operating System.
ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen1 Offline Mass Data Processing using Online Computing Resources at HERA-B José Hernández DESY-Zeuthen.
Device Management.
Computer Organization and Architecture
On-Demand Media Streaming Over the Internet Mohamed M. Hefeeda, Bharat K. Bhargava Presented by Sam Distributed Computing Systems, FTDCS Proceedings.
NovaBACKUP 10 xSP Technical Training By: Nathan Fouarge
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
The D0 Monte Carlo Challenge Gregory E. Graham University of Maryland (for the D0 Collaboration) February 8, 2000 CHEP 2000.
CHAPTER 2 OPERATING SYSTEM OVERVIEW 1. Operating System Operating System Definition A program that controls the execution of application programs and.
Operating System. Architecture of Computer System Hardware Operating System (OS) Programming Language (e.g. PASCAL) Application Programs (e.g. WORD, EXCEL)
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
GLAST LAT ProjectDOE/NASA Baseline-Preliminary Design Review, January 8, 2002 K.Young 1 LAT Data Processing Facility Automatically process Level 0 data.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
Chapter 5 Operating System Support. Outline Operating system - Objective and function - types of OS Scheduling - Long term scheduling - Medium term scheduling.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
CHEP'07 September D0 data reprocessing on OSG Authors Andrew Baranovski (Fermilab) for B. Abbot, M. Diesburg, G. Garzoglio, T. Kurca, P. Mhashilkar.
Block1 Wrapping Your Nugget Around Distributed Processing.
Stephen Wolbers CHEP2000 February 7-11, 2000 Stephen Wolbers CHEP2000 February 7-11, 2000 CDF Farms Group: Jaroslav Antos, Antonio Chan, Paoti Chang, Yen-Chu.
CDF Offline Production Farms Stephen Wolbers for the CDF Production Farms Group May 30, 2001.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Online Reconstruction used in the Antares-Tarot alert system J ü rgen Brunner The online reconstruction concept Performance Neutrino doublets and high.
CLASS Information Management Presented at NOAATECH Conference 2006 Presented by Pat Schafer (CLASS-WV Development Lead)
1 Database mini workshop: reconstressing athena RECONSTRESSing: stress testing COOL reading of athena reconstruction clients Database mini workshop, CERN.
Data reprocessing for DZero on the SAM-Grid Gabriele Garzoglio for the SAM-Grid Team Fermilab, Computing Division.
Lee Lueking 1 The Sequential Access Model for Run II Data Management and Delivery Lee Lueking, Frank Nagy, Heidi Schellman, Igor Terekhov, Julie Trumbo,
Outline: Tasks and Goals The analysis (physics) Resources Needed (Tier1) A. Sidoti INFN Pisa.
The KLOE computing environment Nuclear Science Symposium Portland, Oregon, USA 20 October 2003 M. Moulson – INFN/Frascati for the KLOE Collaboration.
Silberschatz, Galvin and Gagne  Operating System Concepts UNIT II Operating System Services.
SECTION 5: PERFORMANCE CHRIS ZINGRAF. OVERVIEW: This section measures the performance of MapReduce on two computations, Grep and Sort. These programs.
L3 DAQ Doug Chapin for the L3DAQ group DAQShifters Meeting 10 Sep 2002 Overview of L3 DAQ uMon l3xqt l3xmon.
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
June 17th, 2002Gustaaf Brooijmans - All Experimenter's Meeting 1 DØ DAQ Status June 17th, 2002 S. Snyder (BNL), D. Chapin, M. Clements, D. Cutts, S. Mattingly.
CIS250 OPERATING SYSTEMS Chapter One Introduction.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Outline: Status: Report after one month of Plans for the future (Preparing Summer -Fall 2003) (CNAF): Update A. Sidoti, INFN Pisa and.
Input/Output Problems Wide variety of peripherals —Delivering different amounts of data —At different speeds —In different formats All slower than CPU.
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
CPU on the farms Yen-Chu Chen, Roman Lysak, Miro Siket, Stephen Wolbers June 4, 2003.
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, R. Brock,T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina,
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Alien and GSI Marian Ivanov. Outlook GSI experience Alien experience Proposals for further improvement.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
Oct. 6, 1999PHENIX Comp. Mtg.1 CC-J: Progress, Prospects and PBS Shin’ya Sawada (KEK) For CCJ-WG.
LHCb Computing activities Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group.
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
Status report NIKHEF Willem van Leeuwen February 11, 2002 DØRACE.
U.S. ATLAS Grid Production Experience
Operating System.
Bernd Panzer-Steindel, CERN/IT
SAM at CCIN2P3 configuration issues
TYPES OFF OPERATING SYSTEM
Near Real Time Reconstruction of PHENIX Run7 Minimum Bias Data From RHIC Project Goals Reconstruct 10% of PHENIX min bias data from the RHIC Run7 (Spring.
Status report NIKHEF Willem van Leeuwen February 11, 2002 DØRACE.
Presentation transcript:

D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina, L.Lueking, L.Loebel-Carpenter, I.Mandrichenko, C.Moore, S.Naymola, A.Moibenko, D.Petravick, M.Przybycien, H.Schellman, K.Shepelak, I.Terekhov, S.Timm, J.Trumbo, S.Veseli, M.Vranicar, R.Wellner, S.White, V.White

D0 Farms 2 D0 Farm needs 250K event size 50Hz trigger rate –peak rate of 12.5 MB/sec –DC is less but reprocessing will bring back up Reconstruction seconds/event on 750 MHz PIII –need 250->>500 CPU’s to handle peak rate –DC is 40% of peak –time constant for 1 GB file is hours. Reality Goal

D0 Farms 3

4 I/O machine Purpose –split/merge of farm output –Serve home areas –Batch system control –File delivery master D0bbin –4 CPU SGI –2 GB ethernet cards –4 72 GB disk partitions (2 way stripe) –peak I/O rates of MB/sec

D0 Farms 5 Worker Nodes 40 Dual Pentium III 500MHz –256MB/CPU 50 Dual Pentium III 750MHz –512 MB/CPU 2 data disks (18 GB) + 6GB system 100Mb ethernet CD/floppy for system configuration

D0 Farms 6 Design Principles Use existing facilities –SAM/Enstore for data access and file tracking –Farm batch system (FBS) for most job control Keep D0 farm control scripts to a minimum –Batch system assigns machines –Data access system decides which file you get If worker process or machine dies, lose minimal number of files and don’t affect other processes No heroic recovery measures, track and resubmit those files

D0 Farms 7 Worker Configuration Workers act as generic FNAL farm machines –Only customization is pnfs for file delivery, home area mount and startup of sam daemons on reboot. –D0 code environment downloads at job start –data access through SAM/encp/rcp, database server Batch system assigns workers to job, not D0FARM control process. D0FARM control never knows which workers are assigned to a job and does not need to.

D0 Farms 8 Data Access is SAM/enstore Integrated data handling system File and process data base Data base server File servers Enstore File delivery systems Pnfs file system Farm Perspective Can tell it you want a set of files Can ask for the ‘next’ file Can flag file as processed or error Can get detailed accounting on what happened Data transfers are from ~ 12 mover nodes to 90 farm nodes through 6509 switch – theoretically could move 100’s of MB/sec Reality – online system has priority for drives.

D0 Farms 9 Farm Batch System Typical Farm Job SECTION START EXEC=startjob parameters QUEUE=D0bbin SECTION WORKER EXEC=runjob parameters NWORKERS=20 QUEUE=D0worker SECTION END EXEC=stopjob parameters QUEUE=D0bbin DEPEND WORKER(done) Queue tells the system what kind of machine to run on and how many. EXEC gives the script name and parameters DEPEND allows cleanup section to run when all worker sections are done. FBS assigns temporary disk on workers On end yanks disk and kills all processes.

D0 Farms 10 START: –Tell SAM which files you will want –Go into wait state until get end signal WORKER: runs on N nodes –Download D0 environment –Inform SAM ready for data –Ask for SAM for next file –Process file and store output to output buffer –Inform SAM of success and ask for next file –On error or end of list, terminate. END: –Create job summary –Send message to Start process telling it to shut down the SAM connection for input Structure of a Farm Job

D0 Farms 11 50% of dual CPU 100% of dual Farm Batch System Monitor Efficient use of of CPU

D0 Farms 12 Query to see which input files were processed by a job

D0 Farms 13 Status System has been in use for MC processing since before CHEP 2000 System has been processing data as it comes off the D0 detector since March 2001 Hardware/control/monitoring can handle full data rates well but… Major problem is speed of executable and data expansion during detector debugging –Output size is ~ input size by design –Currently factor of 2-3 larger due to debugging info. –Better thresholds and less noise will make life much easier Farms get more stress at beginning of run than later!!

D0 Farms 14 Results of typical farm startup 24.5 MB/sec Cold start of ½ of the D0 farm. 90 receiver nodes 141 files of average size 376 MB Read from 2-3 network mounted Mammoth II tapes over 100 MB ethernet at ~10MB/sec/drive. Elapsed time of 44 minutes. This is twice peak rate from the detector.

D0 Farms 15 Current Production Plot from FBSWWW product – out of the box

D0 Farms 16 W  e candidate Run Event Electron candidate recorded using EM trigger p T = 38 GeV isolation<0.2 EM fraction = 0.97 Pierre Petroff and Laurent Duflot Layer 1 8 GeV Layer 2 14 GeV Layer 3 15 GeV Layer GeV hot cell

D0 Farms 17 Future It works now! but we will still: Add ~100 more nodes over next 6 months Make Improvements in automated running Datasets currently defined and submitted by hand ~.25 FTE but still too much Local caching of files Guarantee tape streams at full speed Don’t waste tape mounts if process file multiple times