May. 11, 2015 David Lawrence JLab Counting House Operations.

Slides:



Advertisements
Similar presentations
Elliott Wolin CLAS12 Software Workshop U of Richmond 25-May-2010.
Advertisements

GlueX Collaboration Meeting, Newport News, October 25-27, 2007 GlueX Simulations Status Report Richard Jones GlueX Collaboration Meeting, Newport News,
DAQ Considerations For Increased Rep. Rate J. Leaver 01/12/2009.
Servers Redundant Array of Inexpensive Disks (RAID) –A group of hard disks is called a disk array FIGURE Server with redundant NICs.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
BLOCK DIAGRAM OF COMPUTER
Offline Software Status Jan. 30, 2009 David Lawrence JLab 1.
Data oriented job submission scheme for the PHENIX user analysis in CCJ Tomoaki Nakamura, Hideto En’yo, Takashi Ichihara, Yasushi Watanabe and Satoshi.
Online Update Elliott Wolin GlueX Collaboration Meeting 3-Jun-2013.
Trigger Database Dmitry Romanov. Run Control Super- visor Agent CODA ROC Write Configuration tag – run number Run Config DB CODA ROC Agent    Hardware.
Feb. 19, 2015 David Lawrence JLab Counting House Operations.
Hall D Trigger and Data Rates Elliott Wolin Hall D Electronics Review Jefferson Lab 23-Jul-2003.
DAQ Status & Plans GlueX Collaboration Meeting – Feb 21-23, 2013 Jefferson Lab Bryan Moffit/David Abbott.
Online Data Challenges David Lawrence, JLab Feb. 20, /20/14Online Data Challenges.
GLAST LAT ProjectDOE/NASA Baseline-Preliminary Design Review, January 8, 2002 K.Young 1 LAT Data Processing Facility Automatically process Level 0 data.
GlueX electronics Collaboration Meeting May, 2004 Paul Smith.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
Access Across Time: How the NAA Preserves Digital Records Andrew Wilson Assistant Director, Preservation.
HPS Online Software Discussion Jeremy McCormick, SLAC Status and Plans.
The GlueX Detector 5/29/091CIPANP The GlueX Detector -- David Lawrence (JLab) David Lawrence (JLab) Electron beam accelerator continuous-wave (1497MHz,
DAQ Status Graham. EMU / EB status EMU framework prototype is complete. Prototype read, process and send modules are complete. XML configuration mechanism.
DAQ Status Report GlueX Collaboration – Jan , 2009 – Jefferson Lab David Abbott (In lieu of Graham) GlueX Collaboration Meeting - Jan Jefferson.
Online Monitoring Status David Lawrence JLab Oct. 2, /2/14Monitoring Status -- David Lawrence1.
DAQ Issues for the 12 GeV Upgrade CODA 3. A Modest Proposal…  Replace aging technologies  Run Control  Tcl-Based DAQ components  mSQL  Hall D Requirements.
Data Acquisition for the 12 GeV Upgrade CODA 3. The good news…  There is a group dedicated to development and support of data acquisition at Jefferson.
David Abbott - JLAB DAQ group Embedded-Linux Readout Controllers (Hardware Evaluation)
4/5/2007Data handling and transfer in the LHCb experiment1 Data handling and transfer in the LHCb experiment RT NPSS Real Time 2007 FNAL - 4 th May 2007.
JANA and Raw Data David Lawrence, JLab Oct. 5, 2012.
A.Golunov, “Remote operational center for CMS in JINR ”, XXIII International Symposium on Nuclear Electronics and Computing, BULGARIA, VARNA, September,
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Clara Gaspar, March 2005 LHCb Online & the Conditions DB.
Computing Division Requests The following is a list of tasks about to be officially submitted to the Computing Division for requested support. D0 personnel.
Online Software 8-July-98 Commissioning Working Group DØ Workshop S. Fuess Objective: Define for you, the customers of the Online system, the products.
David Abbott - Jefferson Lab DAQ group Data Acquisition Development at JLAB.
Hall-D/GlueX Software Status 12 GeV Software Review III February 11[?], 2015 Mark Ito.
Oct. 8, 2015 David Lawrence JLab Counting House Operations.
The KLOE computing environment Nuclear Science Symposium Portland, Oregon, USA 20 October 2003 M. Moulson – INFN/Frascati for the KLOE Collaboration.
MOLLER DAQ Aug 2015 meeting Team : R. Michaels, P. M. King, M. Gericke, K. Kumar R. Michaels, MOLLER Meeting, Aug, 2015.
Online Reconstruction 1M.Ellis - CM th October 2008.
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
Online Monitoring for the CDF Run II Experiment T.Arisawa, D.Hirschbuehl, K.Ikado, K.Maeshima, H.Stadie, G.Veramendi, W.Wagner, H.Wenzel, M.Worcester MAR.
Linda R. Coney – 5 November 2009 Online Reconstruction Linda R. Coney 5 November 2009.
DoE Review January 1998 Online System WBS 1.5  One-page review  Accomplishments  System description  Progress  Status  Goals Outline Stu Fuess.
MICE CM28 Oct 2010Jean-Sebastien GraulichSlide 1 Detector DAQ o Achievements Since CM27 o DAQ Upgrade o CAM/DAQ integration o Online Software o Trigger.
Online Monitoring System at KLOE Alessandra Doria INFN - Napoli for the KLOE collaboration CHEP 2000 Padova, 7-11 February 2000 NAPOLI.
Activity related CMS RPC Hyunchul Kim Korea university Hyunchul Kim Korea university 26th. Sep th Korea-CMS collaboration meeting.
Monitoring Update David Lawrence, JLab Feb. 20, /20/14Online Monitoring Update -- David Lawrence1.
Oct HPS Collaboration Meeting Jeremy McCormick (SLAC) HPS Web 2.0 OR Web Apps and Databases (Oh My!) Jeremy McCormick (SLAC)
DAQ Status & Plans GlueX Collaboration Meeting – Feb 21-23, 2013 Jefferson Lab Bryan Moffit/David Abbott.
November 1, 2004 ElizabethGallas -- D0 Luminosity Db 1 D0 Luminosity Database: Checklist for Production Elizabeth Gallas Fermilab Computing Division /
CODA Graham Heyes Computer Center Director Data Acquisition Support group leader.
Event Management. EMU Graham Heyes April Overview Background Requirements Solution Status.
11/01/20081 Data simulator status CCRC’08 Preparatory Meeting Radu Stoica, CERN* 11 th January 2007 * On leave from IFIN-HH.
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
The ALICE data quality monitoring Barthélémy von Haller CERN PH/AID For the ALICE Collaboration.
January 2010 – GEO-ISC KickOff meeting Christian Gräf, AEI 10 m Prototype Team State-of-the-art digital control: Introducing LIGO CDS.
Online Software November 10, 2009 Infrastructure Overview Luciano Orsini, Roland Moser Invited Talk at SuperB ETD-Online Status Review.
October 19, 2010 David Lawrence JLab Oct. 19, 20101RootSpy -- CHEP10, Taipei -- David Lawrence, JLab Parallel Session 18: Software Engineering, Data Stores,
Super BigBite DAQ & Trigger Jens-Ole Hansen Hall A Collaboration Meeting 16 December 2009.
1 Calice TB Review DESY 15/6/06D.R. Ward David Ward Post mortem on May’06 DESY running. What’s still needed for DESY analysis? What’s needed for CERN data.
TRTViewer: the ATLAS TRT detector monitoring and diagnostics tool 4 th Workshop on Advanced Transition Radiation Detectors for Accelerator and Space Applications.
David Lawrence JLab May 11, /11/101Reconstruction Framework -- GlueX Collab. meeting -- D. Lawrence.
Compute and Storage For the Farm at Jlab
WP18, High-speed data recording Krzysztof Wrona, European XFEL
CLAS12 Ready for Science Answers to the Homework Questions & Summary/Path Forward Latifa Elouadrhiri Jefferson Lab CLAS12 Ready for Science Review.
Gluondb1 (online) Resides on gluon25, own IP address
Off-line & GRID Computing
Emanuele Leonardi PADME General Meeting - LNF January 2017
Example of DAQ Trigger issues for the SoLID experiment
Hall D Trigger and Data Rates
Presentation transcript:

May. 11, 2015 David Lawrence JLab Counting House Operations

Spring Commissioning Run Written to tape – 72 TB – 4297 EVIO files – 635 Runs (maybe TB of useful beam data for offline analysis)

Select runs runI solenoid I beam modestartkB/eventradiatorNevents A100nA :4128 kBJ1A5074.6M A45nA :2519 kB1 x M 31380A60-100nA85-1 4:49189 kB1 x M A70nA75-2 3:1733 kB1 x M A12nA :0728 kBJ1A5019.9M A10-20nA :5833 kBJ1A50*6.5M 32000A nA85-3 5:55224 kB1 x M * Run 3186 has ~13M events, but the magnet quenched about halfway through the run. The first 8 files (= 6.5M events) should be good.

12.4 kB/event I beam = 0 mode 7 BCAL/FCAL trigger 227 kB/event I beam = 180nA mode 8 FCAL/ST trigger run 3201 run 3205

Mar. 11, 2015 May. 4, 2015 Integrated disk usage for EVIO files only calculated from Nevents*avg. event size in datmon DB

32-bit word types in EVIO event

Run 2931 beam cosmics 74.6M events Diamond radiator Mode 7 (i.e. integrals only) Apparent beam burst while writing file 93 of 107 Data footprint changed after this… From EPICS archive via Myaviewer From EPICS events in EVIO files 120nA

Before incident After incident ALL fADC125 modules show same behavior CDC FDC

Run 3185: -Diamond radiator -1300A solenoid -10nA beam current -Mode M events Probably closest to GlueX production running conditions MONITORING FARM USAGE 100 CPU NETWORK

Well documented on wiki Some info automatically entered every run Most data from “Sean” DB inserted into RCDB DB in counting house is being replicated to DB outside of counting house Public web interface not quite live yet (Dmitry is still touching up some things) Work needed from others (e.g. me) to populate RCDB with things being written to other Run info. DB Run Conditions DB (RCDB)

“Sean” DB

“Dave” DB

“Dave” DB

Automatic E-log entries

CODA versions EPICS values for Beamline EPICS values for BCAL EPICS values for CDC EPICS values for FDC Trigger/DAQ config.

Improved Hall Operations Rewired console computers to UPS (thanks power outage!) Printer now works from CH computers RHEL7 tested with CSS JTabs issue identified and is being corrected (by accelerator) RAID Disk mapping and semiautomatic-deletion script tested Added BCAL_Eff, BCAL_inv_mass, HLDetectorTiming plugins to monitoring

Backup Slides Online Status -- David Lawrence20

Counting house computer systems Computer(s) processor General Purpose Network DAQ Network I.B. Network comments gluonfs1 N/A X ~1.6TB with snapshot backup gluonraid1-2 Intel E XXX RAID disk host ER process gluon01-05 X Shift taker consoles gluon20-23 AMD 2347 X Controls 8core gluon24-30 X Controls (gluon24 is web/DB/cMsg server) 12core + 12ht gluon40-43 AMD 6380 XXX 16core + 16”ht” gluon46-49 E XX (gluon47 &49) X 16core + 16ht gluon E XX 16core + 16ht rocdev1 Pentium X RHEL5 system for compiling ROLs for DAQ hdguest0-3 X (outside network) Guest consoles in cubicles (outside network) Online Status -- David Lawrence21

Run 2931 data words beam cosmics f125 block header f125 filler Places where red is higher are digitized values (others are headers, trailers, timestamps,etc…) 32-bit word types in EVIO event

Data Rates ROC Event Builder Event Recorder Tape Library ROC Spec: 100MB/sec Tested: ~30MB/sec Spec: 3000MB/sec Tested: 600MB/sec Spec: 300MB/sec Tested: 600MB/sec Spec: 300MB/sec Tested: 450MB/sec “Tested” means with actual data while it was being acquired. In some cases, offline testing has achieved significantly higher rates. 72TB x2 RAID disk Online Status -- David Lawrence23 (L3 farm) 125.9TB in 147,355 files written to tape in 2014 commissioning run

Mode 7 (fADC integrals) Mode 7 (full samples) fADC250 fADC125 fADC250 fADC250/F1TDCfADC125 Mode 7 (fADC integrals) Online Status -- David Lawrence24 FCAL BCAL FCAL BCAL FDC CDC fADC250/F1TDC

Mode 7 (fADC integrals) Mode 8 (fADC full samples) Online Status -- David Lawrence kB/event 69 kB/event

Online Monitoring Online Status -- David Lawrence26

Rough Specs. Review 10 8  /s on LH 2 target -> ~400kHz hadronic rate L1 trigger goal is to cut away ~50% leaving 200kHz L3 trigger goal is to reduce by ~90% leaving 20kHz Early simulation suggested ~15kB/event Design specs*: – 200 kHz = 3000 MB/s (front end) – L3 reduction by factor of 10 = 300MB/s to RAID disk – 3 days storage on RAID = 300MB/s*3days = 78TB – Maintain 300MB/s transfer from RAID to tape Online Status -- David Lawrence27 *L3 not officially part of 12GeV upgrade project

Mode 7 (fADC Integrals) Mode 8 (fADC full samples) Each 32bit word in the EVIO file tallied to identify what file space is being used for Comparison between mode 7 and mode 8 data made Example: some of the fADC250 word types Online Status -- David Lawrence28

Event Size Simulation was consistent with initial estimate of event size Actual data was more than x4 larger Much of the data was taken in “raw” mode where fADC samples were saved Online Status -- David Lawrence29

DAQ to Detector Translation Table The Translation Table is used to convert from DAQ system coordinates (rocid, slot, channel) into detector-specific coordinates (e.g. BCAL module, layer, sector, end) ~23k channels defined in SQLite DB file Stored in CCDB as XML string for offline analysis with complete history: – /Translation/DAQ2detector Online Status -- David Lawrence30

hdmon Monitoring Plugins Online Status -- David Lawrence31 BCAL_online CDC_online DAQ_online FCAL_onlineFDC_onlinePS_onlineST_onlineTAGH_onlineTAGM_onlineTOF_online Each detector system provides 1 or more plugins that create histograms for monitoring All plugins are attached to a Common DANA process (hdmon) A “rootspy” plugin publishes all histograms to the network rootspy

Raw Data Formatted Files (from simulated data) Online Status -- David Lawrence32 CCDB hdgeant_smeared.hddm run0002.evio (Data file in same format as will be produced by CODA DAQ system) roc002.evio roc003.evio roc004.evio......

L3 and monitoring architecture EB ER L3 and monitoring processes are decoupled. They could run on same nodes though if desired. gluon53gluonraid1 gluon46 (Data flows from left to right) farm manager Online Status -- David Lawrence33

hdmongui Online Status -- David Lawrence34 multiple “levels” supported processes run multi-threaded

Online Status -- David Lawrence35