Online SW and Database Status

Slides:



Advertisements
Similar presentations
Backup and Recovery Part 1.
Advertisements

PPOUG, 05-OCT-01 Agenda RMAN Architecture Why Use RMAN? Implementation Decisions RMAN Oracle9i New Features.
May 14, 2001E. Gallas/Trigger Database1 Status of the Trigger Database Elizabeth Gallas, Rich Wellner, Vicky White Fermilab - Computing Division See my.
Imperial College Tracker Slow Control & Monitoring.
HPS Online Software Discussion Jeremy McCormick, SLAC Status and Plans.
2nd September Richard Hawkings / Paul Laycock Conditions data handling in FDR2c  Tag hierarchies set up (largely by Paul) and communicated in advance.
The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within.
Gnam Monitoring Overview M. Della Pietra, D. della Volpe (Napoli), A. Di Girolamo (Roma1), R. Ferrari, G. Gaudio, W. Vandelli (Pavia) D. Salvatore, P.
1 Control Software (CAT) Introduction USB Interface implementation Calorimeter Electronics Upgrade Meeting Frédéric Machefert Wednesday 5 th May, 2010.
Integrating EPICS and LabVIEW on Windows using DCOM Freddie Akeroyd ISIS Computing Group.
Multistep Runs with ROD Crate DAQ Murrough Landon, QMUL Outline: Overview Implementation Comparison with existing setup Readout Status ModuleServices API.
Ideas about Tests and Sequencing C.N.P.Gee Rutherford Appleton Laboratory 3rd March 2001.
SL1Calo Input Signal-Handling Requirements Joint Calorimeter – L1 Trigger Workshop November 2008 Norman Gee.
Why A Software Review? Now have experience of real data and first major analysis results –What have we learned? –How should that change what we do next.
MICE CM28 Oct 2010Jean-Sebastien GraulichSlide 1 Detector DAQ o Achievements Since CM27 o DAQ Upgrade o CAM/DAQ integration o Online Software o Trigger.
TDAQ Experience in the BNL Liquid Argon Calorimeter Test Facility Denis Oliveira Damazio (BNL), George Redlinger (BNL).
FTS monitoring work WLCG service reliability workshop November 2007 Alexander Uzhinskiy Andrey Nechaevskiy.
October Test Beam DAQ. Framework sketch Only DAQs subprograms works during spills Each subprogram produces an output each spill Each dependant subprogram.
L1Calo Installation Status Murrough Landon, QMUL Level-1 Calorimeter Trigger (University of Birmingham, University of Heidelberg, University of Mainz,
The DCS Databases Peter Chochula. 31/05/2005Peter Chochula 2 Outline PVSS basics (boring topic but useful if one wants to understand the DCS data flow)
1 Tracker Software Status M. Ellis MICE Collaboration Meeting 27 th June 2005.
MAUS Status A. Dobbs CM43 29 th October Contents MAUS Overview Infrastructure Geometry and CDB Detector Updates CKOV EMR KL TOF Tracker Global Tracking.
ATLAS Detector Resources & Lumi Blocks Enrico & Nicoletta.
L1Calo Databases ● Overview ● Trigger Configuration DB ● L1Calo OKS Database ● L1Calo COOL Database ● ACE Murrough Landon 16 June 2008.
Schedule ● External Links ● Calorimeter Signals ● Internal Tests ● System Tests ● Combined Runs ● Schedule Murrough Landon 19 April 2007 A brief update.
CT-PPS DB Info (Preliminary) DB design will be the same as currently used for CMS Pixels, HCAL, GEM, HGCAL databases DB is Oracle based A DB for a sub-detector.
ICAICT201A USE COMPUTER OPERATING SYSTEM. USING THE CONTROL PANEL The Control Panel contains many options for configuring your computer, including: adding.
L1Calo DBs: Status and Plans ● Overview of L1Calo databases ● Present status ● Plans Murrough Landon 20 November 2006.
DB and Information Flow Issues ● Selecting types of run ● L1Calo databases ● Archiving run parameters ● Tools Murrough Landon 28 April 2009.
Online Software Status ● Overview ● Recent Changes ● Required Changes ● Desirable Changes ● Upgrade ● Summary Murrough Landon 1 July 2009.
L1Calo Databases ● Overview ● Recent Activitity ● To Do List ● Run types Murrough Landon 4 June 2007.
Online Database Developments ● Overview ● OKS database status and plans ● COOL database developments ● Validating calibrations ● Tools ● Summary Murrough.
M4 Operations ● Operational model for M4 ● Shifts and Experts ● Documentation and Checklists ● Control Room(s) ● AOB Murrough Landon 24 July 2007.
Calibration Operations ● Recent Changes ● Required Changes ● Calibration by Shifters Murrough Landon 23 March 2011.
Rainer Stamen, Norman Gee
Online Database Work Overview Work needed for OKS database
How to Contribute to System Testing and Extract Results
Essentials of UrbanCode Deploy v6.1 QQ147
Advanced Programing practices
User-Written Functions
Calibration Status Energy Calibration Stabilities/instabilities
CMS High Level Trigger Configuration Management
Run Control (and Other) Work
Configuration Database
SRM2 Migration Strategy
Online Database Status
EUDAQ Status Report Emlyn Corrin, 29 September 2010
AMI – Status November Solveig Albrand Jerome Fulachier
Generator Services planning meeting
MiniDAQ2 Workshop Control System.
Online Software Status
LCGAA nightlies infrastructure
Conditions Database Current conditions folders Proposed new folders
Online Software Status
Recent Online SW/DB Changes
Online SW Readiness (or not)
Online SW, DB & Calibration
Online Software Status
Remaining Online SW Tasks
Level 1 (Calo) Databases
Test-rigs outside CERN
Cabling Lengths and Timing
Online Software “To Do” List
Level-1 Calo Monitoring
CPM plans: the short, the medium and the long
Tracker Software Status
Publishing image services in ArcGIS
Banafsheh Hajinasab Based on presentation by K. Strnisa, Cosylab
TOS Control – main idea and scheme
Map Reduce, Types, Formats and Features
Presentation transcript:

Online SW and Database Status Murrough Landon 5 November 2008 Overview Tools Database Run Control and Module Services Usual Suspects To do list for the shutdown

Overview Main changes since the last joint meeting actually the basic software has been fairly stable moved to tdaq-01-09-01 (for collisions!) major update of COOL database schema, mainly for PPM added rate metering for PPM and CMM see talk by Felix more recently also reading PPM histograms various improvements to monitoring see talk by Taylor further development of calibration procedures, mainly CPM see talk by Joseph many little changes ROD related SW, hardware byteswapping, etc

Tools (1) Mapping Tool added display of rates from IS recently also statistics from PPM rate histograms l1rates tabular summary of many level 1 rates no histograms, but faster and more detail than trigger presenter ACE last batch of improvements from Alvin better time range selection, handling of different types will have less support in future, alas

Tools (2) l1chuck.py still a work in progress selective interface to DB for killing hot towers update DB and allow on the fly changes (making a new lumi block) latter not yet implemented present version updates both PpmDeadChannels and PprChanCalib folders better not to change LUT noise cut from calibration but add override value?

Tools (3) Readout monitoring rodmon application previously used to compare simulation and hardware in playback tests recently extended to provide monitoring of ROD fragments publishing summaries of errors to IS complementary to other monitoring producing histograms should we envisage combining this IS info with other ROD information, making a better readout diagnostic panel? eg combine/revamp/improve the L1Calo Busy panel

COOL Database (1) PPM schema changed after summer discussions keep DAC scan results as floats online calculation of correct DAC for expected pedestal but store other information as precalculated integers ensures offline has identical values to those used online now have separate folder for dead/killed PPM inputs those not due to internal L1Calo calibration problems still bringing this slowly into use already realise need for modest further changes noise cut from pedestal run override via dead channels folder NB need stable schema before moving to use the production COOL database! though we could try using non-production Oracle DB

COOL Database (2) Various other schema updates extra run parameters attributes for some scans bypass PPM LUT in playback (and set trivial FIR coefficients) disable module status updates during CAM scans common calibration error code attribute implemented similar code for dead/killed channels (Not much) Validation but no further work done on validation of calibrations and still need mechanism for tracking module IDs nor of code to extract/plot time history of values

OKS Database Recent L1Calo schema changes removed old OKS calibration cleaned up OKS run parameters added defined sets of TTCvi/LTP configurations selected by user via the L1Calo RunPars IGUI panel Forthcoming TDAQ changes “Check in” scheme for OKS database at point 1 No changes required outside point 1 It will no longer be possible just to write or edit new or updated files to the right directories - they will have to be checked in via a new CVS-like application If valid, they will then be made available on a readonly filesystem (only writable by a OKS server)

Module Services and Run Control Run controllers now publish “full statistics” used for rates (every 2s) and PPM histograms (less often) additional objects for statistics of PPM histograms responsibility in several packages: statsL1Calo: containers for IS variables and ROOT histograms module services: fills information using statsL1Calo API rcL1Calo: takes care of publishing the results On-the-fly commands To support on the fly changes of noise cuts, disabling of hot channels (anything else?) Support for set of known commands in run controller Passed to module services for implementation External SW (scripts) needs to take care of changing lumi block

All the Usual Stuff! We finally moved to hardware byteswapping! And have started talking about enabling bus errors... But we have not written much more documentation TDAQ versions currently using tdaq-01-09-01 (01-09-00 now unusable due to improvements in monitoring tdaq-02-00-00 is imminent (and 02-00-01 in February) new version has yet more changes to the IGUI API this time with more consultation than when tdaq-01-09-00 hit us

To Do List (1) Run control states reorganisation? We are currently using “substates”, keeping the ancient “Load” and “Configure” steps as our internal states TDAQ implementation of substates has some problems cannot restart our low level controllers, only top level L1Calo error states are not properly recovered (needs TDAQ rewrite!) Would like to reorganise our module services code a bit so we can use single “Configure” state to do all configuration of module (apart from its connections) Use “Connect” transition for connections between modules also for loading generated test vectors from the simulation? Good time to review, clean up and optimise (if possible) Need to check if there are really any “gotchas” Will also need time to test carefully

To Do List (2) Build/development environment Fix dependencies in CMT requirements files so we can do parallel makes (hope to improve elapsed time for builds) Should also make it easy to check out and build only one or two packages (not really the current model) Proper patch procedure for point 1 Original idea of tagged builds for point 1 completely undermined by little fixes and library rebuilds useful, but means we never really know what SW we are running Aim to make patches only from new tags Disallow changes to header files (requires complete new build)

To Do List (3) Errors and ERS The full ATLAS partition generates a huge amount of error messages from ERS But L1Calo is probably too quiet Some important problems only get reported in verbose log files which may never be noticed We should review all this and ensure we do issue suitable error messages via ERS (so visible to the IGUI) for real problems Long term intention has also been to convert HDMC and module service exceptions to the ERS exception framework Bus errors We still ignore bus errors: should try reporting them!

To Do List (4) Automatic disabling of LAr/Tile inputs to L1Calo The present L1Calo IGUI CaloMask panel allows individual LAr or Tile regions to be disabled from the trigger But this must be done by hand (after the partition is booted) After recent discussions with a few interested parties we propose to automatically disable LAr or Tile trigger inputs if their corresponding TTC partitions are disabled Scheme uses resources and resource sets (needs some L1Calo SW) http://indico.cern.ch/conferenceDisplay.py?confId=43097 Automatic disabling of RoIB inputs The same discussions also proposed a similar automatic scheme to disable RoIB inputs if eg L1Calo is disabled Implementation requires SW work on the RoIB side BUSY issues need RoIB firmware work (no effort available)

To Do List (5) Conditions database So far we only read data from COOL We also need to write conditions data for each run IS parameters (still not done although fairly easy) Module IDs eg at start of run Timestamp when calibration data was read This may be earlier than the run start and there might be database updates made between the configure step and start of run Archiving Longer term archive for rates Use of Monitoring Data Archive (MDA) for calibrations Application parameters Take more application parameters from DB instead of hard coding them...

PPM Calibration Attributes in COOL Common ModuleId [UInt32] ErrorCode [UInt32] *ResultsTimeStamp [UInt63] DAC/Pedestal DacOffset [Double] DacSlope [Double] DacCrossCor [Double] PedValue [UInt32] PedNoiseCut [UInt32] PedMean [Double] PedSigma [Double] PedNumEvents [UInt32] Coarse Timing (R/O Ptr) SyncDelayData [Uchar] SyncDelayBcid [UChar] Fine Timing (PHOS4) InDataNegedge [UChar] Phos4Delay [UChar] BCID InBcidNegedge [Uchar] ExtBcidThreshold [Uint16] SatBcidThreshLow [Uint16] SatBcidThreshHigh [Uint16] SatBcidLevel [Uint16] BcidEnergyRangeLow [Uint16] BcidEnergyRangeHigh [Uint16] FirCoeff* [UChar] FirStartBit [Uchar] LUT LutOffset [UInt16] LutSlope [Uint16] LutNoiseCut [Uint16] LutStrategy [UChar] LutParameters [BLOB] Notes In *Results folder only