CM26 March 2010Jean-Sebastien GraulichSlide 1 Online Summary o The heplnw17 case o DAQ o CAM o Online Reconstruction o Data Base o Data Storage Jean-Sebastien Graulich, Geneva Software also discussed in the same session, not reported here
Online Activities Main Issue: Breakdown of heplnw17 Not discussed in Online session Collaboration Forum Revealed Lack of robustness, single point of failure Original misunderstanding: Private network Protected subnet Need for formally agreed support (from PPD or ISIS ?) Consequence 3 months of relative chaos (bad) Start working on a general computing and network requirement document (good) CM26 March 2010Jean-Sebastien GraulichSlide 2
CM26 March 2010Jean-Sebastien GraulichSlide 3 DAQ achievements DAQ system upgrade is ready Luminosity monitors integrated Trigger system cabling optimized DAQ and Trigger System Consolidation Cabling documentation in progress Progress in EMR front-end electronics
CM26 March 2010Jean-Sebastien GraulichSlide 4 Event Building The synchronization problem between the two crates persists We incriminate the PCI/VME interface It couldn’t be replaced because all the spares used for the mirror DAQ system Massive failure of boards: 4 out of 10 boards had to be send for repair In the meanwhile Online monitoring histogram allow to spot the problem A VME and PC power cycle solve it temporarily Shifter’s attention is required
CM26 March 2010Jean-Sebastien GraulichSlide 5 Schedule Milestones From CM25 CAM data in Online Data StreamNov 09 -> Mai 10 Tracker integrated in DAQ and OLMJan 10 -> July 10 TOF TDC Clock SynchronizationMarch 10 -> Aug 10 More complicated than first thought. Need a dedicated board Burst Gate Signal in the Trigger System Burst Gate Signal in the Trigger System Need support here The priority has been set to the DAQ and Trigger system upgrade and consolidation DAQ System upgradeMai 10 Production of SW/EMR Front End ElectronicsJan 10 -> Started
Control and Monitoring Outstanding Progress “Control is under control” Computer management ---> OK Software Management ---> OK Data Management ---> OK Documentation ---> OK All this sustained by mainly two individuals James Leaver and Pierrick Hanlet CM26 March 2010Jean-Sebastien GraulichSlide 6
CM26 March 2010Jean-Sebastien GraulichSlide 7
CM26 March 2010Jean-Sebastien GraulichSlide 8
CM26 March 2010Jean-Sebastien GraulichSlide 9
CM26 March 2010Jean-Sebastien GraulichSlide 10
CM26 March 2010Jean-Sebastien GraulichSlide 11
CM26 March 2010Jean-Sebastien GraulichSlide 12
CAM Decay solenoid included in the alarm handler Linde control panel mirrored into EPICS Very useful for expert remote monitoring Next: remote gateway and remote archive viewer new IOCs for new equipments The “Long, hard road to ramp up CaM infrastructure and knowledge base” has lead us to a point where we no longer foresee difficult hurdles to regularly add new IOCs, monitoring, alarm handling, and archiving… CM26 March 2010Jean-Sebastien GraulichSlide 13
CM26 March 2010Jean-Sebastien GraulichSlide 14
CM26 March 2010Jean-Sebastien GraulichSlide 15
CM26 March 2010Jean-Sebastien GraulichSlide 16
Data base What it does Store Configuration = Set values != read values Document hardware status - Geometry G4MICE - Cabling - Alarm Handler settings, etc Record automatically the magnet settings, ‘ISIS settings’, target information and DAQ information Superset of what is currently entered manually into the run configuration spreadsheet on the MICO page Allow retrieving these settings at the start of the run Also allow saving settings not attached to a run E.g. Pion at 300 MeV/c EPICS client developed by James Leaver for this CM26 March 2010Jean-Sebastien GraulichSlide 17
Data base status Progress was suspended in January due to failure of heplnw17 Local copy of DB system under development in Glasgow, progress resumed The main server functionality requested has now been implemented (except cabling) Proper migration to Rutherford Lab scheduled the bulk of outstanding work See David Forrest’s talk for details CM26 March 2010Jean-Sebastien GraulichSlide 18
Data Storage The only formally-agreed route for access to data (DAQ output) is via the Grid. The Grid Transfer Box (miceacq05) is located in the MLCR. It will eventually run an autonomous agent that reads the data from the RAID system in the MLCR and uploads it to the Grid, in particular the CASTOR tape system at RAL In the meantime data IS being uploaded to the Grid, but on a manual, next-day timescale. CM26 March 2010Jean-Sebastien GraulichSlide 19
Data Access Henry Nebrensky presented a tutorial on how to access the data using the grid Open Issues Permanent storage is on Tape at RAL Long access time (Robot loading the tape) We should foresee a place where actively used data is stored on disk Files on tape must be at least 200 MB… Eventually, someone on duty (MOM or shifter) will need to have a Grid certificate Once again we have a single (human) point of failure here CM26 March 2010Jean-Sebastien GraulichSlide 20
General Comment The MOG still suffers for a loose leadership Compensated by the enthusiasm and commitment of the individuals inside the group Most members are on short term contracts Linda and Pierrick depend on NFS grant David has to write his Ph.D. James will leave in January 2011 I’ll leave on June 2011 CM26 March 2010Jean-Sebastien GraulichSlide 21