Online Update On behalf of the Online group. Introduction The main categories for the online schedule are based around the following items, so I will.

Slides:



Advertisements
Similar presentations
Computer Basics Hit List of Items to Talk About ● What and when to use left, right, middle, double and triple click? What and when to use left, right,
Advertisements

MIS 2000 Class 20 System Development Process Updated 2014.
MICE Target Electronics, DAQ and BPS MICE/ISIS Target Meeting 17 th September 2010 P J Smith – University of Sheffield.
MOM Report Ray Gamet MICE Operations Manager University of Liverpool MICO 18 th May 2009.
Software Summary Database Data Flow G4MICE Status & Plans Detector Reconstruction 1M.Ellis - CM24 - 3rd June 2009.
MOM Summary 19/01/10 Water Now the ice has cleared I am able to get on the roof again. Could do with getting the conductivity readings fed back similarly.
1 MICE PM Report General Update –Phase I civil engineering –R5.2 in June –Phase II design work –Phase II hardware –A few questions –Schedule, milestones.
Linda R. Coney – 24th April 2009 MOM Update Linda R. Coney 21 September, 2009.
Linda R. Coney – 24th September 2009 MOM Update End of Sept Run Linda R. Coney 05 October, 2009.
MICO 1 st February 2010 Terry Hart (MOM). MICO 1 st February June 2015 Decay Solenoid DS will not be ready for Machine Physics. Matt Hills spent.
MICE CM15 June 2006Jean-Sébastien GraulichSlide 1 DAQ Status o Detector DAQ Test bench in Geneva o Preparing for Test Beam o FE electronics tests o Detector.
MOM Report Ray Gamet MICE Operations Manager University of Liverpool MICO 11 th May 2009.
Status of Electronics & Control System P J Smith University of Sheffield 16/12/2009.
Online Reconstruction Update Linda R. Coney UCR Dec 17, 2009.
MICO Meeting, 1 st March 2010 Terry Hart (MOM) - Decay Solenoid - Targets - DAQ - March User Runs Plans - Suggestions.
Target Monitoring and Control Current status DAQ card – NI 6254 Analog/Digital card Problems with Linux version drivers Currently reading up to 12 channels.
1 Analysis code for KEK Test-Beam M. Ellis Daresbury Tracker Meeting 30 th August 2005.
MOM Report Paul Soler MICE Operations Manager University of Glasgow MICO May 5, 2009.
Spectrometer Solenoid Update Michael S. Zisman Center for Beam Physics Accelerator & Fusion Research Division Lawrence Berkeley National Laboratory MICO.
1 Timescales Construction LCR MICE Beam Monitoring counters + DAQ My understanding Jan 07 Feb 07 Mar 07 Apr 07 May 07 Jun 07 Jul 07 Aug 07 Sep 07 ISIS.
Online Servers and Schedule Paul J Smith. Online Servers Recent Changes: NUCS - > micethins -> there has been some discussion of changing the name – suggest.
Working with SharePoint Document Libraries. What are document libraries? Document libraries are collections of files that you can share with team members.
MICE CM26 March '10Jean-Sebastien GraulichSlide 1 Detector DAQ Issues o Achievements Since CM25 o DAQ System Upgrade o Luminosity Monitors o Sequels of.
CM26 March 2010Jean-Sebastien GraulichSlide 1 Online Summary o The heplnw17 case o DAQ o CAM o Online Reconstruction o Data Base o Data Storage Jean-Sebastien.
Software Summary 1M.Ellis - CM23 - Harbin - 16th January 2009  Four very good presentations that produced a lot of useful discussion: u Online Reconstruction.
1M. Ellis - MICE Video Conference - 15th March 2007 Software Report  New G4MICE Users  TOF Simulation and Digitisation  SciFi Reconstruction  Tracker.
Control and Monitoring System / EPICS Pete Owens Daresbury Laboratory.
L.E.A. Data Technologies L.E.A. Data Technologies Introduction.
 To explain the importance of software configuration management (CM)  To describe key CM activities namely CM planning, change management, version management.
1 Operations Summary L. Coney – UCR MICE CM34 – Oct 2012.
1 Operations Overview L. Coney UC Riverside MICE CM32 – Glasgow – June 2012.
Linda Coney MC VC May 19, 2011 MICE Operations Manager Report.
Offline Software Summary 1M.Ellis - CM25 - 6th November 2009  Release Schedule  Simulation and Analysis Update  Detector software: u CKOV u EMR u KL.
Control & Monitoring System Update Contributors: Brian Martlew - Daresbury Laboratory - STFC James Leaver - Imperial College Pierrick Hanlet - Fermilab.
DAQ MICO Report Online Monitoring: –Status All histograms are now implemented Still not fully online –Only monitoring from data file (slightly offline,
MICE Collaboration meeting (CM20) At RAL 10 – 13 Feb 2008 AFC Magnet – Status and Progress W. Lau, Oxford University.
CIDB The PSI Controls Inventory DataBase Timo Korhonen, PSI (for the CIDB Team)
OFFLINE TRIGGER MONITORING TDAQ Training 5 th November 2010 Ricardo Gonçalo On behalf of the Trigger Offline Monitoring Experts team.
Software Status  Last Software Workshop u Held at Fermilab just before Christmas. u Completed reconstruction testing: s MICE trackers and KEK tracker.
The Alternative Larry Moore. 5 Nodes and Variant Input File Sizes Hadoop Alternative.
Click to add Text L. CONEY1 MICE Shifter Training June 2015.
G4MICE Status and Plans 1M.Ellis - CM24 - RAL - 31st May 2009  Firstly, a correction to the agenda:  I failed to spot a mistake in the agenda that I.
Online Software 8-July-98 Commissioning Working Group DØ Workshop S. Fuess Objective: Define for you, the customers of the Online system, the products.
9: Troubleshooting Your Network
Virtual Classes Provides an Innovative App for Education that Stimulates Engagement and Sharing Content and Experiences in Office 365 MICROSOFT OFFICE.
Online Reconstruction 1M.Ellis - CM th October 2008.
MICE CM January 2009Jean-Sébastien GraulichSlide 1 Detector DAQ Issues o Achievements Since CM23 o Detector DAQ o Trigger o Online Software o Front End.
Ckov status David Lange, Doug Wright LLNL May 4, 2004.
Linda R. Coney – 5 November 2009 Online Reconstruction Linda R. Coney 5 November 2009.
Matthew Glenn AP2 Techno for Tanzania This presentation will cover the different utilities on a computer.
MICE Project Report Alan Bross (for Paul Drumm). Project Issues ● Key dates: – ISIS Synchrotron start-up scheduled for 1st August ● Shielded area around.
Tevatron BPM Upgrade Stephen Wolbers CD Accelerator Coordination Meeting September 21, 2004.
MICE CM28 Oct 2010Jean-Sebastien GraulichSlide 1 Detector DAQ o Achievements Since CM27 o DAQ Upgrade o CAM/DAQ integration o Online Software o Trigger.
Avalanche Motorsports
Electronics Spares for StepIV Operations MICE CM39 C.Macwaters 25/6/14.
11 th February 2008Brian Martlew EPICS for MICE Status of the MICE slow control system Brian Martlew STFC, Daresbury Laboratory.
Software Overview 1M. Ellis - CM21 - 7th June 2008  Simulation Status  Reconstruction Status  Unpacking Library  Tracker Data Format  Real Data (DATE)
Technical Board Summary Alan Bross MICE CM17 CERN February 25, 2007.
MICE Online Group Linda R. Coney UCR MICE CM31 UMiss October 2011.
Office 365 is cloud- based productivity, hosted by Microsoft. Business-class Gain large, 50GB mailboxes that can send messages up to 25MB in size,
Coney - CM36 - June Online Overview L. Coney – UCR MICE CM36 – June 2013.
MAUS Status A. Dobbs CM43 29 th October Contents MAUS Overview Infrastructure Geometry and CDB Detector Updates CKOV EMR KL TOF Tracker Global Tracking.
LIU Configuration Management EDMS Documentation, Layout and ECRs This presentation follows the LIU-PSB specific presentation done on 9 th October 2014.
Project Managers Report CM38 – Napa Roy Preece 24th February 2014.
L. Coney UC Riverside MICE CM32 – Glasgow – June 2012
MICE Computing and Software
MICE Collaboration Meeting Saturday 22nd October 2005 Malcolm Ellis
Project Manager’s Report + DC/MoM Report
CMS November Upgrade Week
SET USER GROUP MEETING May 17, 2017
Presentation transcript:

Online Update On behalf of the Online group

Introduction The main categories for the online schedule are based around the following items, so I will give a quick run through of how we are progressing: MLCR Upgrade Nagios and System Health Backups and Spares Documentation DAQ Online commissioning run Network Switches On-call Rotas Online Monitoring I’m not going to include lots of schedule detail – that was included in the parallel talk on Monday, which is available if you want to take a look. Apart from dropping the schedule items, this presentation is similar to the one given on Monday. My apologies to those in the online parallel! 2

MLCR Upgrade This is about 75% complete. The MLCR is now functional. More Space in the MLCR – the new desks are slightly narrower, overflow space outside of MLCR. Moved over to small local Intel NUC machines. These machines are not function specific so any machine can be used to run MICE operations from a menu. As the machines are local we have been able to utilise full HD displays so there is more display space – 2 monitors per NUC. What is left to do is a combination of a few things that are necessary and a few things that will be nice to have. Add the 4:3 monitors for the CCTV system and the Decay Solenoid status (brackets need raising). Install the larger wall monitors. (Alarm handler, electronic whiteboard) Finish installing KVM system. Touching up of overflow area. New(er) chairs. 3

MLCR Upgrade 4 4:3 Monitor brackets to be raised. Additional monitors still to be added. NUC’s to be tidied away behind monitors and wiring tidied up. KVM system to be completed (KVM interface will be where Pierrick is seated!)

Nagios and System Health A PC has been installed that performs a simple up/down check on all the other machines on MICENet. The functionality of this system can (and will) extend well beyond these simple checks. If individual have important processes running on machines or requirements such as processor utilisation/hdd space that needs to be monitored then please contact Paul. We will see if this can be incorporated into the Nagios system. Nagios needs to be integrated with EPICS alarm handler system – As I understand at the moment a good/not good signal will be sufficient. There is a longer term aim to add Cacti (Network monitoring) and Ganglia (Machine load monitoring) but I consider these less critical. (Nice to have) 5

Backups and Spares More detail on this was presented in the parallel session. The main message that requires presenting to the collaboration is: There is a mirrored raid (1TB) NAS in the MLCR where we can store backups as long as they are not too big. Currently 2/3 of this space is free. Note that there is no off- site storage for this kind of backup at present! Mathew Robinson has added instructions on how to back your critical data up to this drive on MICEmine. Note that the online group does not back up all of the machines/data/software but if you need help utilising this facility for critical data/files then please contact us. If you want to take advantage of backup facilities offered by the online group it is the system owner’s responsibility to let us know and to ensure that what we are offering is suitable. There may be other options. We (the online group) also need to take a closer look at the issue with backups to ensure our own critical files are suitably backed up. If MICE wants a full backup of every machine (probably not ideal) then significant additional infrastructure would be required. 6

Backups and Spares I have toured the MLCR with Craig and Henry and I have a list of what spare machines we are likely to need. I don’t think we are in too bad a position – The main conclusion is that we may need to buy a few additional servers, depending on whether we feel we can tolerate the risk (or not) of the current ones failing. A more comprehensive list detailing the additional machines that need to be purchased will be available soon. Spares on DAQ – Yordan has specified that we need one spare TDC and one spare digitiser. I understand that the funds for these items have been found. There are three additional DAQ boards that have issues, the impact of this has yet to be determined. 7

Documentation The documentation needs improving, I see several areas: 1) Details of Hardware 2) Details of Processes and Checks 3) Nagios 4) Online On-call An additional large portion of documentation will be required for the online on-call (4) but this will rely upon the technical documentation (1-3) being improved. The timing for the completion of the documentation is when we require online on- call experts to be trained. It will require some effort to sort the documentation out – but from a scheduling perspective this is not seen as a problem. 8

DAQ & DAQ/MICE Commissioning Run 7 new DAQ machines installed in the MLCR – OS and software installed. 5 are for active use, 2 are spares. 1 of the spares is earmarked as a tracker DAQ spare (it is in the tracker rack) and 1 is for general DAQ spare. There will be a period in November (16 th -26 th ) where the trackers will be tested. Most of the tracker DAQ group will be around at this time. “We will want to connect waveguides to the solenoid… As for detectors, the most invasive thing I'd like to do is inject light into TOF1 (to trigger) at the same time as firing the tracker LEDs to get a sense for the triggering delays. It's not clear to me whether direct light injection into the TOFs has ever been done, or whether it is permitted.” – David Adey So we may need to power up some of the PID. I (Paul) will try and use this time to become a bit more familiar with the DAQ. There is a final commissioning date of 21 st Jan where the online group needs to have all the machines ready. 9

10 Previous known puzzle with new trigger system – discussed at previous CM: Cut and pasted from Yordan’s parallel talk on Monday

11

Network Switches The old network stack is proving unreliable and there is limited support for the HW. An order for the new network stack has gone in. We are purchasing a HP 5406R zl2 stack from Hewlett Packard. The stack should arrive by the end of October – we will then schedule an installation date and plan the cabling details. “The warranty is free next business day replacement for the lifetime of the product. We tend to buy a spare and have it on the shelf for a quick replacement whilst raising a ticket for the free replacement of the spare.” – Quote from RAL Networks 12

On-Call Rota A draft document has been produced, but not yet published. this includes: 1)Introduction – definition of what the On-Call responsibilities are in general terms. 2) A list of responsibilities of the on-call team (this currently is a spreadsheet) 3) Staffing – How do we staff? – Share responsibility with other groups. Need for clear documentation and training. 4) Recruitment – Number of individuals required is not yet ascertained – requirement for some technical expertise. 5)Training - Defines the need to give enough time to write the documentation and train suitable and willing individuals. Chris Rogers has kindly offered to edit this document – required for the next MPB. 13

Online Monitoring 14 The Online Monitoring is working but there was a comment that users weren’t getting anything useful out of it. The system will be been changed to compare live plots with a set reference histograms Rhys Gardener is in the process of finding out exactly what needs monitoring and how it can be turned into something that gives timely and useful information to the individuals who are on shift. Rhys has commented that he can do some testing of the SW with empty plots i.e. the proposed January run - but ideally he requires beam to fully test the software. So the target activation run in Feb will be an ideal time to test this.

Conclusions Servers – There have been some additions to the server list with the MLCR upgrade, some changes with the DAQ machine upgrade, the Nagios (Test) Server and with the proposed changes to the way that the IOC’s are handled by Pierrick. Urgently I believe that we need to address: Backups – What is being backed up, is there enough backups, can we recover? Contact the online group if you have critical files that need backing up! The schedule is coming together. It’s kind of evolved over the last couple of months into something that is representative of what needs to be accomplished for the Step IV run. The documentation requires some attention. However there is always something more urgent to do! But it is scheduled as it is important Are there any unknown unknowns? The upcoming testing should help to answer this. I don’t believe that the schedule is resource limited – but there are a lot of things to do! 15