ATLAS HLT and DAQ physical layout

Slides:



Advertisements
Similar presentations
Electrical distribution for ALICE experiment & Electronic Rack Control S. Philippin.
Advertisements

André Augustinus 16 June 2003 DCS Workshop Safety.
SGI Confidential Copan 400 Installation. SGI Confidential Installation Overview The approximate minimum time to complete this procedure is.
LECC V. Bobillier1 Overview of LHCb Electronics installation aspects V. Bobillier; L. Roy; J. Christiansen CERN - LHCb.
Presents The Silver Family An Integrated Approach to Processors, Data Communication and Head End Integration.
Supervision of Production Computers in ALICE Peter Chochula for the ALICE DCS team.
DCS aspects of TDAQ racks in USA15 and SDX1 Y. Ermoline et al. ATLAS Week, CERN, 15 February 2005.
Status of DCS for TDAQ racks Y. Ermoline et al. ATLAS TDAQ Week, CERN, 27 April 2005.
DCS LEB Workshop ‘98, Rome, Detector Control System, H.J.Burckhart,1 Detector Control System H.J Burckhart, CERN u Motivation and Scope u Detector and.
The Detector Safety System for LHC Experiments Stefan Lüders ― CERN EP/SFT & IT/CO CHEP03 ― UC San Diego ― March 27 th, 2003.
All content in this presentation is protected – © 2008 American Power Conversion Corporation Symmetra ® LX – redundant UPS High performance, redundant.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
SR1 ongoing workH.Pernegger21/2/02 Ongoing work in SR1 and preparations for detector macro assembly Heinz Pernegger.
SCADA FOR WATER DISTRIBUTION IC DEPT. GECGn SEC28.
Robert Gomez-Reino on behalf of PH-CMD CERN group.
Calo Piquet Training Session - Xvc1 ECS Overview Piquet Training Session Cuvée 2012 Xavier Vilasis.
5 March DCS Final Design Review: RPC detector The DCS system of the Atlas RPC detector V.Bocci, G.Chiodi, E. Petrolo, R.Vari, S.Veneziano INFN Roma.
High Sensitivity Smoke Detection
JPV, TB of Nov.15 th, ‘Rack Status‘ Jean-Pierre Vanuxem, PH-AIS.
Towards a Detector Control System for the ATLAS Pixeldetector Susanne Kersten, University of Wuppertal Pixel2002, Carmel September 2002 Overview of the.
JCOP Workshop September 8th 1999 H.J.Burckhart 1 ATLAS DCS Organization of Detector and Controls Architecture Connection to DAQ Front-end System Practical.
IMPLEMENTATION OF SOFTWARE INPUT OUTPUT CONTROLLERS FOR THE STAR EXPERIMENT J. M. Burns, M. Cherney*, J. Fujita* Creighton University, Department of Physics,
Clara Gaspar, October 2011 The LHCb Experiment Control System: On the path to full automation.
1 5 December 2012 Laurent Roy Infrastructure / Electronics Upgrade -Cabling (long distance) -Rack and Crate (space, electrical distribution, cooling) -Detector.
15/12/10FV 1 Common DCS hardware. General ideas Centralized system: o Common computers. Detector dedicated computers (with detector participation) only.
Status and plans for online installation LHCb Installation Review April, 12 th 2005 Niko Neufeld for the LHCb Online team.
1 Low Voltage Power Supply Specification DCS Workshop 8 Sept 03 L.Jirden.
1 Responsibilities & Planning DCS WS L.Jirdén.
LNL 1 SLOW CONTROLS FOR CMS DRIFT TUBE CHAMBERS M. Bellato, L. Castellani INFN Sezione di Padova.
Control in ATLAS TDAQ Dietrich Liko on behalf of the ATLAS TDAQ Group.
A water-cooling solution for PC-racks of the LHC experiments  The cooling concept  Rack Model  Cooling Test  Rack Performance  Failure Tests  Future.
André Augustinus 6 October 2005 Cabling, Interlocks, Racks Some remarks.
1 Outline Firmware upgrade of the HV_LED_DAC boards. HV Status Bits board. Status of the board integration into the LHCb TFC system. CALO HV system and.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
R. Fantechi. TDAQ commissioning Status report on Infrastructure at the experiment PC farm Run control Network …
The LHC PC Rack Project Fred Wickens On behalf of the LHC PC-Rack Study Group 10th LECC Workshop Boston Sep 2004.
CLAS12-RICH Mechanical Design Status-Report CLAS12 RICH Review September 5-6 th 2013 S. Tomassini, D. Orecchini1 D. Orecchini, S. Tomassini.
André Augustinus 16 June 2003 DCS Workshop General Purpose Monitor.
Source Controller software Ianos Schmidt The University of Iowa.
DCS meeting27.Sep.2002 Serguei Zelepoukine (IHEP, Protvino) Contents: 1. Front-end electronics: problems & solutions. 2. DCS cabling outside the ECAL detector.
Control, monitoring and safety aspects of electrical distribution in the Atlas experiment W.Iwanski PH/ESE-BE.
1 Calorimeters LED control LHCb CALO meeting Anatoli Konoplyannikov /ITEP/ Status of the calorimeters LV power supply and ECS control Status of.
ATLAS DCS ELMB PRR, CERN, March 2002Fernando Varela ELMB Networks CAN/CANopen Interoperability of the ELMB Usage of the ELMB in ATLAS ELMB Networks Full.
ATLAS TDAQ RoI Builder and the Level 2 Supervisor system R. E. Blair, J. Dawson, G. Drake, W. Haberichter, J. Schlereth, M. Abolins, Y. Ermoline, B. G.
60kW Thermosiphon control system
André Augustinus Infrastructure  Turbines  Radmon  Thermocouple readout  Netgear redundancy  Other cavern items  Cooling CRs  Other CR and surface.
1 13 Octobre 2011 Laurent Roy ATCA Crate / Infrastructure in place -Rack in D3: dimension -Electrical power availability -Tell1 racks running in D3 (air.
14 November 08ELACCO meeting1 Alice Detector Control System EST Fellow : Lionel Wallet, CERN Supervisor : Andre Augustinus, CERN Marie Curie Early Stage.
AFP Trigger DAQ and DCS Krzysztof Korcyl Institute of Nuclear Physics - Cracow on behalf of TDAQ and DCS subsystems.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Infrastructures and Installation of the Compact Muon Solenoid Data AcQuisition at CERN on behalf of the CMS DAQ group TWEPP 2007 Prague,
R. Krempaska, October, 2013 Wir schaffen Wissen – heute für morgen Controls Security at PSI Current Status R. Krempaska, A. Bertrand, C. Higgs, R. Kapeller,
A.Henriques (JCOV meeting 31/5/2001) ATLAS Calorimeter cooling project Goal: u Definition of a common frame work for the cooling of the ATLAS calorimeter.
ATCA COOLING PROJECT INTERSHIP FINAL PRESENTATION Piotr Koziol.
Update on Farm Monitor and Control Domenico Galli, Bologna RTTC meeting Genève, 14 april 2004.
ATLAS DCS ELMB PRR, March 4th 2002, H.J.Burckhart1 Embedded Local Monitor Board ELMB  Context  Aim  Requirements  Add-ons  Our aims of PRR.
E Ethernet C CAN bus P Profibus HV HV cables LV LV cables (+busbar) S Serial (RS232) Signal cable Other/Unknown Liquid or Gas Cable and/or Bus PCI-XYZ.
Marc R. StockmeierDCS-meeting, CERN DCS status ● DCS overview ● Implementation ● Examples – DCS board.
André Augustinus 13 June 2005 User Requirements, interlocks, cabling, racks, etc. Some remarks.
Status of integration in USA15. Status of rack layout 20/03/2015FTK Italia - A. Lanza2 Rack Y A2 equipped with turbine cooling system and nominal.
20OCT2009Calo Piquet Training Session - Xvc1 ECS Overview Piquet Training Session Cuvée 2009 Xavier Vilasis.
1 04 April 2007 L. Roy Detector Safety System for infrastructure - General fire detection (barracks, cavern) - Sniffer system (smoke detection above the.
CERN Summerschool 2007 Monitor and Control of the CANbus power supplies of ATLAS’ MDTs Wilco Koppert (UU)
FTK infrastructures and DCS 02/12/20101FTK Initial Design Review - A. Lanza  Crates  Single rack layout  Cabling  Rack placement  DSS  DCS.
Institute of Nuclear Physics Polish Academy of Sciences
How SCADA Systems Work?.
Electrical and electronic systems projects
Robert Gomez-Reino Garrido
TEC assembly and commissioning at Lyon
Presentation transcript:

ATLAS DAQ/HLT Infrastructure H.P. Beck, M. Dobson, Y. Ermoline, D. Francis, M. Joos, B. Martin, G. Unel, F. Wickens. Acknowledgements: H. Burckhart, W. Iwanski, K. Korcyl, M. Leahu 11th Workshop on Electronics for LHC and future Experiments, 12-16 September 2005, Heidelberg, Germany

ATLAS HLT and DAQ physical layout SDX1 ROS – Read Out System SV – Supervisor DFM – Data Flow Manager USA15 Read-Out Subsystems underground 152 ROSes (max 12 ROLs per ROS) HLT and EB processor farms on the surface Rack-mounted PCs and network switches ~2500 components in ~120 racks

HLT and DAQ operational infrastructure components Housing of HLT/DAQ system - 2 floor counting room in SDX1 building Metallic structure (electronics is heavy nowadays) Lighting, ventilation, mixed water piping and cable trays Power supply (from network and UPS) Housing and operation of rack-mounted PCs and network switches Rack selection and component mounting Power distribution and heat removal Power and network cabling Operational infrastructure monitoring Standard ATLAS DCS tools for rack parameter monitoring Linux and IPMI tools for individual PC internal monitoring Computing farm installation, operation and maintenance Rack configuration and cabling databases System administration

Counting room in SDX1 for DAQ/HLT system Size of barrack constrained by crane, shaft to USA15 and existing walls of SDX Housing 100 racks 600 mm x 1000 mm, up to 52U (height floor to ceiling ~2600 mm) Metallic structure initially designed for 500 kg/rack, re-designed for 800 kg/rack Air-conditioning removes ~10% of the heat dissipated, other 90% are removed via water-cooling Design Temporary power supply Construction Lighting and ventilation Cable trays Water piping

Housing of equipment “Standard” ATLAS Investigated 2 options 52U rack Modified “standard” 52U ATLAS rack for ROSes in USA15 positions already defined weight not an issue in USA15 uniform to other racks in USA15 Industry standard Server Rack for SDX1 (e.g. RITTAL TS8) bayed racks with partition panes for fire protection height and weight limits lighter racks, more flexible mounting for PCs Mounting of equipment Front mounted PC’s on supplied telescopic rails Rear/front mounted switches (on support angles if heavy) All cabling at rear inside the rack RITTAL TS8 47/52U rack

Cooling of equipment “Standard” ATLAS 52U rack Common horizontal air-flow cooling solution for ATLAS and RITTAL racks Outcome of joint “LHC PC Rack Cooling Project” Requirement: ~10 kW per rack A water-cooled heat exchanger fixed to the rear door of the rack CIAT cooler mounted on the rear door (+150 mm) 1800x300x150 mm3 Cooling capacity – 9.5 kW Water: in 15°C, out 19.1°C Air: in 33°C, out 21,5 °C 3 fan rotation sensors RITTAL 47U rack Water in/out

Powering of equipment in SDX1 (prototype rack study) Powering problems and (as simple as possible) solutions High inrush current – D-curve breakers and sequential powering Harmonic content – reinforced neutral conductor with breaker 11 kVA of apparent power delivered per rack on 3 phases, 16A each ~10 kW of real power available (and dissipated) with typical PFC of 0.9 3 individually controllable breakers with D-curve for high inrush current First level of sequencing 3 sequential power distribution units inside the rack (e.g. SEDAMU from PGEP) Second level of sequencing 4 groups of 4(5) outlets, max 5A per group, power-on is separated by 200 ms Investigated possibilities for individual power control via IPMI or Ethernet controlled power units Ventilation Auxiliary Power to equipment Power Control

Powering of equipment – proposed final implementation (1) Transformer (in front of SDX1) to power the switchboard Power from the switchboard delivered to 6 distributing cupboards on 2 levels Each cupboard controls and powers 1 row of racks (up to 18 racks) Two 3-phase cables under the false floor from the cupboard to each rack Switchboard Distributing cupboards Distribution cupboard 400 A breaker on the input 1 Twido PLC Read of ON/OFF and Electrical Fault status of all breakers in the row 16A breakers on the output One breaker per power cable Each breaker is individually turned ON by DCS via PLC

Powering of equipment – proposed final implementation (2) 3U distribution box on the bottom of each rack (front side) to provide distribution inside a rack 2 manual 3 phase switches on the front panel to cut power on 2 power lines Input and output cables on the rear panel Distribution from two 3 phase power lines to 6 single phase lines Flat cable distribution on the back side of the rack 6 cables from the distribution box to 6 flat cables 6-7 connectors on each flat cable PLC 2 x 16A D type 3 phase breakers 2 x 3 phase manual switches Rack Side Installation is designed to sustain 35 A of inrush current (peak) from 1 PC for ~35-40 PCs per rack

Monitoring of operational infrastructure SDX1 TDAQ computing room environment monitored by: CERN infrastructure services (electricity, C&V, safety) and ATLAS central DCS (room temperature, etc) Two complementary paths to monitor TDAQ rack parameters by available standard ATLAS DCS tools (sensors, ELMB, etc.) by PC itself (e.g. – lm_sensors) or farm management tools (e.g. – IPMI) What is monitored by DCS inside the racks: Air temperature – 3 sensors Inlet water temperature – 1 sensor Relative humidity – 1 sensor Cooler’s fan operation – 3 sensors What is NOT monitored by DCS inside the racks: Status of the rear door (open/closed) Water leak/condensation inside the cooler Smoke detection inside the rack Quartz TF 25 NTC 10 kOhm Precon HS-2000V RH

Standard ATLAS DCS tools for sensor readout The SCADA system (PVSS II) with OPC client / server The PC (Local Control Station) with Kvaser PCIcan-Q (4 ports) CAN power crate (16 branches of 32 ELMBs) and CANbus cable The ELMB, motherboards and sensor adapters Kvaser ELMB CAN power crate Motherboard

Sensors location and connection to ELMB Sensors location on the rear door of TDAQ rack All sensor signals (Temperature, Rotation, Humidity) and power lines are routed to a connector on the rear door to simplify assembly Flat cables connect these signals to 1 of 4 ELMB motherboard connectors 3 connectors receive signals from 3 racks, 1 spare connector for upgrades 1 ELMB may be used for 3 racks CANbus cables To next rack To PSU

Use of internal PC’s monitoring Most PC's now come with a hardware monitoring chips (e.g. LM78) Onboard voltages, fan status, CPU/chassis temperature, etc. a program, running on every TDAQ PC, may use lm_sensors package to access parameters and send this information using DIM to DCS PC IPMI (Intelligent Platform Management Interface) specification platform management standard by Intel, Dell, HP, and NEC a standard methodology for accessing and controlling bare-metal hardware, even without software installed or running based on a specialized micro-controller, or Baseboard Management Controller (BMC) - it is available even if system is powered down and no OS loaded http://it-dep-fio-ds.web.cern.ch/it-dep-fio-ds/presentations.asp Supermicro IPMI Card

Preparation for equipment installation Design drawing Rack content 6 Pre-Series racks Rack numbering (Y.03-04.D1)

Computing farm – Pre-Series installation Pre-Series components in CERN (few % of the final size) Fully functional, small scale, version of the complete HLT/DAQ Installed in SDX1 lower level and USA15 Effort for physical installation Highlight and solve procedural problems before we get involved in the much larger scale installation of the full implementation Will grow in time – 2006: + 14 racks, 2007: + 36 racks…

Cabling of equipment All cables are defined and labeled: 608 individual optical fibers from ROS to Patch Panels 12 bundles of 60 fibers from USA15 patch panels to SDX1 patch panels Individual cables from Patch Panels to the central switches and then to PCs Cables labeling is updated after installation Cable installation: tries to minimize cabling between racks, to keep cabling tidy and to conform to minimum bend radii not using cable arms but uncable a unit before removal

Computing farm – system management System Management of HLT/DAQ has been considered by SysAdmin Task Force, topics addressed: Users / Authentication Networking in general Booting / OS / Images Software / File Systems Farm Monitoring How to Switch on/off nodes Remote Access & Reset with IPMI IPMI daughter card for PC motherboard - experience with v1.5. which allows access via LAN; to reset, to turn off & on; to login as from the console; to monitor all sensors (fan, temperature, voltages etc.) Cold start procedure tested recently for Pre-Series: IPMI used to powerdown/boot/monitor machines scripts for IPMI operations (booting all EF nodes etc...) are being written

Computing farm – Point 1 system architecture Central Server 1 Central Server 2 Local Server 1 Local Server 2 Local Server n Client 1 Client n Service path Sync path Sync paths Alternative sync paths ATCN CPN Gateway (login necessary) CERN IT Bypass Users access CTN Clients are netbooted. CERN Public Network CERN Technical Network ATLAS Technical and Control Network

Computing farm – file servers/clients infrastructure Gateway & Firewall services to CERN Public Network ssh access Tree servers/clients structure Central File Server - 6 Local File Servers - ~70 clients all files come from a single (later a mirrored pair) Central File Server All clients are net-booted and configured from Local File Servers (PC’s & SBC’s) maximum of ~30 clients per boot-server to allow scaling up to 2500 nodes a top-down configuration (from machine specific / detector specific / function specific / node specific) provided when modified, the atlas software is (push) synced from top down to each node with disk software installation mechanism and responsibilities being discussed with Offline and TDAQ Node logs are collected on the local servers for post-mortem analysis if needed

Computing farm – Nagios farm management All farm management related functions are unified under a generic management tool Nagios (LFS and Clients): a unique tool to view the overall status, issue commands, etc… using IPMI where available otherwise ssh and lm_sensors mail notification for alarms (e,g, - temperature) DCS tools will be integrated

Conclusions: Good progress made in developing final infrastructure for the DAQ/HLT system power and cooling - has become a major challenge in computer centers installation monitoring and farm management Pre-series installation has provided invaluable experience to tune/correct the infrastructure, handling and operation Making good progress towards ATLAS operation in 2007 Looking forward to the start of physics running