Download presentation
Presentation is loading. Please wait.
Published byDominick Farmer Modified over 9 years ago
1
Peter Chochula
2
DAQ architecture and databases DCS architecture Databases in ALICE DCS Layout Interface to external systems Current status and experience Future plans
3
The presented slides are summary of 2 talks delivered to DATABASE FUTURES workshop No major changes of status or planning since the june workshop
4
ALICE is primary interested in ion collisions Focus on last weeks of LHC operation in 2011 (Pb-Pb collisions) During the year ALICE is being improved In parallel, ALICE participates in p-p programme So far, in 2011 ALICE delivered: 1000 hours of stable physics data taking 2.0 10 9 events collected 2.1 PB of data 5300 hours of stable cosmics datataking, calibration and technical runs 1.7 10 10 events 3.5 PB of data ▪ IONS STILL TO COME IN 2011! 4
5
2011 ions approaching First collisions already registered during LHC commissioning last weekend!
6
ALICE DAQ (slides taken from presentation of Sylvain Chapeland)
12
All DB services based on MySQL Managed by DAQ team Tools, GUI and API developed and maintained by ALICE DAQ team No major concerns Satisfactory performance and scalability Excellent stability over 3 years
13
ALICE DCS is responsible for safe and correct operation of the experiment DCS interacts with devices, configures them, monitors the operation and executes corrective actions There are about 1200 network attached devices and ~300 directly connected devices controlled by the DCS About 1 000 000 parameters are actively supervised by the system DCS interacts with many external systems Part of the acquired data is made available to offline for analysis (conditions data)
14
Detector Controls System DETECTORS and DETECTOR-like systems External Services and Systems Electricity Ventilation Cooling Gas Magnets Safety Access Control LHC Configuratio n Database Archival Database Devices SCADA 1000 ins/s Up to 6GB Infrastructure B-field Space Frame Beam Pipe Radiation Environment Alice Systems DAQ TRIGGER HLT Conditions Database ECS OFFLINE 14 SPDPHS FMDT00V00PMDMTRMCHZDCACO SDDSSDTPCTRDTOFHMP Controls Context AD0TRILHC
15
Detector Controls System Configuration Database Archival Database Devices SCADA 1000 ins/s Up to 6GB Conditions Database OFFLINE CONFIGURATION DATABASE: configuration of PVSS systems device settings front-end configuration Stored mostly as code which is compiled online and sent to devices at the start of a run
16
Detector Controls System Configuration Database Archival Database Devices SCADA 1000 ins/s Up to 6GB Conditions Database OFFLINE ARCHIVAL DATABASE: Parameters acquired from devices
17
Detector Controls System Configuration Database Archival Database Devices SCADA 1000 ins/s Up to 6GB Conditions Database OFFLINE CONDITIONS DATABASE Stores a subset of archived data Implemented at OFFLINE side Populated after each run with data acquired during the run
18
All PVSS systems (sitting on ALICE network) have direct access to all archived data using PVSS built-in interface Debugging Interactive analysis by shift crew and experts External and non-PVSS systems can access data only via dedicated client/server (AMANDA) suite Protection of archive Load balancing
19
AMANDA is a ALICE-grown client- server software used to access ALICE DCS data Client sends request for data, indicating names of the required parameters and requested time interval (without knowing the archive structure) Server retrieves the data and sends it back to the client Several AMANDA servers are deployed in ALICE Multiple requests are queued in the servers and processed sequentially In case of overload, it is enough to kill AMANDA server AMANDA servers operate across the secured network boundary AMANDA client AMANDA server DATA REQUEST SQL req. DATA Firewall
20
The main database service for the DCS is based on ORACLE The DBA tasks are provided by DSA section of the IT-DB, based on a SLA between ALICE and IT PRIMARY database servers and storage are located in ALICE pit STANDBY database and tape backups are located in IT
21
PRIMARY DATABASE - ALICE P2 STANDBY DATABASE - IT Storage Backup SAN DB Servers ~100 DB Clients: Cofiguration Archive Offline clients Backup Streaming DATABASE – IT (limited amount of data)
22
The DB is backed-up directly in ALICE site to a dedicated array Fast recovery Full backup The whole DB is mirrored on STANDBY database in IT The STANDBY database is backed up on tapes In case of DB connectivity problems, the clients can accumulate data in local buffers and dump them to DB once the connection is restored. Lifetime of local buffers is ~days
23
PRIMARY DATABASE - ALICE P2 STANDBY DATABASE - IT Storage Backup SAN DB Servers ~100 DB Clients: Cofiguration Archive Offline Backup Disaster scenario tested in 2010 All ALICE DCS redirected to standby database for several days SUCESS!!!
24
Number of clients: ~100 The ALICE DCS DB is tuned and tested for: steady insertion rate of ~1000 inserts/s peak rate of 150 000 inserts/s Current DB size: ~3TB 2-3 schemas/detector Served by 6 servers and 4 SAN arrays
25
ALICE DB service is in uninterrupted and stable operation since more than 3 years Initial problems caused by instabilities of RAID arrays solved by firmware upgrades Operational procedures fine-tuned, to match IT and ALICE requirements ▪ Updates only during LHC technical stops, etc.. The typical operational issues are caused by clients: Misconfigured smoothing (client overload) Missing data (stuck client, lost SOR/EOR signals) ▪ However, big improvements on stability during the last year (credits to EN-ICE guys)!
26
The smooth and stable operation of the ORACLE database for ALICE DCS is a big achievement Hereby we wish to express out thanks to the members of the IT-DB DS team for their highly professional help and approach!
27
There are some additional databases deployed in ALICE DCS, but their use is very light: MySQL – for bookkeeping on file exchange servers ▪ Zero maintenance, local backup solution SQL Server – as a storage for system monitoring tools (MS SCOM) ▪ Used as a out-of-the box solution, but growing quickly (will need to move to a new server) Private ORACLE server in the LAB for development
28
Currently the service fulfils ALICE needs No major architectural changes planned in the near future (before the long LHC shutdown) HW and SW upgrades still foreseen: A replacement of the DB service in ALICE counting room is prepared for this year winter shutdown ▪ Hardware (blade servers, SAN infrastructure and arrays) installed ▪ Software: 11G would be nice to have
29
No significant increase of data volume from detectors planned before the long LHC shutdown During the shutdown new detector modules will be added to ALICE. This might double the amount of data New project – the ALICE-LHC interface currently store data to files (luminosities, trigger counters, etc.) Aim to move to ORACLE Currently estimating the load – comparable with present DCS archival
30
We are getting more request for accessing the data from local analysis code Currently were are able to satisfy the needs with AMANDA and PVSS code, but reaching the limits Request for remote access (clients sitting on GPN or in institutes) Read only Mainly for debugging and monitoring purposes – latency is not an issue Possible scenarios: Currently we use streaming for a limited amount of data We could use the retired service or a light DB server to contain a browsable copy of the database 11g active dataguard seems to be a very attractive solution for ALICE DCS ▪ Basic requirement – PVSS compatibility
31
ALICE DAQ is based on MySQL databases managed by the DAQ team The main DB service in ALICE DCS is based on ORACLE managed by IT The operational experience is very positive (stability, reliability) on server side Small issues on clients side, being constantly improved No major modifications expected before the LHC long shutdown Several upgrades ongoing Again, thank to IT-DB experts for smooth operation of this critical service
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.