IT-DB Physics Services Planning for LHC start-up

Slides:



Advertisements
Similar presentations
Introduction to DBA.
Advertisements

RLS Production Services Maria Girone PPARC-LCG, CERN LCG-POOL and IT-DB Physics Services 10 th GridPP Meeting, CERN, 3 rd June What is the RLS -
12. March 2003Bernd Panzer-Steindel, CERN/IT1 LCG Fabric status
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
D0 Run IIb Review 15-Jul-2004 Run IIb DAQ / Online status Stu Fuess Fermilab.
ASGC 1 ASGC Site Status 3D CERN. ASGC 2 Outlines Current activity Hardware and software specifications Configuration issues and experience.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
ATLAS Metrics for CCRC’08 Database Milestones WLCG CCRC'08 Post-Mortem Workshop CERN, Geneva, Switzerland June 12-13, 2008 Alexandre Vaniachine.
Workshop Summary (my impressions at least) Dirk Duellmann, CERN IT LCG Database Deployment & Persistency Workshop.
Roberto Divià, CERN/ALICE 1 CHEP 2009, Prague, March 2009 The ALICE Online Data Storage System Roberto Divià (CERN), Ulrich Fuchs (CERN), Irina Makhlyueva.
CERN Physics Database Services and Plans Maria Girone, CERN-IT
LFC Replication Tests LCG 3D Workshop Barbara Martelli.
CERN - IT Department CH-1211 Genève 23 Switzerland t Oracle Real Application Clusters (RAC) Techniques for implementing & running robust.
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
IDE disk servers at CERN Helge Meinhard / CERN-IT CERN OpenLab workshop 17 March 2003.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
CERN-IT/DB After-C5 Presentation EDMS Service Migration Monica Marinucci Lopez October 4th 2002 EDMS Document:
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Last update 29/01/ :01 LCG 1Maria Dimou- cern-it-gd Maria Dimou IT/GD CERN VOMS server deployment LCG Grid Deployment Board
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
LCG LHC Computing Grid Project From the Web to the Grid 23 September 2003 Jamie Shiers, Database Group IT Division, CERN, Geneva, Switzerland
BNL Oracle database services status and future plans Carlos Fernando Gamboa, John DeStefano, Dantong Yu Grid Group, RACF Facility Brookhaven National Lab,
CERN/IT/DB DB US Visit Oracle Visit August 20 – [ plus related news ]
Maria Girone CERN - IT Tier0 plans and security and backup policy proposals Maria Girone, CERN IT-PSS.
CNAF Database Service Barbara Martelli CNAF-INFN Elisabetta Vilucchi CNAF-INFN Simone Dalla Fina INFN-Padua.
LHC Logging Cluster Nilo Segura IT/DB. Agenda ● Hardware Components ● Software Components ● Transparent Application Failover ● Service definition.
The RAL Tier-1 and the 3D Deployment Andrew Sansum 3D Meeting 22 March 2006.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
A quick summary and some ideas for the 2005 work plan Dirk Düllmann, CERN IT More details at
Lead from the front Texas Nodal 1 TDWG Nodal Update – June 6, Texas Nodal Market Implementation Server.
Storage & Database Team Activity Report INFN CNAF,
Architecture of a platform for innovation and research Erik Deumens – University of Florida SC15 – Austin – Nov 17, 2015.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 6 April 2005.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Oracle Clustering and Replication Technologies UK Metadata Workshop - Oxford Barbara Martelli Gianluca Peco.
Dynamic Extension of the INFN Tier-1 on external resources
WLCG IPv6 deployment strategy
ALICE Computing Data Challenge VI
DCS Status and Amanda News
Maria Girone, CERN – IT, Data Management Group
(on behalf of the POOL team)
High Availability Linux (HA Linux)
LCG 3D Distributed Deployment of Databases
Database Services at CERN Status Update
Service Challenge 3 CERN
Elizabeth Gallas - Oxford ADC Weekly September 13, 2011
Database Readiness Workshop Intro & Goals
Castor services at the Tier-0
WLCG Service Interventions
Scalable Database Services for Physics: Oracle 10g RAC on Linux
Ákos Frohner EGEE'08 September 2008
The INFN Tier-1 Storage Implementation
Bernd Panzer-Steindel CERN/IT
Introduction of Week 6 Assignment Discussion
Oracle Storage Performance Studies
Grid Canada Testbed using HEP applications
ASM-based storage to scale out the Database Services for Physics
Database Services for CERN Deployment and Monitoring
Distributed computing deals with hardware
Scalable Database Services for Physics: Oracle 10g RAC on Linux
Deploying Production GRID Servers & Services
Presentation transcript:

IT-DB Physics Services Planning for LHC start-up Maria Girone, IT-DB 6 July 2004 http://cern.ch/db/

Introduction Summary of Current Situation Requirements for LHC and proposed solution

Physics Services – Today Replica Location Service (RLS) One WN as AS per VO (6 VO) for production service Two DS, hosting the DB, coupled via hot standby 2-3 spare nodes for failover Other parallel set-up for test One WN (AS) and one DS (DB) for POOL Two WN (AS) and one DS (DB) for IT/GD certification team >10 dedicated disk servers (COMPASS, HARP, Cristal II, CMS Tracker, …) Sun Cluster (public 2-node for applications related to detector construction, calibration and physics production processing) Run Oracle 9iR2 (DB) 9.0.4 (iAS) on RHEL 2.1

Dedicated Nodes (Disk Servers) COMPASS: 8 Disk Servers, +2/3 per year Typically only 1-2 DBs actively used – all others idle Good candidate for consolidation into a RAC HARP: 2 Disk Servers Cristal II: 1 Disk Server X5Tracker: 1 Disk Server CASTOR nameserver: 1 Disk Server PVSS: 1 Disk Server (foreseen) Basically everything that doesn’t ‘fit’ in Sun cluster (next…)

Sun Cluster 2 x Sun 280Rs (ex-Sundev) 360GB of SAN storage Roughly 100GB used for DB files… Hosts all physics applications with storage requirements << 100GB Production DBs (e.g. LHCb scanbook, CMS TMDB, …) Detector construction DBs Neither CPU capacity nor storage sufficient for LHC exploitation phase Current requests suggest that storage will be insufficient before end 2005! Needs to be replaced 2005 - 2006

Future Service Needs Increasing Service needs from the LCG side Grid File Catalog (high availability) CASTOR name server (high availability) Event-level Metadata Upcoming Conditions DB and POOL RDBMS production services Uncertain requirements on storage volume resource split among different services access I/O patterns

Investigating Oracle RAC / Linux RAC - Real Application Clusters provides High availability – proven technology used in production at CERN to avoid single points of failure with transparent application failover Consolidation – allows a smallish number (8? 16?) of database services to be consolidated into a single management entity Scalability – SAN-based infrastructure means CPU nodes and storage can be added to meet demand RAC (OPS) used on Sun since 1996 2 – 3 node clusters

SAN-based DB infrastructure Mid-range Linux PCs: dual power supply, mirrored system disk, with dual HBAs multiple GbitE (3) F/C connected switches: up to 64 ports (Qlogic) (Infortrend) Storage: 16 x 256GB disks

Short Term (July 2004-Aug 2004) Build prototype Linux RAC using existing h/w Farm nodes, Infortrend storage, Brocade/Qlogic switch etc Plan to use Oracle 10g with 9iR2 as fall-back Obtain offers for Qlogic Switches and HBAs Obtain offer(s) for ‘mid-range’ PCs (FIO) Order pre-production RAC h/w Define test cases for all target apps Castor n/s, RLS, COMPASS, …

Medium Term (Sep 2004 – Dec 2004) Build 2 RACs using shared SAN infrastructure to handle RLS and CASTOR production in 2005 Using same h/w, evaluate needs for COMPASS/HARP, Sun cluster replacement, POOL RDBMS b/e Understand deployment and manageability impact How many RACs do we need? Order additional h/w and plan migration of these services Extra PCs, HBAs, storage as required…

Long(-er) Term (mid-2006?) Scalable SAN-based infrastructure for all DB services for Physics Can add storage and / or CPU capacity as needed Can re-configure resources to meet demand Runs Oracle 10g R2? Announced December 2004; available H2 2005? Re-use same architecture for other services? AIS; IDB ?

Conclusions Today’s database services need to evolve to be ready for the LHC era Consolidation, manageability, high availability, scalability Investigating RAC as possible solution Linux based F/C Storage seems to be an affordable way to meet requirements IT-DB proposes a prototype phase using RAC for RLS and CASTOR for 2005 Data Challenges Plan to participate in the LCG Service Challenges beforehand All other services to follow later These services will be prototyped on Oracle 10g