RAC aware change In order to reduce the cluster contention, a new scheme for the insertion has been developed. In the new scheme: - each “client” receives.

Slides:



Advertisements
Similar presentations
ITEC474 INTRODUCTION.
Advertisements

Chapter 9. Performance Management Enterprise wide endeavor Research and ascertain all performance problems – not just DBMS Five factors influence DB performance.
Module 13: Performance Tuning. Overview Performance tuning methodologies Instance level Database level Application level Overview of tools and techniques.
The Architecture of Oracle
Database Architectures and the Web
INTRODUCTION TO ORACLE Lynnwood Brown System Managers LLC Oracle High Availability Solutions RAC and Standby Database Copyright System Managers LLC 2008.
Acknowledgments Byron Bush, Scott S. Hilpert and Lee, JeongKyu
Overview of Database Administrator (DBA) Tools
Oracle9i Database Administrator: Implementation and Administration 1 Chapter 2 Overview of Database Administrator (DBA) Tools.
INTRODUCTION TO ORACLE DATABASE ADMINISTRATION Lynnwood Brown System Managers LLC Introduction – Lecture 1 Copyright System Managers LLC 2007 all rights.
Cacti Workshop Tony Roman Agenda What is Cacti? The Origins of Cacti Large Installation Considerations Automation The Current.
Technical Architectures
1 - Oracle Server Architecture Overview
Harvard University Oracle Database Administration Session 2 System Level.
A Guide to Oracle9i1 Introduction to Oracle9i Database Administration Chapter 11.
Database System Architectures  Client-server Database System  Parallel Database System  Distributed Database System Wei Jiang.
Chapter 9 Overview  Reasons to monitor SQL Server  Performance Monitoring and Tuning  Tools for Monitoring SQL Server  Common Monitoring and Tuning.
Module 15: Monitoring. Overview Formulate requirements and identify resources to monitor in a database environment Types of monitoring that can be carried.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
Database System Concepts and Architecture Lecture # 3 22 June 2012 National University of Computer and Emerging Sciences.
CERN IT Department CH-1211 Geneva 23 Switzerland t Experience with NetApp at CERN IT/DB Giacomo Tenaglia on behalf of Eric Grancher Ruben.
CERN - IT Department CH-1211 Genève 23 Switzerland t The High Performance Archiver for the LHC Experiments Manuel Gonzalez Berges CERN, Geneva.
Chapter Oracle Server An Oracle Server consists of an Oracle database (stored data, control and log files.) The Server will support SQL to define.
Oracle on Windows Server Introduction to Oracle10g on Microsoft Windows Server.
Oracle10g RAC Service Architecture Overview of Real Application Cluster Ready Services, Nodeapps, and User Defined Services.
By Lecturer / Aisha Dawood 1.  You can control the number of dispatcher processes in the instance. Unlike the number of shared servers, the number of.
CSE 781 – DATABASE MANAGEMENT SYSTEMS Introduction To Oracle 10g Rajika Tandon.
March 19981© Dennis Adams Associates Tuning Oracle: Key Considerations Dennis Adams 25 March 1998.
1 Moshe Shadmon ScaleDB Scaling MySQL in the Cloud.
Oracle9i Performance Tuning Chapter 1 Performance Tuning Overview.
Update on Database Issues Peter Chochula DCS Workshop, June 21, 2004 Colmar.
1 RAC Internals Julian Dyke Independent Consultant Web Version juliandyke.com © 2007 Julian Dyke.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
Oracle Tuning Considerations. Agenda Why Tune ? Why Tune ? Ways to Improve Performance Ways to Improve Performance Hardware Hardware Software Software.
Oracle Tuning Ashok Kapur Hawkeye Technology, Inc.
Oracle 10g Database Administrator: Implementation and Administration Chapter 2 Tools and Architecture.
Oracle9i Performance Tuning Chapter 12 Tuning Tools.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Achieving Scalability, Performance and Availability on Linux with Oracle 9iR2-RAC Grant McAlister Senior Database Engineer Amazon.com Paper
Copyright ©2003 Digitask Consultants Inc., All rights reserved Cluster Concepts Digitask Seminar November 29, 1999 Digitask Consultants, Inc.
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
Process Architecture Process Architecture - A portion of a program that can run independently of and concurrently with other portions of the program. Some.
Mark E. Fuller Senior Principal Instructor Oracle University Oracle Corporation.
CERN IT Department CH-1211 Geneva 23 Switzerland t Oracle Tutorials CERN June 8 th, 2012 Performance Tuning.
VMware vSphere Configuration and Management v6
INTRODUCTION TO ORACLE DATABASE ADMINISTRATION Lynnwood Brown President System Managers LLC Introduction – Lecture 1 Copyright System Managers LLC 2003.
Infrastructure for Data Warehouses. Basics Of Data Access Data Store Machine Memory Buffer Memory Cache Data Store Buffer Bus Structure.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Database Competence Centre openlab Major Review Meeting nd February 2012 Maaike Limper Zbigniew Baranowski Luigi Gallerani Mariusz Piorkowski Anton.
Database Performance Eric Grancher - Nilo Segura Oracle Support Team IT/DES.
MND review. Main directions of work  Development and support of the Experiment Dashboard Applications - Data management monitoring - Job processing monitoring.
Oracle Architecture - Structure. Oracle Architecture - Structure The Oracle Server architecture 1. Structures are well-defined objects that store the.
GPFS: A Shared-Disk File System for Large Computing Clusters Frank Schmuck & Roger Haskin IBM Almaden Research Center.
Eric Grancher, Maaike Limper, CERN IT department Oracle at CERN Database technologies Image courtesy of Forschungszentrum.
Considerations for database servers Castor review – June 2006 Eric Grancher, Nilo Segura Chinchilla IT-DES.
Thomas Baus Senior Sales Consultant Oracle/SAP Global Technology Center Mail: Phone:
I NTRODUCTION OF W EEK 2  Assignment Discussion  Due this week:  1-1 (Exam Proctor): everyone including in TLC  1-2 (SQL Review): review SQL  Review.
CERN IT Department CH-1211 Genève 23 Switzerland t Load testing & benchmarks on Oracle RAC Romain Basset – IT PSS DP.
1 PVSS Oracle scalability Target = changes per second (tested with 160k) changes per client 5 nodes RAC NAS 3040, each with one.
The PVSS Oracle Archiver FW WG 6 th July Credits Many people involved IT/DES: Eric Grancher, Nilo Segura, Chris Lambert IT/PSS: Luca Canali ALICE:
How To Pass Oracle 1z0-060 Exam In First Attempt?
PVSS Evolution in Relation to Databases
Open Source distributed document DB for an enterprise
Oracle 11g Real Application Clusters Advanced Administration
Oracle Database Monitoring and beyond
Case studies – Atlas and PVSS Oracle archiver
Index Index.
Performance And Scalability In Oracle9i And SQL Server 2000
Presentation transcript:

RAC aware change In order to reduce the cluster contention, a new scheme for the insertion has been developed. In the new scheme: - each “client” receives a unique number - each of the tables is partitioned - each of the “clients” inserts into a separate partition. Thanks to these changes, the system achieved 150’000 events per second at peak. It has been observed that under heavy load the event loading can sometimes freeze. Space allocation A number of modifications have been made to pre-allocate the space (tablespace re-utilization as a background task). Thanks to these changes, the system now achieves a consistent insertion rate of 150’000 events per second. Introduction – Aim Oracle Real Application Cluster (RAC) is the Oracle cluster technology. It has been used at CERN since 1996 (it was called Oracle Parallel Server up to Oracle version 8.1). RAC is increasingly used in High Energy Physics. It is used at the LCG Tier1 sites and at CERN. It is a “shared-everything” clustered system where each of the clustered nodes directly accesses the storage via a Storage Area Network (typically Fibre Channel) or via Network Attached Storage (typically NFS over Ethernet). RAC provides a scalable solution in terms of high-availability and performance. The mechanism for the coordination of the data changes from the different cluster nodes is called “cache fusion”. Applications which are not “read-mostly” may not easily scale with the number of nodes in a RAC cluster. Siemens/ETM’s PVSS software is a key component of the supervisory layer of the control systems for the LHC accelerator and experiments. It is a SCADA system. One of its components is the module that archives the monitoring data events in a database. Current requirements are as high as 150’000 events per second. Methodology The methodology used for performance tuning is based on the Oracle wait interface. Before version 7, developers and database administrators have been working with ratios that provide a statistical view over some database key indicators (in memory / disk sorts, full table scans / indexed table access…). As of Oracle version 7, the database presents the instrumentation results (event 10046) through the wait interface. It has completely changed the methodology for performance tuning. In version 10, Oracle has introduced Active Session History (ASH) and Automatic Workload Repository (AWR). They provide automatic gatherings of session and instance- wide wait events aggregation. The methodology used for this project is based on an iterative AWR/ASH based tuning exercise, until the required performance was achieved. Acknowledgments We thank Manuel Gonzalez Berges, Piotr Golonka, Wayne Salter (CERN IT-CO), Chris Lambert, Nilo Segura Chinchilla (CERN IT-DES), Luca Canali (CERN IT-PSS), Svetozár Kapusta (ALICE experiment), Bjørn Engsig and Lothar Flatz (Oracle). Results Initial Initial tests have shown that “out of the box” performance was about 100 events per second archiving. The database was waiting on the clients “SQL*Net message from client”. Investigations have shown that: - a generic C++ database API was used (no bulk loading) - a trigger was launched after each event upload. API change The company providing the software made a set of radical changes following the initial tests and recommendations. They changed the API to use OCCI (Oracle C++ Call Interface) and after bulk loading a set of events into a temporary table, the software now makes a call to a PL/SQL package. Thanks to these changes, the system achieved 2’000 events per second. Schema change In order to reduce the I/O needs (“db file sequential read” is a wait event that is linked with I/O request), several changes have been tested and implemented: - index usage analysis and reduction - table structure change (move to Index Organised Table) - replacement of the merge by an insert - use of direct path insert (bypassing the buffer cache). Thanks to these changes, the system achieved 16’000 events per second. The top wait events indicate clearly that the system is slowed down because of cluster related wait events. Conclusions We have shown that nearly linear scalability can be obtained with RAC for some typical write-mostly workload. The tests have been done on a 6 node RAC cluster utilizing 32 SATA disks connected via Fibre Channel to the hosts. It requires an adequate data model and a database client interface capable of supporting bulk operations. Instrumentation in the database and in the application is necessary for achieving fine tuning results. We have used extensively the table partitioning feature to achieve such transaction rates. The iterative methodology for RAC scalability using the wait interface features provided in Oracle 10g has been key for the success of this work. References Shiers J 2006 The Evolution of Databases in HEP Computing in High Energy and Nuclear Physics (13-17 February, 2006) Dyke J and Shaw S 2006 Pro Oracle Database 10g RAC on Linux (Apress) Morle J 2006 Connection management in an Oracle RAC Configuration Lahiri T, Srihari V, Chan W, Macnaughton N and Chandrasekaran S 2001 Cache Fusion: Extending Shared-Disk Clusters with Shared Caches Proceedings of the 27th international Conference on Very Large Data Bases (September , 2001) Cache Fusion and the Global Cache Service Oracle9i Real Application Clusters Concepts Release 2 (9.2) Part Number A Figure 1. RAC logical view (shared disk, servers, clients). Figure 2. iterative methodology for performance tuning. Disks Database servers Clients (interactive, batch) 1. run the workload, gather ASH/AWR information 2. find the top event that slows down the processing 3. understand why time is spent on this event 4. modify client code database schema database code hardware configuration EventWaits Time (s) % Total Call Time Wait Class db file sequential read 29, User I/O enq: TX - contention Other CPU time Table 1. AWR report during a test after iteration 1. EventWaits Time (s) % Total Call Time Wait Class gc buffer busy27, Cluster CPU time gc current block busy 6, Cluster gc current grant busy 24, Cluster gc current block 2- way 118, Cluster Table 2. AWR report during a test after iteration 2. EventWaits Time (s) % Total Call Time Wait Class row cache lock Concurrency gc current multi block request 7, Cluster CPU time log file parallel write1, System I/O undo segment extension 785, Configuration events_ history eventlastval update eventlastval set … trigger: update eventlast … insert into events_history (…) values (…) events_ history 1 bulk insert into temp table PL/SQL: insert /*+ APPEND */ into events_history (…) partition PARTITION (1) select … from temp 3 2 x 150 Table 3. AWR report during a test after iteration 3. temp table instance 1instance 5instance 6 Global Cache Service 1. request 2. forward 3. copy the block 4. inform GCS Figure 3. data block has been changed on instance 5. This block is held in exclusive mode by Instance 5. Instance 1 attempts to modify the block. The data block is not written to disk, all communication is done over the cluster interconnect. Figure 4. description of the architecture changes. Test setup with a 6 nodes RAC and 150 clients (lxplus) each generating 1’000 changes/s iteration# Graph 1. Events per second loaded in the database x 6 CERN - IT Department CH-1211 Genève 23 Switzerland Oracle RAC (cluster) application scalability Eric Grancher and Anton Topurov CERN IT department / openlab ’000 events/s + stable rate 16’000 events/s 100 events/s 2’000 events/s 1000 Hz