Database Performance Measurement

Slides:



Advertisements
Similar presentations
Database Architectures and the Web
Advertisements

Database Management3-1 L3 Database Management Santa R. Susarapu Ph.D. Student Virginia Commonwealth University.
C-Store: Data Management in the Cloud Jianlin Feng School of Software SUN YAT-SEN UNIVERSITY Jun 5, 2009.
1 Chapter 2 Database Environment Transparencies © Pearson Education Limited 1995, 2005.
20 Spatial Queries for an Astronomer's Bench (mark) María Nieto-Santisteban 1 Tobias Scholl 2 Alexander Szalay 1 Alfons Kemper 2 1. The Johns Hopkins University,
Chapter 3 Database Management
Distributed DBMSs A distributed database is a single logical database that is physically distributed to computers on a network. Homogeneous DDBMS has the.
Organizing Data & Information
Chapter 2 Database Environment Pearson Education © 2014.
Chapter 1 An Overview of Database Management. 1-2 Topics in this Chapter What is a Database System? What is a Database? Why Database? Data Independence.
Object-Oriented Methods: Database Technology An introduction.
Using Standard Industry Benchmarks Chapter 7 CSE807.
Design Considerations CS2312. Conceptual Design includes Operational Use Mini World Requirements collection & analysis Conceptual design Data model design.
Database Environment 1.  Purpose of three-level database architecture.  Contents of external, conceptual, and internal levels.  Purpose of external/conceptual.
Database Architectures and the Web Session 5
Memory/Storage Architecture Lab Computer Architecture Performance.
OnLine Analytical Processing (OLAP)
Data Warehousing 1 Lecture-24 Need for Speed: Parallelism Virtual University of Pakistan Ahsan Abdullah Assoc. Prof. & Head Center for Agro-Informatics.
Session-8 Data Management for Decision Support
Data Warehouse Fundamentals Rabie A. Ramadan, PhD 2.
April 26, CSE8380 Parallel and Distributed Processing Presentation Hong Yue Department of Computer Science & Engineering Southern Methodist University.
Database Architectures Database System Architectures Considerations – Data storage: Where do the data and DBMS reside? – Processing: Where.
10 1 Chapter 10 Distributed Database Management Systems Database Systems: Design, Implementation, and Management, Sixth Edition, Rob and Coronel.
Distributed DBMSs- Concept and Design Jing Luo CS 157B Dr. Lee Fall, 2003.
1 Seoul National University Performance. 2 Performance Example Seoul National University Sonata Boeing 727 Speed 100 km/h 1000km/h Seoul to Pusan 10 hours.
Bayu Adhi Tama, M.T.I 1 © Pearson Education Limited 1995, 2005.
Introduction to the new mainframe © Copyright IBM Corp., All rights reserved. 1 Main Frame Computing Objectives Explain why data resides on mainframe.
© 2003 Prentice Hall, Inc.3-1 Chapter 3 Database Management Information Systems Today Leonard Jessup and Joseph Valacich.
Chapter 2 Database Environment.
IT 5433 LM1. Learning Objectives Understand key terms in database Explain file processing systems List parts of a database environment Explain types of.
Distributed Systems Architectures Chapter 12. Objectives  To explain the advantages and disadvantages of different distributed systems architectures.
Data Mining and Data Warehousing: Concepts and Techniques What is a Data Warehouse? Data Warehouse vs. other systems, OLTP vs. OLAP Conceptual Modeling.
Client/Server Technology
An Intro. to Database Benchmarks
Applied Operating System Concepts
Introduction to Computing Lecture # 13
Data warehouse.
TPC Benchmarks: TPC-A and TPC-B
Operational & Analytical Database
MVC and other n-tier Architectures
The Client/Server Database Environment
Physical Database Design for Relational Databases Step 3 – Step 8
Database Architectures and the Web
Data Warehouse.
Database Performance Tuning and Query Optimization
CSCI/CMPE 3334 Systems Programming
Chapter 16 Designing Distributed and Internet Systems
Programmable Logic Controllers (PLCs) An Overview.
Chapter 2 Database Environment Pearson Education © 2009.
Database Management System (DBMS)
Basic Concepts in Data Management
03 - Database Design, UML and (Extended) Entity Relationship Modeling
Introduction to Databases Transparencies
Lecture 1: Multi-tier Architecture Overview
Database Environment Transparencies
CSE8380 Parallel and Distributed Processing Presentation
MORE ON ARCHITECTURES The main reasons for using an architecture are maintainability and performance. We want to structure the software into reasonably.
Chapter 11 Database Performance Tuning and Query Optimization
Why Threads Are A Bad Idea (for most purposes)
Data Warehousing Concepts
Chapter 3 Database Management
Performance And Scalability In Oracle9i And SQL Server 2000
Why Threads Are A Bad Idea (for most purposes)
Information Retrieval and Web Design
Why Threads Are A Bad Idea (for most purposes)
Analytics, BI & Data Integration
DSS Architecture MBA 572 Craig K. Tyran Fall 2002.
Outline Context for database development Goals of database development
Data Warehouse and OLAP Technology
Presentation transcript:

Database Performance Measurement Spring 2002 Sang Hyup Lee School of Computing, Soongsil Univ. lsh95ss@netsgo.com Database Benchmark

Benchmarks To rate the performance of various hardware/software platforms in the same way To measure performance in different application domains on different platforms. To compare ratings of different platforms in terms of their relative price-performance ratios before deciding on an acquisition. Database Benchmark

Custom & Generic Benchmark Custom benchmark Created by a particular customer based on a specific application Measured on a number of platforms to aid in an acquisition decision Very cost and hard to maintain Generic benchmark Created to be generally useful for a large group of potential customers with applications that are relatively standard across an application domain. Also known as domain-specific benchmark Hybrid benchmark generic benchmark + custom benchmark Database Benchmark

How to Normalize Results Difficult to compare the performance of the wide variety of database software/hardware platforms (from IBM mainframe to PC) Current practice Requires domain-specific qualifications that must be met e.g.) ACID properties of transaction for OLTP benchmarks Reports two specific measures a measure of peak performance, say transaction per second (tps) a measure of price/performance, say dollars per tps Database Benchmark

Scalar measures vs. Vector measures Scalar measures: the idea of using a single performance measure Vector measures Set Query benchmark (CPU, I/O, Elapsed Time) Vector measures allow users to distinguish different performance effects in terms of functional origins, namely separability Database Benchmark

Criteria for Domain-Specific Benchmarks Relevance performs typical operations within the problem domain Portability easy to implement on many different systems and architectures Scalability applicable small and large computer systems Simplicity easily understandable otherwise it would lack credibility Database Benchmark

A genus of Benchmarks Wisconsin benchmark (1983) TP1/DebitCredit benchmark (1985) Transaction Processing Council formed (Aug. 1988) TPC-A (Nov. 1989) TPC-B (Aug. 1990) AS3AP (1991) SUN benchmark (001, object operations version benchmark) (1992) TPC-C (July 1992) 007 Benchmark (1993) Sequoia 2000 Storage Benchmark (1993) TPC-D (April, 1995) Database Benchmark

TPC Benchmarks Originally eight vendors, currently 35 vendors joined, 1988 To define transaction processing and database benchmarks To limits the possibility that one vendor will change the rules to its own advantage. Database Benchmark

Benchmarks and Application Domains OLTP benchmark DSS (Decision Support System) OODBS (Object-Oriented Database System) ECAD (Electronic Computer-Aided Design) DRS (Document Retrieval Service) Database Benchmark

OLTP Domain (1) 1985: DebitCredit benchmark 1985: TP1 benchmark Dealt with concurrent order entries by clerks or tellers on thousands of terminals in large applications Stylized representation of a large bank and a multi-user workload consisting of a single database transaction which stressed the bottlenecks just listed. 1985: TP1 benchmark Removed the multi-thread requirement and drove the logic with a small number of batch threads. Easier to run and gave better price-performance numbers Database Benchmark

OLTP Domain (2) 1989 : TPC-A 1990 : TPC-B 1992 : TPC-C Standardized DebitCredit with some variations from the original 1990 : TPC-B Standardized TP1 No network environment 1992 : TPC-C The current OLTP benchmark Quite complex and difficult to run. Database Benchmark

Decision Support System (DSS) Domain (1) relatively small numbers of users read-only queries Update activity may be undesirable in DSS an extract of operational data is often placed in a data warehouse Parallelism is another important capability in DSS applications. Database Benchmark

Decision Support System (DSS) Domain (2) 1983 : Wisconsin Measures the elapsed time for a relational database system To perform a set of queries requested one after another by a single user. 1993 : AS3AP Created to improve on the Wisconsin benchmark by building scaleability considerations into the fundamental design Introduce the concept of equivalent database size 1993 : Set Query A single-user benchmark with several suites of queries Attempted to offer guidance on how to use the vector of measures Database Benchmark

Decision Support System (DSS) Domain (3) 1995 : TPC-D benchmark Thought of as executing multi-user queries to analyze the sort of operational business data entered by TPC-C Complex decision support Multitable join, sorting, aggregation, ... Queries per hour (QPH) Dollars per query per hour ($/QPH) Database Benchmark

OODBS, Engineering Workstation 1993 : 001 benchmark relatively simple single-user workstation emphasizes main-memory-resident access to data and claims to be independent of the underlying data model 1993 : 007 benchmark Created to provide comprehensive measure of some of the capabilities of OODBS systems that were not covered by OO1. Database Benchmark

Full Text Retrieval(FTR) created to represent the workload generated by a generic document retrieval service modeled after the TPC benchmarks, with a multi-user workload and a performance rating in search transactions per minute (tpm-S) created based on a good deal of experience with real document retrieval applications. Database Benchmark

Research issues What are good criteria for domain-specific benchmarks Currently, relevance, portability, scalability, simplicity What about scalar versus vector measures How to incorporate actual costs of system ownership Multi-user benchmark How to achieve saparability of performance effects How to develop an analytic framework New applications or models BLOBs Database Benchmark