LECTURE 3: ANALYSIS OF EXPERIMENTAL DATA

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Lesson 10: Linear Regression and Correlation
Forecasting Using the Simple Linear Regression Model and Correlation
Chapter 7 Statistical Data Treatment and Evaluation
Hypothesis Testing Steps in Hypothesis Testing:
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Probabilistic & Statistical Techniques Eng. Tamer Eshtawi First Semester Eng. Tamer Eshtawi First Semester
Correlation and Regression
Simple Linear Regression and Correlation
Objectives (BPS chapter 24)
© The McGraw-Hill Companies, Inc., 2000 CorrelationandRegression Further Mathematics - CORE.
Chapter 10 Simple Regression.
Chapter 13 Introduction to Linear Regression and Correlation Analysis
SIMPLE LINEAR REGRESSION
Chapter Topics Types of Regression Models
Engineering Computation Curve Fitting 1 Curve Fitting By Least-Squares Regression and Spline Interpolation Part 7.
Chapter 11 Multiple Regression.
CHAPTER 6 Statistical Analysis of Experimental Data
Simple Linear Regression Analysis
SIMPLE LINEAR REGRESSION
© 2000 Prentice-Hall, Inc. Chap Forecasting Using the Simple Linear Regression Model and Correlation.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Chapter 14 Introduction to Linear Regression and Correlation Analysis
CHAPTER 6 Statistical Analysis of Experimental Data
Simple Linear Regression and Correlation
SIMPLE LINEAR REGRESSION
Introduction to Linear Regression and Correlation Analysis
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Statistics for Data Miners: Part I (continued) S.T. Balke.
1 G Lect 10a G Lecture 10a Revisited Example: Okazaki’s inferences from a survey Inferences on correlation Correlation: Power and effect.
© The McGraw-Hill Companies, Inc., Chapter 11 Correlation and Regression.
Basic Probability (Chapter 2, W.J.Decoursey, 2003) Objectives: -Define probability and its relationship to relative frequency of an event. -Learn the basic.
Introduction to Linear Regression
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Topic 10 - Linear Regression Least squares principle - pages 301 – – 309 Hypothesis tests/confidence intervals/prediction intervals for regression.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Y X 0 X and Y are not perfectly correlated. However, there is on average a positive relationship between Y and X X1X1 X2X2.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Probability = Relative Frequency. Typical Distribution for a Discrete Variable.
CHEMISTRY ANALYTICAL CHEMISTRY Fall Lecture 6.
Review Lecture 51 Tue, Dec 13, Chapter 1 Sections 1.1 – 1.4. Sections 1.1 – 1.4. Be familiar with the language and principles of hypothesis testing.
Chapter 8: Simple Linear Regression Yang Zhenlin.
Analysis of Experimental Data; Introduction
Copyright (C) 2002 Houghton Mifflin Company. All rights reserved. 1 Understandable Statistics Seventh Edition By Brase and Brase Prepared by: Lynn Smith.
Chapter 20 Statistical Considerations Lecture Slides The McGraw-Hill Companies © 2012.
Lecture 8: Measurement Errors 1. Objectives List some sources of measurement errors. Classify measurement errors into systematic and random errors. Study.
©The McGraw-Hill Companies, Inc. 2008McGraw-Hill/Irwin Linear Regression and Correlation Chapter 13.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
REGRESSION AND CORRELATION SIMPLE LINEAR REGRESSION 10.2 SCATTER DIAGRAM 10.3 GRAPHICAL METHOD FOR DETERMINING REGRESSION 10.4 LEAST SQUARE METHOD.
Statistics and probability Dr. Khaled Ismael Almghari Phone No:
Inference about the slope parameter and correlation
Regression and Correlation
Regression Analysis AGEC 784.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Modify—use bio. IB book  IB Biology Topic 1: Statistical Analysis
CHAPTER 29: Multiple Regression*
PENGOLAHAN DAN PENYAJIAN
Correlation and Regression
Simple Linear Regression
SIMPLE LINEAR REGRESSION
SIMPLE LINEAR REGRESSION
Lecture # 2 MATHEMATICAL STATISTICS
Regression and Correlation of Data
Introductory Statistics
Presentation transcript:

LECTURE 3: ANALYSIS OF EXPERIMENTAL DATA MEASUREMENT AND INSTRUMENTATION BMCC 3743 LECTURE 3: ANALYSIS OF EXPERIMENTAL DATA Mochamad Safarudin Faculty of Mechanical Engineering, UTeM 2010

Contents Introduction Measures of dispersion Parameter estimation Criterion for rejection questionable data points Correlation of experimental data

Introduction Needed in all measurements with random inputs, e.g. random broadband sound/noise Tyre/road noise, rain drops, waterfall Some important terms are: Random variable (continuous or discrete), histogram, bins, population, sample, distribution function, parameter, event, statistic, probability.

Terminology Population : the entire collection of objects, measurements, observations and so on whose properties are under consideration Sample: a representative subset of a population on which an experiment is performed and numerical data are obtained

Contents Measures of dispersion Introduction Parameter estimation Criterion for rejection questionable data points Correlation of experimental data

Measures of dispersion =>Measures of data spreading or variability Deviation (error) is defined as Mean deviation is defined as Population standard deviation is defined as

Measures of dispersion Sample standard deviation is defined as is used when data of a sample are used to estimate population std dev. Variance is defined as

Exercise Find the mean, median, standard deviation and variance of this measurement: 1089, 1092, 1094, 1095, 1098, 1100, 1104, 1105, 1107, 1108, 1110, 1112, 1115

Answer to exercise Mean = 1103 (1102.2) Median = 1104 Std deviation = 5.79 (7.89) Variance = 33.49 (62.18)

Contents Parameter estimation Introduction Measures of dispersion Criterion for rejection questionable data points Correlation of experimental data

Parameter estimation Generally, Estimation of population mean, is sample mean, . Estimation of population standard deviation, is sample standard deviation, S.

Interval estimation of the population mean Confidence interval is the interval between to , where is an uncertainty. Confidence level is the probability for the population mean to fall within specified interval:

Interval estimation of the population mean Normally referred in terms of , also called level of significance, where confidence level If n is sufficiently large (> 30), we can apply the central limit theorem to find the estimation of the population mean.

Central limit theorem If original population is normal, then distribution for the sample means’ is normal (Gaussian) If original population is not normal and n is large, then distribution for sample means’ is normal If original population is not normal and n is small, then sample means’ follow a normal distribution only approximately.

Normal (Gaussian) distribution When n is large, where Rearranged to get Or with confidence level

Level of Significance (%) Table z Confidence Interval Confidence Level (%) Level of Significance (%) 3.30 99.9 0.1 3.0 99.7 0.3 2.57 99.0 1.0 2.0 95.4 4.6 1.96 95.0 5.0 1.65 90.0 10.0 68.3 31.7 Area under 0 to z

Student’s t distribution When n is small, where Rearranged to get Or with confidence level t table

Interval estimation of the population variance Similarly as before, but now using chi- squared distribution, , (always positive) where

Interval estimation of the population variance Hence, the confidence interval on the population variance is Chi squared table

Contents Criterion for rejection questionable data points Introduction Measures of dispersion Parameter estimation Criterion for rejection questionable data points Correlation of experimental data

Criterion for rejection questionable data points To eliminate data which has low probability of occurrence => use Thompson test. Example: Data consists of nine values, Dn = 12.02, 12.05, 11.96, 11.99, 12.10, 12.03, 12.00, 11.95 and 12.16. = 12.03, S = 0.07 So, calculate deviation:

Criterion for rejection questionable data points From Thompson’s table, when n = 9, then Comparing with where then D9 = 12.16 should be discarded. Recalculate S and to obtain 0.05 and 12.01 respectively. Hence for n = 8, and so remaining data stay. Thompson’s t table

Contents Correlation of experimental data Introduction Measures of dispersion Parameter estimation Criterion for rejection questionable data points Correlation of experimental data

Correlation of experimental data Correlation coefficient Least-square linear fit Linear regression using data transformation

A) Correlation coefficient Case I: Strong, linear relationship between x and y Case II: Weak/no relationship Case III: Pure chance => Use correlation coefficient, rxy to determine Case III

Linear correlation coefficient Given as where +1 means positive slope (perfectly linear relationship) -1 means negative slope (perfectly linear relationship) 0 means no linear correlation

Linear correlation coefficient In practice, we use special Table (using critical values of rt) to determine Case III. If from experimental value of |rxy| is equal or more than rt as given in the Table, then linear relationship exists. If from experimental value of |rxy| is less than rt as given in the Table, then only pure chance => no linear relationship exists.

B) Least-square linear fit To get best straight line on the plot: Simple approach: ruler & eyes More systematic approach: least squares Variation in the data is assumed to be normally distributed and due to random causes To get Y = ax + b, it is assumed that Y values are randomly vary and x values have no error.

Least-square best fit For each value of xi, error for Y values are Then, the sum of squared errors is

Least-square best fit Minimising this equation and solving it for a & b, we get

Least-square best fit Substitute a & b values into Y = ax + b, which is then called the least-squares best fit. To measure how well the best-fit line represents the data, we calculate the standard error of estimate, given by where Sy,x is the standard deviation of the differences between data points and the best-fit line. Its unit is the same as y.

Coefficient of determination …Is another good measure to determine how well the best-fit line represents the data, using For a good fit, must be close to unity.

C) Linear regression using data transformation For some special cases, such as Applying natural logarithm at both sides, gives where ln(a) is a constant, so ln(y) is linearly related to x.

Example Determine the linear correlation between T and V Thermocouples are usually approximately linear devices in a limited range of temperature. A manufacturer of a brand of thermocouple has obtained the following data for a pair of thermocouple wires: T(0C) 20 30 40 50 60 75 100 V(mV) 1.02 1.53 2.05 2.55 3.07 3.56 4.05 Determine the linear correlation between T and V

Tabulate the data using this table: Solution: Tabulate the data using this table: rxy= 0.980392

The following measurements were obtained in the calibration of Another example The following measurements were obtained in the calibration of a pressure transducer: Voltage DP H2O 0.31 1.96 0.65 4.20 0.75 4.90 0.85 5.48 0.91 5.91 1.12 7.30 1.19 7.73 1.38 9.00 1.52 9.90 Determine the best fit straight line Find the coefficient of determination for the best fit

Y=6.56x-0.06

r2= From the result before we can find coeff of determination r2 by tabulating the following values r2= 0.999926

Next Lecture Experimental Uncertainty Analysis End of Lecture 3