1 Chapter 2: Simple Comparative Experiments (SCE) Simple comparative experiments: experiments that compare two conditions (treatments) –The hypothesis.

Slides:



Advertisements
Similar presentations
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 9 Inferences Based on Two Samples.
Advertisements

11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
BPS - 5th Ed. Chapter 241 One-Way Analysis of Variance: Comparing Several Means.
CHAPTER 2 Building Empirical Model. Basic Statistical Concepts Consider this situation: The tension bond strength of portland cement mortar is an important.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Some Basic Statistical Concepts
Design of Engineering Experiments - Experiments with Random Factors
Copyright © Cengage Learning. All rights reserved. 9 Inferences Based on Two Samples.
10-1 Introduction 10-2 Inference for a Difference in Means of Two Normal Distributions, Variances Known Figure 10-1 Two independent populations.
Topic 2: Statistical Concepts and Market Returns
Chapter 11 Multiple Regression.
EEM332 Design of Experiments En. Mohd Nazri Mahmud
Chapter 2 Simple Comparative Experiments
Testing the Difference Between Means (Small Independent Samples)
Chapter 2Design & Analysis of Experiments 7E 2009 Montgomery 1 Chapter 2 –Basic Statistical Methods Describing sample data –Random samples –Sample mean,
Chapter 11: Inference for Distributions
Inferences About Process Quality
Chapter 9 Hypothesis Testing.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
5-3 Inference on the Means of Two Populations, Variances Unknown
Chapter 9 Title and Outline 1 9 Tests of Hypotheses for a Single Sample 9-1 Hypothesis Testing Statistical Hypotheses Tests of Statistical.
Statistical Inference for Two Samples
Chapter 24: Comparing Means.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 9 Hypothesis Testing.
Experimental Statistics - week 2
ISE 352: Design of Experiments
Statistika & Rancangan Percobaan
Copyright © Cengage Learning. All rights reserved. 13 Linear Correlation and Regression Analysis.
Regression Analysis (2)
STAT 5372: Experimental Statistics Wayne Woodward Office: Office: 143 Heroy Phone: Phone: (214) URL: URL: faculty.smu.edu/waynew.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Copyright © Cengage Learning. All rights reserved. 10 Inferences Involving Two Populations.
More About Significance Tests
NONPARAMETRIC STATISTICS
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Statistical Inferences Based on Two Samples Chapter 9.
1 Design of Engineering Experiments Part 2 – Basic Statistical Concepts Simple comparative experiments –The hypothesis testing framework –The two-sample.
Topics: Statistics & Experimental Design The Human Visual System Color Science Light Sources: Radiometry/Photometry Geometric Optics Tone-transfer Function.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Basic concept Measures of central tendency Measures of central tendency Measures of dispersion & variability.
1 10 Statistical Inference for Two Samples 10-1 Inference on the Difference in Means of Two Normal Distributions, Variances Known Hypothesis tests.
Design of Engineering Experiments Part 2 – Basic Statistical Concepts
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Section Inference about Two Means: Independent Samples 11.3.
DOX 6E Montgomery1 Design of Engineering Experiments Part 9 – Experiments with Random Factors Text reference, Chapter 13, Pg. 484 Previous chapters have.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 8 Hypothesis Testing.
1 9 Tests of Hypotheses for a Single Sample. © John Wiley & Sons, Inc. Applied Statistics and Probability for Engineers, by Montgomery and Runger. 9-1.
Lesson Comparing Two Means. Knowledge Objectives Describe the three conditions necessary for doing inference involving two population means. Clarify.
AP Statistics Chapter 24 Comparing Means.
DOX 6E Montgomery1 Design of Engineering Experiments Part 2 – Basic Statistical Concepts Simple comparative experiments –The hypothesis testing framework.
Chapter 10 The t Test for Two Independent Samples
Copyright (C) 2002 Houghton Mifflin Company. All rights reserved. 1 Understandable Statistics S eventh Edition By Brase and Brase Prepared by: Lynn Smith.
© Copyright McGraw-Hill 2004
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Copyright © Cengage Learning. All rights reserved. 9 Inferences Based on Two Samples.
ENGR 610 Applied Statistics Fall Week 7 Marshall University CITE Jack Smith.
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
SUMMARY EQT 271 MADAM SITI AISYAH ZAKARIA SEMESTER /2015.
1 Design and Analysis of Experiments (2) Basic Statistics Kyung-Ho Park.
Lecture 7: Bivariate Statistics. 2 Properties of Standard Deviation Variance is just the square of the S.D. If a constant is added to all scores, it has.
Chapter 11: Categorical Data n Chi-square goodness of fit test allows us to examine a single distribution of a categorical variable in a population. n.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 7 Inferences Concerning Means.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
4-1 Statistical Inference Statistical inference is to make decisions or draw conclusions about a population using the information contained in a sample.
Chapter 7 Hypothesis Testing with One Sample.
Statistical Quality Control, 7th Edition by Douglas C. Montgomery.
Chapter 4. Inference about Process Quality
Math 4030 – 10b Inferences Concerning Variances: Hypothesis Testing
Chapter 2 Simple Comparative Experiments
Chapter 8 Hypothesis Testing with Two Samples.
9 Tests of Hypotheses for a Single Sample CHAPTER OUTLINE
Presentation transcript:

1 Chapter 2: Simple Comparative Experiments (SCE) Simple comparative experiments: experiments that compare two conditions (treatments) –The hypothesis testing framework –The two-sample t-test –Checking assumptions, validity

2 Portland Cement Formulation (page 23) Average tension bond sterngths (ABS) differ by what seems nontrivial amount. Not obvois that this difference is large enough imply that the two formulations really are diff. Diff may be due to sampling fluctuation and the two formulations are really identical. Possibly another two samples would give opposite results with strength of MM exceeding that of UM. Hypothesis testing can be used to assist in comparing these formulations. Hypothesis testing allows the comparison to be made on objective terms, with knowledge of risks associated with searching the wrong conclusion

3 Graphical View of the Data Dot Diagram, Fig. 2-1, pp. 24 Response variable is a random variable Random variable: 1.Discrete 2.continuous

4 Box Plots, Fig. 2-3, pp. 26 Displays min, max, lower and upper quartile, and the median Histogram

Probability Distributions 5 Probability structure of a Random variable, y, is described by its probability distribution. y is discrete: p(y) is the probability function of y (F2- 4a) y is continuous: f(y) is the probability density function of y (F2-4b)

Probability Distributions Properties of probability distributions 6 y-discrete: y-continuous:

Probability Distributions mean, variance, and expected values 7

Probability Distributions Basic Properties 1.E(c) = c 2.E(y) =  3.E(cy) = c E(y) = c  4.V(c) = 0 5.V(y) =  2 6.V(cy) = c 2 V(y) = c 2  2 8

Probability Distributions Basic Properties E(y 1 +y 2 ) = E(y 1 )+E(y 2 ) =  1 +  2 Cov(y 1,y 2 ) = E[(y 1 -  1 )(y 2 -  2 )] Covariance: measure of the linear association between y 1 and y 2. E(y 1.y 2 ) = E(y 1 ).E(y 2 ) =  1.  2 (y 1 and y 2 are indep) 9

Sampling and Sampling Distributions The objective of statistical inference is to draw conclusions about a population using a sample from that population. Random Sampling: each of N!/(N-n)!n! samples has an equal probability of being chosen. Statistic: any function of observations in a sample that does not contain unknown parameters. Sample mean and sample variance are both statistics. 10

Properties of sample mean and variance Sample mean is a point estimator of population mean  Sample variance is a point estimator of population variance   The point estimator should be unbiased. Long run average should be the parameter that is being estimated. An unbiased estimator should have min variance. Min variance point estimator has a variance that is smaller than the variance of any other estimator of that parameter. 11

Degrees of freedom (n-1) in the previous eq is called the NDOF of the sum of squares. NDOF of a sum of squares is equal to the no. of indep elements in the sum of squares Because, only (n-1) of the n elements are indep, implying that SS has (n- 1) DOF 12

The normal and other sampling distributions Normal Distribution y is distributed normally with mean  and variance  2 Standard normal distribution:  =0 and  2 =1 13

Central Limit Theorem If y 1, y 2, …, y n is a sequence of n independent and identically distributed random variables with E(y i ) =  and V(y i ) =  2 (both finite) and x = y 1 +y 2 +…+y n, then has an approximate N(0,1) distribution. This implies that the distribution of the sample averages follows a normal distribution with This approximation require a relatively large sample size (n≥30) 14

Chi-Square or  2 distribution If z 1, z 2, …, z n are normally and independently distributed random variables with mean 0 and variance 1 NID(0,1), then the random variable follows the chi-square distribution with k DOF. 15

Chi-Square or  2 distribution The distribution is asymmetric (skewed)  = k  2 = 2k Appendix III 16

Chi-Square or  2 distribution y 1, y 2, …, y n is a random sample from N( ,  ), then SS/  2 is distributed as chi-square with n-1 DOF 17

Chi-Square or  2 distribution If the observations in the sample are NID(  ), then the distribution of S 2 is and,,,, Thus, the sampling distribution of the sample variance is a constant times the chi- square distribution if the population is normally distributed 18

Chi-Square or  2 distribution Example: The Acme Battery Company has developed a new cell phone battery. On average, the battery lasts 60 minutes on a single charge. The standard deviation is 4.14 minutes. a)Suppose the manufacturing department runs a quality control test. They randomly select 7 batteries. The standard deviation of the selected batteries is 6 minutes. What would be the chi-square statistic represented by this test? b)If another sample of 7 battery was selected, what is the probability that the sample standard deviation is greater than 6? DOX 6E Montgomery19

Chi-Square or  2 distribution Solution a) We know the following: –The standard deviation of the population is 4.14 minutes. –The standard deviation of the sample is 6 minutes. –The number of sample observations is 7. To compute the chi-square statistic, we plug these data in the chi-square equation, as shown below.  2 = [ ( n - 1 ) * s 2 ] / σ 2  2 = [ ( ) * 6 2 ] / = 12.6 DOX 6E Montgomery20

b) To find the probability of having a sample standard deviation S > 6, we refer to the Chi- square distribution tables, we find the value of  corresponding to chi-square = 12.6 and 6 degrees of freedom. This will give  0.05 which is the probability of having S > 6 DOX 6E Montgomery21 Chi-Square or  2 distribution

DOX 6E Montgomery22

t distribution with k DOF If z and are indpendent normal and chi- square random variables, the random variable Follow t distribution with k DOF as follows: 23

t distribution with k DOF  = 0 and  2 = k/(k-2) for k>2 If k=infinity, t becomes standard normal If y 1, y 2, …, y n is a random sample from N( ,  ), then is distributed as t with n-1 DOF 24

t distribution - example Example: Acme Corporation manufactures light bulbs. The CEO claims that an average Acme light bulb lasts 300 days. A researcher randomly selects 15 bulbs for testing. The sampled bulbs last an average of 290 days, with a standard deviation of 56 days. If the CEO's claim were true, what is the probability that 15 randomly selected bulbs would have an average life of no more than 290 days? DOX 6E Montgomery25

t distribution – Example Solution To find P(x-bar<290), The first thing we need to do is compute the t score, based on the following equation: –t = [ x - μ ] / [ s / sqrt( n ) ] t = ( ) / [ 56 / sqrt( 15) ] = P(t 0.692) DOX 6E Montgomery26

t distribution – Example From the t-distribution tables, we have  = 25% which correspond to the probability of having the sample average less than 290 DOX 6E Montgomery27

DOX 6E Montgomery28

F distribution If and are two independent chi- square random variables with u and v DOF, then the ratio Follows the F dist with u numerator DOF and v denominator DOF 29

F distribution Two independent normal populations with common variance  2. If y 11, y 12, …, y 1n1 is a random sample of n 1 observations from 1 st population and y 21, y 22, …, y 2n2 is a random sample of n 2 observations from 2 nd population, then 30

31 The Hypothesis Testing Framework Statistical hypothesis testing is a useful framework for many experimental situations We will use a procedure known as the two- sample t-test

32 The Hypothesis Testing Framework Sampling from a normal distribution Statistical hypotheses:

33 Estimation of Parameters

34 Summary Statistics (pg. 36) Formulation 1 “New recipe” Formulation 2 “Original recipe”

35 How the Two-Sample t-Test Works:

36 How the Two-Sample t-Test Works:

37 How the Two-Sample t-Test Works: Values of t 0 that are near zero are consistent with the null hypothesis Values of t 0 that are very different from zero are consistent with the alternative hypothesis t 0 is a “distance” measure-how far apart the averages are expressed in standard deviation units Notice the interpretation of t 0 as a signal-to-noise ratio

38 The Two-Sample (Pooled) t-Test

39 The Two-Sample (Pooled) t-Test So far, we haven’t really done any “statistics” We need an objective basis for deciding how large the test statistic t 0 really is. t 0 = -2.20

40 The Two-Sample (Pooled) t-Test A value of t 0 between –2.101 and is consistent with equality of means It is possible for the means to be equal and t 0 to exceed either or –2.101, but it would be a “rare event” … leads to the conclusion that the means are different Could also use the P-value approach t 0 = -2.20

Use of P-value in Hypothesis testing P-value: smallest level of significance that would lead to rejection of the null hypothesis H o It is customary to call the test statistic significant when H o is rejected. Therefore, the P-value is the smallest level  at which the data are significant. 41

42 Minitab Two-Sample t-Test Results

Checking Assumptions – The Normal Probability Plot Assumptions 1.Equal variance 2.Normality Procedure : 1.Rank the observations in the sample in an ascending order. 2.Plot ordered observations vs. observed cumulative frequency (j- 0.5)/n 3.If the plotted points deviate significantly from straight line, the hypothesized model in not appropriate. 43

44 Checking Assumptions – The Normal Probability Plot

The mean is estimated as the 50 th percentile on the probability plot. The standard deviation is estimated as the differnce between the 84 th and 50 th percentiles. The assumption of equal population variances is simply verified by comparing the slopes of the two straight lines in F

46 Importance of the t-Test Provides an objective framework for simple comparative experiments Could be used to test all relevant hypotheses in a two-level factorial design.

47 Confidence Intervals (See pg. 43) Hypothesis testing gives an objective statement concerning the difference in means, but it doesn’t specify “how different” they are General form of a confidence interval The 100(1- α)% confidence interval on the difference in two means:

Hypothesis testing The test statitic becomes This statistic is not distributed exactly as t. The distribution of t o is well approximated by t if we use as the DOF 48

Hypothesis testing The test statitic becomes If both populations are normal, or if the sample sizes are large enough, the distribution of z o is N(0,1) if the null hypothesis is true. Thus, the critical region would be found using the normal distribution rather than the t. We would reject H o, if where z  is the upper  2 percentage point of the standard normal distribution 49

Hypothesis testing The 100(1-  ) percent confidence interval: 50

Hypothesis testing Comparing a single mean to a specified value The hypothesises are: H o :  =  o and H 1 :  ≠  o If the population is normal with known variance, or if the population is non-normal but the sample size is large enough, then the hypothesis may be tested by direct application of the normal distribution. Test statistic If Ho is true, then the distribution of zo is N(0,1). Therefore, H o is rejected if 51

Hypothesis testing Comparing a single mean to a specified value The value of  o is usually determined in one of three ways: 1.From past evidence, knowledge, or experimentation 2.The result of some theory or model describing the situation under study 3.The result of contractual specifications 52

Hypothesis testing Comparing a single mean to a specified value If the variance of the population is known, we assume that the population is normally distributed. Test statistic H o is rejected if The 100(1-  ) percent confidence interval 53

The paired comparison problem The two tip hardness experiment Statistical model j th paired difference Expected value of paired difference Testing hypothesis: H o :  d =0 and H 1 :  d ≠0 Test statistic: 54

The paired comparison problem The two tip hardness experiment Randomized block design Block: homogenous experimental unit The block represents a restriction on complete randomization because the treatment combinations are only randomized within the block 55 Randomized BlockComplete Randomization DOFn-1 = 92n-2 = 18 Standard deviationS d = 1.2S p = 2.32 Confidence interval on  1 -  ± ±2.18

Inferences about the variablity of normal distributions H o :  2 =  o and H 1 :  2 ≠  o Test statistic Ho is rejected if or The 100(1-a) percent confidence interval 56

Inferences about the variablity of normal distributions Test statistic H o is rejected if or 57