11 Simple Linear Regression and Correlation CHAPTER OUTLINE

Slides:



Advertisements
Similar presentations
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Advertisements

Forecasting Using the Simple Linear Regression Model and Correlation
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Probability & Statistical Inference Lecture 9
12-1 Multiple Linear Regression Models Introduction Many applications of regression analysis involve situations in which there are more than.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
12 Multiple Linear Regression CHAPTER OUTLINE
Learning Objectives Copyright © 2004 John Wiley & Sons, Inc. Bivariate Correlation and Regression CHAPTER Thirteen.
Probabilistic & Statistical Techniques Eng. Tamer Eshtawi First Semester Eng. Tamer Eshtawi First Semester
Linear regression models
Objectives (BPS chapter 24)
1 Chapter 2 Simple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Chapter 12 Simple Linear Regression
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
1 Chapter 3 Multiple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Chapter 13 Introduction to Linear Regression and Correlation Analysis
SIMPLE LINEAR REGRESSION
8 Statistical Intervals for a Single Sample CHAPTER OUTLINE
Simple Linear Regression Analysis
Introduction to Probability and Statistics Linear Regression and Correlation.
SIMPLE LINEAR REGRESSION
© 2000 Prentice-Hall, Inc. Chap Forecasting Using the Simple Linear Regression Model and Correlation.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Simple Linear Regression and Correlation
Introduction to Regression Analysis, Chapter 13,
Simple Linear Regression Analysis
13 Design and Analysis of Single-Factor Experiments:
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS & Updated by SPIROS VELIANITIS.
Lecture 5 Correlation and Regression
Correlation & Regression
Chapter 9 Title and Outline 1 9 Tests of Hypotheses for a Single Sample 9-1 Hypothesis Testing Statistical Hypotheses Tests of Statistical.
Regression and Correlation Methods Judy Zhong Ph.D.
SIMPLE LINEAR REGRESSION
Introduction to Linear Regression and Correlation Analysis
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 12 Analyzing the Association Between Quantitative Variables: Regression Analysis Section.
Chapter 11 Simple Regression
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Simple Linear Regression Models
1 1 Slide © 2005 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Introduction to Linear Regression
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
1 10 Statistical Inference for Two Samples 10-1 Inference on the Difference in Means of Two Normal Distributions, Variances Known Hypothesis tests.
Chapter 11 Linear Regression Straight Lines, Least-Squares and More Chapter 11A Can you pick out the straight lines and find the least-square?
Basic Concepts of Correlation. Definition A correlation exists between two variables when the values of one are somehow associated with the values of.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
6-1 Introduction To Empirical Models Based on the scatter diagram, it is probably reasonable to assume that the mean of the random variable Y is.
1 11 Simple Linear Regression and Correlation 11-1 Empirical Models 11-2 Simple Linear Regression 11-3 Properties of the Least Squares Estimators 11-4.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
MARKETING RESEARCH CHAPTER 18 :Correlation and Regression.
1 9 Tests of Hypotheses for a Single Sample. © John Wiley & Sons, Inc. Applied Statistics and Probability for Engineers, by Montgomery and Runger. 9-1.
VI. Regression Analysis A. Simple Linear Regression 1. Scatter Plots Regression analysis is best taught via an example. Pencil lead is a ceramic material.
Regression Analysis © 2007 Prentice Hall17-1. © 2007 Prentice Hall17-2 Chapter Outline 1) Correlations 2) Bivariate Regression 3) Statistics Associated.
Correlation & Regression Analysis
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Simple Linear Regression Analysis Chapter 13.
Chapter 12 Simple Linear Regression n Simple Linear Regression Model n Least Squares Method n Coefficient of Determination n Model Assumptions n Testing.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
Chapter Outline EMPIRICAL MODELS 11-2 SIMPLE LINEAR REGRESSION 11-3 PROPERTIES OF THE LEAST SQUARES ESTIMATORS 11-4 SOME COMMENTS ON USES OF REGRESSION.
Applied Statistics and Probability for Engineers
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Slides by JOHN LOUCKS St. Edward’s University.
6-1 Introduction To Empirical Models
Prepared by Lee Revere and John Large
Simple Linear Regression
SIMPLE LINEAR REGRESSION
Product moment correlation
SIMPLE LINEAR REGRESSION
St. Edward’s University
Presentation transcript:

11 Simple Linear Regression and Correlation CHAPTER OUTLINE 11-1 Empirical Models 11-6 Prediction of New Observations 11-2 Simple Linear Regression 11-7 Adequacy of the Regression Model 11-3 Properties of the Least Squares Estimators 11-7.1 Residual analysis 11-4 Hypothesis Test in Simple Linear Regression 11-7.2 Coefficient of determination (R2) 11-8 Correlation 11-4.1 Use of t-tests 11-9 Regression on Transformed Variables 11-4.2 Analysis of variance approach to test significance of regression 11-10 Logistics Regression 11-5 Confidence Intervals 11-5.1 Confidence intervals on the slope and intercept 11-5.2 Confidence interval on the mean response Dear Instructor: This file is an adaptation of the 4th edition slides for the 5th edition. It will be replaced as slides are developed following the style of the Chapters 1-7 slides. Chapter 11 Table of Contents

Learning Objectives for Chapter 11 After careful study of this chapter, you should be able to do the following: Use simple linear regression for building empirical models to engineering and scientific data. Understand how the method of least squares is used to estimate the parameters in a linear regression model. Analyze residuals to determine if the regression model is an adequate fit to the data or to see if any underlying assumptions are violated. Test the statistical hypotheses and construct confidence intervals on the regression model parameters. Use the regression model to make a prediction of a future observation and construct an appropriate prediction interval on the future observation. Apply the correlation model. Use simple transformations to achieve a linear regression model. Chapter 11 Learning Objectives

11-1: Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis is a statistical technique that is very useful for these types of problems. For example, in a chemical process, suppose that the yield of the product is related to the process-operating temperature. Regression analysis can be used to build a model to predict yield at a given temperature level.

11-1: Empirical Models

11-1: Empirical Models Figure 11-1 Scatter Diagram of oxygen purity versus hydrocarbon level from Table 11-1.

11-1: Empirical Models Based on the scatter diagram, it is probably reasonable to assume that the mean of the random variable Y is related to x by the following straight-line relationship: where the slope and intercept of the line are called regression coefficients. The simple linear regression model is given by where  is the random error term.

11-1: Empirical Models We think of the regression model as an empirical model. Suppose that the mean and variance of  are 0 and 2, respectively, then The variance of Y given x is

11-1: Empirical Models The true regression model is a line of mean values: where 1 can be interpreted as the change in the mean of Y for a unit change in x. Also, the variability of Y at a particular value of x is determined by the error variance, 2. This implies there is a distribution of Y-values at each x and that the variance of this distribution is the same at each x.

11-1: Empirical Models Figure 11-2 The distribution of Y for a given value of x for the oxygen purity-hydrocarbon data.

11-2: Simple Linear Regression The case of simple linear regression considers a single regressor or predictor x and a dependent or response variable Y. The expected value of Y at each level of x is a random variable: We assume that each observation, Y, can be described by the model

11-2: Simple Linear Regression Suppose that we have n pairs of observations (x1, y1), (x2, y2), …, (xn, yn). Figure 11-3 Deviations of the data from the estimated regression model.

11-2: Simple Linear Regression The method of least squares is used to estimate the parameters, 0 and 1 by minimizing the sum of the squares of the vertical deviations in Figure 11-3. Figure 11-3 Deviations of the data from the estimated regression model.

11-2: Simple Linear Regression Using Equation 11-2, the n observations in the sample can be expressed as The sum of the squares of the deviations of the observations from the true regression line is

11-2: Simple Linear Regression

11-2: Simple Linear Regression

11-2: Simple Linear Regression Definition

11-2: Simple Linear Regression

11-2: Simple Linear Regression Notation

11-2: Simple Linear Regression Example 11-1

11-2: Simple Linear Regression Example 11-1

11-2: Simple Linear Regression Example 11-1 Figure 11-4 Scatter plot of oxygen purity y versus hydrocarbon level x and regression model ŷ = 74.20 + 14.97x.

11-2: Simple Linear Regression Example 11-1

11-2: Simple Linear Regression Estimating 2 The error sum of squares is It can be shown that the expected value of the error sum of squares is E(SSE) = (n – 2)2.

11-2: Simple Linear Regression Estimating 2 An unbiased estimator of 2 is where SSE can be easily computed using

11-3: Properties of the Least Squares Estimators Slope Properties Intercept Properties

11-4: Hypothesis Tests in Simple Linear Regression 11-4.1 Use of t-Tests Suppose we wish to test An appropriate test statistic would be

11-4: Hypothesis Tests in Simple Linear Regression 11-4.1 Use of t-Tests The test statistic could also be written as: We would reject the null hypothesis if

11-4: Hypothesis Tests in Simple Linear Regression 11-4.1 Use of t-Tests Suppose we wish to test An appropriate test statistic would be

11-4: Hypothesis Tests in Simple Linear Regression 11-4.1 Use of t-Tests We would reject the null hypothesis if

11-4: Hypothesis Tests in Simple Linear Regression 11-4.1 Use of t-Tests An important special case of the hypotheses of Equation 11-18 is These hypotheses relate to the significance of regression. Failure to reject H0 is equivalent to concluding that there is no linear relationship between x and Y.

11-4: Hypothesis Tests in Simple Linear Regression Figure 11-5 The hypothesis H0: 1 = 0 is not rejected.

11-4: Hypothesis Tests in Simple Linear Regression Figure 11-6 The hypothesis H0: 1 = 0 is rejected.

11-4: Hypothesis Tests in Simple Linear Regression Example 11-2

11-4: Hypothesis Tests in Simple Linear Regression 11-4.2 Analysis of Variance Approach to Test Significance of Regression The analysis of variance identity is Symbolically,

11-4: Hypothesis Tests in Simple Linear Regression 11-4.2 Analysis of Variance Approach to Test Significance of Regression If the null hypothesis, H0: 1 = 0 is true, the statistic follows the F1,n-2 distribution and we would reject if f0 > f,1,n-2.

11-4: Hypothesis Tests in Simple Linear Regression 11-4.2 Analysis of Variance Approach to Test Significance of Regression The quantities, MSR and MSE are called mean squares. Analysis of variance table:

11-4: Hypothesis Tests in Simple Linear Regression Example 11-3

11-4: Hypothesis Tests in Simple Linear Regression

11-5: Confidence Intervals 11-5.1 Confidence Intervals on the Slope and Intercept Definition

11-5: Confidence Intervals Example 11-4

11-5: Confidence Intervals 11-5.2 Confidence Interval on the Mean Response Definition

11-5: Confidence Intervals Example 11-5

11-5: Confidence Intervals Example 11-5

11-5: Confidence Intervals Example 11-5

11-5: Confidence Intervals Figure 11-7 Figure 11-7 Scatter diagram of oxygen purity data from Example 11-1 with fitted regression line and 95 percent confidence limits on Y|x0.

11-6: Prediction of New Observations If x0 is the value of the regressor variable of interest, is the point estimator of the new or future value of the response, Y0.

11-6: Prediction of New Observations Definition

11-6: Prediction of New Observations Example 11-6

11-6: Prediction of New Observations Example 11-6

11-6: Prediction of New Observations Figure 11-8 Figure 11-8 Scatter diagram of oxygen purity data from Example 11-1 with fitted regression line, 95% prediction limits (outer lines) , and 95% confidence limits on Y|x0.

11-7: Adequacy of the Regression Model Fitting a regression model requires several assumptions. Errors are uncorrelated random variables with mean zero; Errors have constant variance; and, Errors be normally distributed. The analyst should always consider the validity of these assumptions to be doubtful and conduct analyses to examine the adequacy of the model

11-7: Adequacy of the Regression Model 11-7.1 Residual Analysis The residuals from a regression model are ei = yi - ŷi , where yi is an actual observation and ŷi is the corresponding fitted value from the regression model. Analysis of the residuals is frequently helpful in checking the assumption that the errors are approximately normally distributed with constant variance, and in determining whether additional terms in the model would be useful.

11-7: Adequacy of the Regression Model 11-7.1 Residual Analysis Figure 11-9 Patterns for residual plots. (a) satisfactory, (b) funnel, (c) double bow, (d) nonlinear. [Adapted from Montgomery, Peck, and Vining (2006).]

11-7: Adequacy of the Regression Model Example 11-7

11-7: Adequacy of the Regression Model Example 11-7

11-7: Adequacy of the Regression Model Example 11-7 Figure 11-10 Normal probability plot of residuals, Example 11-7.

11-7: Adequacy of the Regression Model Example 11-7 Figure 11-11 Plot of residuals versus predicted oxygen purity, ŷ, Example 11-7.

11-7: Adequacy of the Regression Model 11-7.2 Coefficient of Determination (R2) The quantity is called the coefficient of determination and is often used to judge the adequacy of a regression model. 0  R2  1; We often refer (loosely) to R2 as the amount of variability in the data explained or accounted for by the regression model.

11-7: Adequacy of the Regression Model 11-7.2 Coefficient of Determination (R2) For the oxygen purity regression model, R2 = SSR/SST = 152.13/173.38 = 0.877 Thus, the model accounts for 87.7% of the variability in the data.